<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2023.1083413</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Medicine</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Fast myocardial perfusion SPECT denoising using an attention-guided generative adversarial network</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Sun</surname> <given-names>Jingzhang</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Yang</surname> <given-names>Bang-Hung</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Li</surname> <given-names>Chien-Ying</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Du</surname> <given-names>Yu</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Yi-Hwa</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2166682/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Wu</surname> <given-names>Tung-Hsin</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c002"><sup>&#x002A;</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Mok</surname> <given-names>Greta S. P.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1651965/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Biomedical Imaging Laboratory (BIG), Department of Electrical and Computer Engineering, Faculty of Science and Technology, University of Macau</institution>, <addr-line>Taipa</addr-line>, <country>Macao SAR, China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Biomedical Imaging and Radiological Sciences, National Yang Ming Chiao Tung University</institution>, <addr-line>Hsinchu</addr-line>, <country>Taiwan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Nuclear Medicine, Taipei Veterans General Hospital</institution>, <addr-line>Taipei City</addr-line>, <country>Taiwan</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Internal Medicine, Yale University School of Medicine</institution>, <addr-line>New Haven, CT</addr-line>, <country>United States</country></aff>
<aff id="aff5"><sup>5</sup><institution>Center for Cognitive and Brain Sciences, Institute of Collaborative Innovation, University of Macau</institution>, <addr-line>Taipa</addr-line>, <country>Macao SAR, China</country></aff>
<aff id="aff6"><sup>6</sup><institution>Ministry of Education Frontiers Science Center for Precision Oncology, Faculty of Health Science, University of Macau</institution>, <addr-line>Taipa</addr-line>, <country>Macao SAR, China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Giorgio Treglia, Ente Ospedaliero Cantonale (EOC), Switzerland</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Federico Caobelli, Bern University Hospital, Switzerland; Sangwon Han, University of Ulsan, Republic of Korea</p></fn>
<corresp id="c001">&#x002A;Correspondence: Greta S. P. Mok, <email>gretamok@um.edu.mo</email></corresp>
<corresp id="c002">Tung-Hsin Wu, <email>tunghsinwu@nycu.edu.tw</email></corresp>
<fn fn-type="equal" id="fn002"><p><sup>&#x2020;</sup>These authors have contributed equally to this work</p></fn>
<fn fn-type="other" id="fn004"><p>This article was submitted to Nuclear Medicine, a section of the journal Frontiers in Medicine</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>03</day>
<month>02</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>10</volume>
<elocation-id>1083413</elocation-id>
<history>
<date date-type="received">
<day>29</day>
<month>10</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>01</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2023 Sun, Yang, Li, Du, Liu, Wu and Mok.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Sun, Yang, Li, Du, Liu, Wu and Mok</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<sec>
<title>Purpose</title>
<p>Deep learning-based denoising is promising for myocardial perfusion (MP) SPECT. However, conventional convolutional neural network (CNN)-based methods use fixed-sized convolutional kernels to convolute one region within the receptive field at a time, which would be ineffective for learning the feature dependencies across large regions. The attention mechanism (Att) is able to learn the relationships between the local receptive field and other voxels in the image. In this study, we propose a 3D attention-guided generative adversarial network (AttGAN) for denoising fast MP-SPECT images.</p>
</sec>
<sec>
<title>Methods</title>
<p>Fifty patients who underwent 1184 MBq <sup>99m</sup>Tc-sestamibi stress SPECT/CT scan were retrospectively recruited. Sixty projections were acquired over 180&#x00B0; and the acquisition time was 10 s/view for the full time (FT) mode. Fast MP-SPECT projection images (1 s to 7 s) were generated from the FT list mode data. We further incorporated binary patient defect information (0 = without defect, 1 = with defect) into AttGAN (AttGAN-def). AttGAN, AttGAN-def, cGAN, and Unet were implemented using Tensorflow with the Adam optimizer running up to 400 epochs. FT and fast MP-SPECT projection pairs of 35 patients were used for training the networks for each acquisition time, while 5 and 10 patients were applied for validation and testing. Five-fold cross-validation was performed and data for all 50 patients were tested. Voxel-based error indices, joint histogram, linear regression, and perfusion defect size (PDS) were analyzed.</p>
</sec>
<sec>
<title>Results</title>
<p>All quantitative indices of AttGAN-based networks are superior to cGAN and Unet on all acquisition time images. AttGAN-def further improves AttGAN performance. The mean absolute error of PDS by AttcGAN-def was 1.60 on acquisition time of 1 s/prj, as compared to 2.36, 2.76, and 3.02 by AttGAN, cGAN, and Unet.</p>
</sec>
<sec>
<title>Conclusion</title>
<p>Denoising based on AttGAN is superior to conventional CNN-based networks for MP-SPECT.</p>
</sec>
</abstract>
<kwd-group>
<kwd>denoising</kwd>
<kwd>attention-guided</kwd>
<kwd>deep learning</kwd>
<kwd>myocardial perfusion</kwd>
<kwd>fast SPECT</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="1"/>
<equation-count count="6"/>
<ref-count count="32"/>
<page-count count="10"/>
<word-count count="5797"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<title>1. Introduction</title>
<p>Myocardial perfusion single photon emission computed tomography (MP-SPECT) is a standard method for the quantitative diagnosis of coronary artery disease (CAD) (<xref ref-type="bibr" rid="B1">1</xref>). However, the acquisition time for the conventional NaI-based MP-SPECT is relatively long (15-20 min) (<xref ref-type="bibr" rid="B2">2</xref>), leading to potential motion artifacts, e.g., upward creep (<xref ref-type="bibr" rid="B3">3</xref>), patient&#x2019;s discomfort, lower patient throughput (<xref ref-type="bibr" rid="B4">4</xref>), and mismatch artifacts between the sequential MP-SPECT and CT (<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B6">6</xref>). New scanner geometries with parallel-hole (<xref ref-type="bibr" rid="B7">7</xref>) or multi-pinhole collimations (<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B9">9</xref>) for MP-SPECT are proposed for better photons detection efficiency and reduced scan time (2-8 min) (<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>). Advanced reconstruction algorithms (<xref ref-type="bibr" rid="B12">12</xref>) also facilitate the possibility of reducing acquisition time without degrading image quality. However, the acquisition time for MP-SPECT is still much longer than CT in general (<xref ref-type="bibr" rid="B4">4</xref>). Therefore, it is necessary to pursue fast MP-SPECT, without compromising the image quality and diagnostic accuracy.</p>
<p>Image noise is a substantial problem for fast MP-SPECT due to the limited detected counts and the fact that it degrades the image quality, hampering clinical diagnosis and quantification results (<xref ref-type="bibr" rid="B9">9</xref>). Recently, deep learning (DL) methods are promising to reduce the noise for MP-SPECT images. Ramon et al. (<xref ref-type="bibr" rid="B13">13</xref>) proposed 3D convolutional neural networks (CNN) to denoise the reconstructed MP-SPECT images with reduced injected dose. Liu et al. used a 3D Unet trained on a noise-to-noise strategy for denoising full dose MP-SPECT reconstructed images and showed improved results as compared to the use of traditional filter (<xref ref-type="bibr" rid="B14">14</xref>). They further evaluated the performance of DL-based denoising according to the area under the curve (AUC) of the total perfusion deficit (TPD) scores results (<xref ref-type="bibr" rid="B15">15</xref>). Aghakhan et al. (<xref ref-type="bibr" rid="B16">16</xref>) used a 2D conditional generative adversarial network (cGAN) for denoising the reduced dose MP-SPECT images from 1/8 to 1/2 dose levels in the projection domain. Shiri et al. (<xref ref-type="bibr" rid="B17">17</xref>) proposed a 2D residual CNN (ResNet) to estimate full time (FT) MP-SPECT projection images. Previously, our group implemented a 3D cGAN to denoise fast and low dose MP-SPECT reconstruction (<xref ref-type="bibr" rid="B18">18</xref>) and projection (<xref ref-type="bibr" rid="B19">19</xref>) images. Our results showed that denoising on the projection domain is superior to the reconstruction domain (<xref ref-type="bibr" rid="B19">19</xref>).</p>
<p>However, conventional CNN-based methods use fixed-sized convolutional kernels to convolute one local region within the receptive field at a time, which would be ineffective for learning the feature dependencies across large regions (<xref ref-type="bibr" rid="B20">20</xref>). The feature dependencies across large regions can only be learned when the feature maps are down-sampled into a relatively small matrix size after passing through several convolutional layers (<xref ref-type="bibr" rid="B21">21</xref>). The attention mechanism has shown to be effective in capturing the long-range dependencies of structural information across large regions (<xref ref-type="bibr" rid="B20">20</xref>). It has been implemented for CT segmentation (<xref ref-type="bibr" rid="B22">22</xref>) and low dose CT denoising (<xref ref-type="bibr" rid="B23">23</xref>). In this study, we propose an attention-based cGAN (AttGAN) in denoising fast MP-SPECT projection images and compare its performance with Unet-based and cGAN-based denoising. We further incorporate the patient defect information into the network to improve the AttGAN performance.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>2. Materials and methods</title>
<sec id="S2.SS1">
<title>2.1. Patient dataset</title>
<p>Fifty anonymized patients who underwent routine stress SPECT/CT scan &#x223C;30 minutes post <sup>99m</sup>Tc-sestamibi injection on a clinical SPECT/CT system (NM/CT 870 CZT, GE Healthcare, USA) were retrospectively enrolled in this study under local ethics approval (IRB number 2022-11-002CC, <xref ref-type="table" rid="T1">Table 1</xref>). Among them, 18 were read as having at least a cardiac defect, which had perfusion abnormalities, according to their medical records from SPECT images and clinical histories. Before the SPECT acquisition, a helical CT scan (120 kVp, smart mA (10-150 mA), 0.375 cm slice thickness) was acquired in the heart region for attenuation correction in SPECT reconstruction. The CT reconstruction matrix size was 512 &#x00D7; 512 &#x00D7; variable axial coverage, with a voxel size of 0.9765 mm. Patients were injected with 1,184 MBq <sup>99m</sup>Tc-sestamibi, and 60 projections were acquired through 180&#x00B0; from the right anterior oblique to the left posterior oblique positions with a matrix size of 64 &#x00D7; 64. The primary photopeak energy window was centered at 140.5 keV with a 20% width and the scatter window was centered at 120 keV with a 10% width.</p>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Demographic information for the patient study.</p></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<td valign="top" align="left" style="color:#ffffff;background-color: #7f8080;"></td>
<td valign="top" align="center" style="color:#ffffff;background-color: #7f8080;">Female</td>
<td valign="top" align="center" style="color:#ffffff;background-color: #7f8080;">Male</td>
<td valign="top" align="center" style="color:#ffffff;background-color: #7f8080;">Total</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Gender</td>
<td valign="top" align="center">13 (26%)</td>
<td valign="top" align="center">37 (74%)</td>
<td valign="top" align="center">50 (100%)</td>
</tr>
<tr>
<td valign="top" align="left">Age (years)</td>
<td valign="top" align="center">69.2 &#x00B1; 9.73<break/>(56-90)</td>
<td valign="top" align="center">64.9 &#x00B1; 11.24<break/>(42-83)</td>
<td valign="top" align="center">66.0 &#x00B1; 10.94<break/>(42-90)</td>
</tr>
<tr>
<td valign="top" align="left">BMI (kg/m<sup>2</sup>)</td>
<td valign="top" align="center">24.5 &#x00B1; 3.13<break/>(21.09-30.47)</td>
<td valign="top" align="center">25.0 &#x00B1; 2.65<break/>(17.91-30.11)</td>
<td valign="top" align="center">25.0 &#x00B1; 2.92<break/>(17.92-31.60)</td>
</tr>
<tr>
<td valign="top" align="left">Perfusion defect size (PDS, %)</td>
<td valign="top" align="center">3.23 &#x00B1; 2.20<break/>(0-7)</td>
<td valign="top" align="center">4.46 &#x00B1; 5.57<break/>(0-29)</td>
<td valign="top" align="center">4.14 &#x00B1; 4.92<break/>(0-29)</td>
</tr>
<tr>
<td valign="top" align="left">Cardiac defect</td>
<td valign="top" align="center">2 (4%)</td>
<td valign="top" align="center">16 (32%)</td>
<td valign="top" align="center">18 (36%)</td>
</tr>
<tr>
<td valign="top" align="left" colspan="4" style="background-color: #dcdcdc;"><bold>CAD risk factors</bold></td>
</tr>
<tr>
<td valign="top" align="left">Hypertension</td>
<td valign="top" align="center">6 (12%)</td>
<td valign="top" align="center">20 (40%)</td>
<td valign="top" align="center">26 (52%)</td>
</tr>
<tr>
<td valign="top" align="left">Dyslipidaemia</td>
<td valign="top" align="center">8 (16%)</td>
<td valign="top" align="center">21 (42%)</td>
<td valign="top" align="center">29 (58%)</td>
</tr>
<tr>
<td valign="top" align="left">Diabetes</td>
<td valign="top" align="center">2 (4%)</td>
<td valign="top" align="center">14 (28%)</td>
<td valign="top" align="center">16 (32%)</td>
</tr>
<tr>
<td valign="top" align="left">Smoker</td>
<td valign="top" align="center">0 (0%)</td>
<td valign="top" align="center">8 (16%)</td>
<td valign="top" align="center">8 (16%)</td>
</tr>
<tr>
<td valign="top" align="left">Family history of CAD</td>
<td valign="top" align="center">4 (8%)</td>
<td valign="top" align="center">9 (18%)</td>
<td valign="top" align="center">13 (26%)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn><p>Mean &#x00B1; SD and range are presented for age, BMI, and PDS.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>The original acquisition time was 10 s/view. We also obtained various fast MP-SPECT projection images by reducing the projection acquisition time to be 7, 5, 3, 2, and 1 s based on the list mode data of the FT images, respectively. All clinical data were reconstructed by the 3D ordered subset expectation maximization (OS-EM) algorithm with 5 iterations and 4 subsets, with CT-based attenuation and dual energy window scatter corrections. The reconstruction matrix size was 64 &#x00D7; 64 &#x00D7; 19 with a voxel size of 0.6096 cm. A 3D post-reconstruction Gaussian filter with a standard deviation of 0.6 voxel was applied on the FT images for data analysis.</p>
</sec>
<sec id="S2.SS2">
<title>2.2. Attention-guided generative adversarial network (AttGAN)</title>
<p>The architecture of the AttGAN used in this study is shown in <xref ref-type="fig" rid="F1">Figure 1</xref> (<xref ref-type="bibr" rid="B20">20</xref>). Similar to cGAN, AttGAN is comprised of two subnetworks: a generator (<xref ref-type="fig" rid="F1">Figure 1A</xref>) and a discriminator (<xref ref-type="fig" rid="F1">Figure 1B</xref>). The generator, which was conditioned with fast MP-SPECT projection images, transformed the fast MP-SPECT images into estimated FT MP-SPECT projection images. The estimated images were later paired with the original fast MP-SPECT projection images as an estimated sample pair. The fast MP-SPECT projection images were also paired with the corresponding FT MP-SPECT projection images as a real sample pair. The discriminator learned to differentiate between the estimated sample pairs and the real sample pairs.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>The AttGAN architecture used in this study. <bold>(A)</bold> Generator; <bold>(B)</bold> discriminator; <bold>(C)</bold> attention module.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g001.tif"/>
</fig>
<p>The Unet-based generator (<xref ref-type="bibr" rid="B24">24</xref>) had subunits of encoding, bottleneck, and decoding layers. Each encoding layer was comprised of convolution (3 &#x00D7; 3 &#x00D7; 3), batch normalization (BN) (<xref ref-type="bibr" rid="B25">25</xref>), rectified linear unit (ReLU) activation, and dropout with a rate of 50%, followed by max-pooling to down-sample feature maps between layers. The decoding layers mirrored the encoding layers, except the up-sample layers replaced the down-sample layers, and skip connection between the encoding and decoding layers was added. The discriminator was a CNN-based network used in our previous study (<xref ref-type="bibr" rid="B19">19</xref>).</p>
<p>The attention modules, which were used for calculating the relationships of each voxel to all other pixels within a feature map, were incorporated together with the skip-connection between the encoding layers and decoding layers in the generator (<xref ref-type="fig" rid="F1">Figure 1C</xref>) (<xref ref-type="bibr" rid="B20">20</xref>). The feature maps <italic>x</italic> from the previous encoding layer <italic>g</italic> and current decoding layer <italic>f</italic> were transformed by a convolution (1 &#x00D7; 1 &#x00D7; 1) respectively, where <italic>g</italic>(<italic>x</italic>) = <italic>W</italic><sub><italic>g</italic></sub><italic>x</italic> and <italic>f</italic>(<italic>x</italic>) = <italic>W</italic><sub><italic>f</italic></sub><italic>x</italic>. <italic>W<sub>g</sub></italic> and <italic>W<sub>f</sub></italic> were trainable parameters. The feature maps of decoding layer <italic>f</italic>(<italic>x</italic>) were then down-sampled to be consistent with the size of <italic>g</italic>(<italic>x</italic>). We performed an inner product of the two vectors <italic>g</italic>(<italic>x</italic>)and <italic>f</italic>(<italic>x</italic>) to obtain the feature dependencies between every two voxels:</p>
<disp-formula id="S2.Ex1">
<mml:math display="block" id="M1"><mml:mrow><mml:mpadded width="+3.3pt"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mtext>j</mml:mtext></mml:mrow></mml:mrow></mml:msub></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:mrow><mml:mrow><mml:mtext>f</mml:mtext><mml:mo>&#x2062;</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mtext>x</mml:mtext><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mtext>T</mml:mtext></mml:mrow></mml:msup></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mtext>g</mml:mtext></mml:mrow><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mtext>x</mml:mtext><mml:mrow><mml:mtext>j</mml:mtext></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula>
<p>The &#x03B1;<sub>i,j</sub> further went through ReLU activation, convolution (1 &#x00D7; 1 &#x00D7; 1), and softmax function to normalize and reshape the feature maps to become <italic>r</italic><sub>i,j</sub>. Finally, we multiplied <italic>r</italic><sub>i,j</sub> with the feature maps <italic>x</italic> from the previous encoding layer <italic>g</italic> to obtain the attention coefficients <italic>Att</italic>:</p>
<disp-formula id="S2.Ex2">
<mml:math display="block" id="M2"><mml:mrow><mml:mpadded width="+3.3pt"><mml:mtext>Att</mml:mtext></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:munderover><mml:mo largeop="true" movablelimits="false" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mtext>N</mml:mtext></mml:mrow></mml:munderover><mml:mrow><mml:mpadded width="+3.3pt"><mml:msub><mml:mtext>x</mml:mtext><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow></mml:msub></mml:mpadded><mml:mo rspace="5.8pt">&#x22C5;</mml:mo><mml:msub><mml:mtext>r</mml:mtext><mml:mrow><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mtext>j</mml:mtext></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula>
<p>The <italic>L<sub>1</sub></italic> loss (<xref ref-type="bibr" rid="B26">26</xref>) and the adversarial loss <italic>L</italic><sub><italic>ADV</italic></sub> were used for training the generator <italic>g</italic>. The discriminator was trained by a cross-entropy loss <italic>L<sub>D</sub></italic> (<xref ref-type="bibr" rid="B19">19</xref>). The final objective function of AttGAN was:</p>
<disp-formula id="S2.Ex3">
<mml:math display="block" id="M3"><mml:mrow><mml:mpadded width="+3.3pt"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mtext>AttGAN</mml:mtext></mml:mrow></mml:msub></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:msub><mml:mtext>argmin</mml:mtext><mml:mrow><mml:mtext>G</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x2062;</mml:mo><mml:msub><mml:mtext>max</mml:mtext><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mtext>ADV</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mtext>G</mml:mtext><mml:mo>,</mml:mo><mml:mtext>D</mml:mtext><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mi>&#x03BB;</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msub><mml:mtext>L</mml:mtext><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mtext>G</mml:mtext><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula>
<p>where &#x03BB; is set to be 100 to adjust the weight of <italic>V<sub>L</sub></italic><sub>1</sub> (<italic>G</italic>) (<xref ref-type="bibr" rid="B26">26</xref>). The AttGAN was trained by minimizing the loss. The Unet and cGAN structures were the same according to our previous studies (<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B19">19</xref>).</p>
</sec>
<sec id="S2.SS3">
<title>2.3. Data preprocessing</title>
<p>All the intensity values of MP-SPECT projection images were normalized to a range of 0&#x2013;1 for training. In addition, we further incorporated the binary patient defect information, i.e., with (1) or without defect (0) from patients&#x2019; own medical records, by embedding four 64 &#x00D7; 64 slices with the same binary values into the projection images (<xref ref-type="fig" rid="F2">Figure 2</xref>).</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>The projection datasets used in this study: <bold>(A)</bold> original projection, <bold>(B)</bold> projection incorporating the patient defect information. The defect information block (red) was encoded by binary values (0 = without defect/1 = with defect).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g002.tif"/>
</fig>
</sec>
<sec id="S2.SS4">
<title>2.4. Network implementation</title>
<p>The AttGAN incorporating patient defect information (AttGAN-def), AttGAN, cGAN, and Unet were implemented using Tensorflow which ran on a NVIDIA GeForce RTX 2080Ti GPU. The Adam optimizer was applied to optimize this proposed model based on an initial learning rate of 0.0001 and trained to 400 epochs.</p>
<p>We performed a 5-fold cross-validation on the clinical datasets to evaluate four DL approaches for various fast SPECT acquisitions. Specifically, for each fold of evaluation, FT and fast SPECT projection images of 35, 5, and 10 patients were selected for training, validation, and testing, respectively. This process was repeated 5 times and all fifty patient datasets were tested and averaged for the final results. The denoised projections were further reconstructed using the same OS-EM algorithm with 5 iterations and 4 subsets with attenuation and scatter correction. No post-reconstruction filter was applied on the reconstructed images generated from the denoised projections for further analysis.</p>
<p>The hyper-parameters, e.g., number of layers and filters within each layer, were determined based on a training-validation procedure for AttGAN. Specifically, the number of layers varied as 2, 3, 4 and 5, while the number of filters within each layer varied as 8, 16, 24, 32 and 40. The hyper-parameters for cGAN and Unet were determined in our previous study (<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B19">19</xref>). The training time for AttGAN-def, AttGAN, cGAN, and Unet was 2.2, 2.2, 2.0, and 1.9 hr, respectively.</p>
</sec>
<sec id="S2.SS5">
<title>2.5. Data analysis</title>
<p>The voxel-based error of the denoised images was assessed by the normalized mean square error (NMSE), structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), joint histogram, and linear regression measured on a 3D volume-of-interest (VOI, 18 &#x00D7; 18 &#x00D7; 18, <xref ref-type="fig" rid="F3">Figure 3A</xref>) which covered the whole heart. The filtered FT reconstructed MP-SPECT images were used as the reference.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p><bold>(A)</bold> The 3D VOI used for voxel-based error calculation. <bold>(B)</bold> Sample ROIs drawn from a polar plot for IR calculation on a selected patient. (ROI<sub>1</sub> was for defect while ROI<sub>2</sub> was for a uniform normal region).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g003.tif"/>
</fig>
<disp-formula id="S2.Ex4">
<mml:math display="block" id="M4"><mml:mrow><mml:mpadded width="+3.3pt"><mml:mtext>NMSE</mml:mtext></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mo largeop="true" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mrow><mml:mtext>k</mml:mtext></mml:mrow></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mtext>N</mml:mtext></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mtext>I</mml:mtext><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mtext>I</mml:mtext><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:msubsup><mml:mo largeop="true" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mrow><mml:mtext>k</mml:mtext></mml:mrow></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mtext>N</mml:mtext></mml:mrow></mml:msubsup><mml:mmultiscripts><mml:mtext>I</mml:mtext><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow><mml:none/><mml:none/><mml:mn>2</mml:mn></mml:mmultiscripts></mml:mrow></mml:mfrac></mml:mrow></mml:math>
</disp-formula>
<disp-formula id="S2.Ex5">
<mml:math display="block" id="M5"><mml:mrow><mml:mpadded width="+3.3pt"><mml:mtext>SSIM</mml:mtext></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mfrac><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x2062;</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03BC;</mml:mi><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x2062;</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03BC;</mml:mi><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mtext>C</mml:mtext><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x2062;</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03C3;</mml:mi><mml:mrow><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mtext>C</mml:mtext><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi mathvariant="normal">&#x03BC;</mml:mi><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi mathvariant="normal">&#x03BC;</mml:mi><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mtext>C</mml:mtext><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mi mathvariant="normal">&#x03C3;</mml:mi><mml:mrow><mml:mtext>D</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi mathvariant="normal">&#x03C3;</mml:mi><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msub><mml:mtext>C</mml:mtext><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:math>
</disp-formula>
<disp-formula id="S2.Ex6">
<mml:math display="block" id="M6"><mml:mrow><mml:mpadded width="+3.3pt"><mml:mtext>PSNR</mml:mtext></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:mrow><mml:mn>10</mml:mn><mml:mo>&#x22C5;</mml:mo><mml:mpadded width="+3.3pt"><mml:msub><mml:mtext>log</mml:mtext><mml:mn>10</mml:mn></mml:msub></mml:mpadded></mml:mrow><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:msub><mml:mtext>MAX</mml:mtext><mml:mrow><mml:mtext>FT</mml:mtext></mml:mrow></mml:msub><mml:msqrt><mml:mtext>MSE</mml:mtext></mml:msqrt></mml:mfrac><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula>
<p>where <italic>I<sub>D</sub></italic> represents the voxel values in denoised reconstructed images, <italic>I</italic><sub><italic>FT</italic></sub> is the voxel values on the filtered FT reconstructed images, <italic>N</italic> (5832) is the number of voxels in the VOI, &#x03BC;<sub><italic>D</italic></sub> and &#x03BC;<sub><italic>FT</italic></sub> are the mean values of the denoised and reference images, &#x03C3;<sub><italic>D</italic></sub> and &#x03C3;<sub><italic>FT</italic></sub> are the standard deviations of the denoised and reference images respectively, and &#x03C3;<sub><italic>D</italic>,<italic>FT</italic></sub> is the cross-covariance between the two images. The constants <italic>C<sub>1</sub></italic> and <italic>C<sub>2</sub></italic> are set to be 0.01 and 0.02, respectively (<xref ref-type="bibr" rid="B17">17</xref>). <italic>MAX</italic><sub><italic>FT</italic></sub> indicates the maximum possible pixel value of the reference images while <italic>MSE</italic>indicates the mean squared error between the denoised and reference images.</p>
<p>Two regions-of-interest (ROI<sub>1</sub> and ROI<sub>2</sub>) were drawn on the defect region and a uniform normal region on the polar plots based on visual assessment (<xref ref-type="fig" rid="F3">Figure 3B</xref>) and were adjusted for each patient, respectively. The same ROIs were applied for all denoised images for the same patient. The intensity ratio (IR) was calculated from the mean value of the defect ROI (ROI<sub>1</sub>) divided by the mean value of the uniform ROI (ROI<sub>2</sub>). The absolute error of IR between FT images and different denoised images was computed.</p>
<p>A clinical relevant index, the perfusion defect size (PDS, %LV), i.e., an index similar to the total perfusion deficit, was measured by the Wackers-Liu CQ&#x2122; (WLCQ) software (Voxelon Inc, Watertown, CT) (<xref ref-type="bibr" rid="B27">27</xref>). The absolute error of PDS between FT and different denoised images was computed. The Bland&#x2013;Altman plots were also computed to quantify the agreement of PDS. For the statistical analysis, a two-tailed paired t-test with Bonferroni correction (SPSS, IBM Corporation, Armonk, NY, USA) was performed between AttGAN-def and other denoising methods at different acquisition time/view for NMSE, PSNR, SSIM, PDS, and IR. A p-value of less than 0.05 was considered as statistically significant.</p>
</sec>
</sec>
<sec id="S3" sec-type="results">
<title>3. Results</title>
<sec id="S3.SS1">
<title>3.1. Reconstructed images, polar plots, and 17-segment analysis</title>
<p><xref ref-type="fig" rid="F4">Figure 4</xref> shows the short axis fast MP-SPECT images, their corresponding difference images as compared with filtered FT SPECT images, polar plots as well as 17-segment plots processed using different DL denoising methods for a normal male patient. <xref ref-type="fig" rid="F5">Figure 5</xref> shows the same results for a male patient with a defect in the left anterior descending (LAD) and left circumflex artery (LCX) region. It can be observed that all the DL-denoised fast SPECT images are similar to the filtered FT SPECT images based on a visual assessment, with the noise level notably suppressed. Furthermore, it is noted that the proposed AttGAN methods have less bias than Unet and cGAN methods according to their corresponding images. Less bias is also observed from the 17-segment images for the AttGAN and AttGAN-def methods. The denoised images consistently exhibit worse resolution, i.e., more blurring, in shorter acquisition times.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Sample images of a normal male patient (age = 67, BMI = 25.4) before and after DL-based denoising for five shorter acquisition times. The images are shown in <bold>(A)</bold> short axis images, <bold>(B)</bold> difference images as compared to FT SPECT, <bold>(C)</bold> polar plots, and <bold>(D)</bold> 17-segment plots.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g004.tif"/>
</fig>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p>Sample images of another male patient (age = 77, BMI = 24.2) with an abnormal perfusion in the LAD and LCX territory before and after DL-based denoising for five shorter acquisition times. The images are shown in <bold>(A)</bold> short axis images, <bold>(B)</bold> difference images, <bold>(C)</bold> polar plots, and <bold>(D)</bold> 17-segment plots.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g005.tif"/>
</fig>
</sec>
<sec id="S3.SS2">
<title>3.2. Quantitative analysis</title>
<sec id="S3.SS2.SSS1">
<title>3.2.1. Physical and clinical indices</title>
<p><xref ref-type="fig" rid="F6">Figure 6</xref> summarizes the average quantitative indices on all 50 testing datasets for original fast SPECT and DL-based denoised images. For all indices, DL-based denoising methods improve the image quality as compared to the original fast SPECT images. AttGAN-def obtains the best performance, followed by the AttGAN, cGAN, and Unet. At 1 s/prj fast SPECT images, the NMSE values are 0.0317 &#x00B1; 0.007, 0.0337 &#x00B1; 0.008, 0.0363 &#x00B1; 0.011 and 0.0380 &#x00B1; 0.013 for AttGAN-def, AttGAN, cGAN, and Unet, respectively, where AttGAN-def has significant difference with the other three DL methods. Similar results are obtained for the PSNR and SSIM values. The absolute errors of IR are 0.0435 &#x00B1; 0.035, 0.0483 &#x00B1; 0.046, 0.0498 &#x00B1; 0.046 and 0.0632 &#x00B1; 0.047 for AttGAN-def, AttGAN, cGAN and Unet on 1 s/prj fast SPECT images. For the absolute error of PDS, the AttGAN-def yields the lowest difference value among all denoising methods on all noise levels. The denoised images achieve better PDS performance, i.e., 1.60 &#x00B1; 1.738, 2.36 &#x00B1; 1.903, 2.76 &#x00B1; 2.056 and 3.02 &#x00B1; 2.428 for AttGAN-def, AttGAN, cGAN, and Unet on 1 s/prj fast SPECT images, as compared to the original fast SPECT images. The AttGAN-def has significant difference with the cGAN and Unet while not significant difference with AttGAN on 1 s/prj fast SPECT images. GAN methods outperform Unet in general.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>Quantitative comparison of NMSE, PSNR, SSIM, PDS, and IR on 50 testing datasets for fast SPECT and different denoised images for five shorter acquisition times. Error bars indicate the standard deviation. The filtered FT SPECT images were used as reference. &#x002A;<italic>p</italic>&#x003C;0.05, &#x002A;&#x002A;<italic>p</italic>&#x003C;0.01, &#x002A;&#x002A;&#x002A;<italic>p</italic>&#x003C;0.001.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g006.tif"/>
</fig>
</sec>
<sec id="S3.SS2.SSS2">
<title>3.2.2. Bland&#x2013;Altman plots</title>
<p>The Bland&#x2013;Altman plots of PDS for different denoised methods and fast SPECT images are shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. The dashed lines denote the 95% confidence interval (CI) of the PDS. For 1 s/prj fast SPECT, the AttGAN-def method shows the smallest variance (95% CI: &#x2212;4.835, +4.435) compared to the reference filtered FT SPECT images, followed by the AttGAN (95% CI: &#x2212;6.383, +5.423), cGAN (95% CI: &#x2212;7.402, +5.802), and Unet (95% CI: &#x2212;8.535, +5.615) methods.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption><p>The Bland-Altman plots of PDS for fast SPECT different denoised images and acquisition times. The filtered FT SPECT images are used as reference.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g007.tif"/>
</fig>
</sec>
<sec id="S3.SS2.SSS3">
<title>3.2.3. Joint histogram and linear regression analysis</title>
<p>The voxel-based joint histogram and linear regression analysis results are shown in <xref ref-type="fig" rid="F8">Figure 8</xref>. Similar to other quantitative analysis, AttGAN-def obtains the best performance (R<sup>2</sup>= 0.8633), followed by the AttGAN (R<sup>2</sup>= 0.8433), cGAN (R<sup>2</sup>= 0.8341), and Unet (R<sup>2</sup>= 0.8123) on 1 s/prj fast SPECT denoised images.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption><p>Joint histogram and linear regression analysis of fast SPECT and different denoised images in the 50 testing datasets for five shorter acquisition times. The filtered FT images are used as the reference.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-10-1083413-g008.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<title>4. Discussion</title>
<p>This study aims to assess the potential of attention-based DL network on fast MP-SPECT using clinical datasets. The attention module, inspired by non-local means, was proposed to enable the remote voxels to contribute to the local receptive filter during convolutional filtering (<xref ref-type="bibr" rid="B28">28</xref>). To the best of our knowledge, we are the first group to propose using an attention-based DL network for denoising the MP-SPECT images. For the quantitative performance comparison, we used filtered FT MP-SPECT images as the reference since there was no ground truth. Results consistently showed that the AttGAN denoised images had better quantitative accuracy from the difference images, polar plots, 17-segment plots, various physical indices, joint correlation histogram, linear regression and the clinical PDS analysis, as compared to conventional cGAN and Unet (<xref ref-type="fig" rid="F4">Figures 4</xref>-<xref ref-type="fig" rid="F8">8</xref>).</p>
<p>Our proposed networks were trained and tested for each dose-specific dataset, respectively, which were acquired on a CZT scanner. The denoising process is similar to an image-to-image translation task, and should be applicable to data acquired from other scanners or other denoised tasks, e.g., low dose SPECT imaging. The difference between low dose and fast SPECT is that the former will be more subject to patient motion with less radiation dose delivered to the patients. This is particularly important for the increasingly young patient population which has higher radiation risk than seniors (<xref ref-type="bibr" rid="B29">29</xref>). In our fast SPECT study, the reduction of acquisition time would be beneficial for patients with a sedation demand and patients with less compliance. The reduced acquisition time would further increase the patient throughput. Both low dose and fast SPECT are of clinical interest and can be potentially achieved using DL techniques (<xref ref-type="bibr" rid="B17">17</xref>).</p>
<p>Shiri et al. (<xref ref-type="bibr" rid="B17">17</xref>) suggested that denoising from half acquisition time per projection outperformed that from half number of projections for fast MP-SPECT by 2D ResNet. According to this reference, our fast MP-SPECT datasets were obtained by reducing the acquisition time per view. Compared with the existing literature, Ramon et al. (<xref ref-type="bibr" rid="B13">13</xref>) reported that CNN-based denoising on 1/2 dose level images could achieve image quality comparable to standard full dose images based on the TPD values. Aghakhan et al. (<xref ref-type="bibr" rid="B16">16</xref>) claimed that all denoised images on 1/2 dose level would be clinically acceptable by using 2D cGAN in the projection domain. Similarly, in our work, the mean value of the absolute difference for PDS, a similar index to TPD between FT and AttGAN-def denoised images on 1/2 acquisition time of FT is 0.66%. Although our 3D AttGAN results could not be directly compared with the previous studies due to the use of different networks, datasets, imaging protocols, and evaluation methods, the results are consistent.</p>
<p>In addition, some studies have shown that concatenating the gender, BMI, state (stress or rest), and scatter window images to the training dataset can improve the DL performance in the attenuation correction task (<xref ref-type="bibr" rid="B30">30</xref>). We first proposed to add patient defect information into the projections for training the AttGAN and showed promising results. Incorporating the defect information into the network structure or loss function could be further investigated but it is beyond the scope of this study. One should note that the defect information was extracted from the patients&#x2019; medical reports and SPECT images in this study, which may be subjected to potential image artifacts. Moreover, the actual defect information may not be able to be verified as other examination results of the patients, e.g., CT angiography, are not available in this study.</p>
<p>Ramon et al. (<xref ref-type="bibr" rid="B13">13</xref>) compared the performance between training on a specific low dose level and training on a collection of various low dose levels at the same time. Their preliminary results showed that a dose-specific network can be more accurate than a &#x201C;one-size-fits-all&#x201D; network. Liu et al. (<xref ref-type="bibr" rid="B31">31</xref>) proposed a denoising method using an image noise index calculated from the normalized standard deviation in the liver ROI for low dose PET images. The image noise index was embedded as a tunable parameter for training. Their results demonstrated that their denoising method achieved better denoising performance than the &#x201C;one-size-fits-all&#x201D; network, while it still could not outperform the dose-specific network. Thus, we trained our DL model using a dose-specific approach, i.e., separately for different image acquisition times, for the best denoising performance. More investigations are warranted for a more generalizable and efficient training strategy, i.e., transfer learning, data preprocessing and adjustment on the loss function for training based on all available data.</p>
<p>Limitations of this study include that the clinical-related evaluation is only conducted by PDS. A more comprehensive clinical analysis and a ROC study of defect detectability are needed to validate the proposed methodology. Another limitation is that our DL networks were trained based on a relatively small patient cohort, i.e., thirty-five patients. Training on a large number of patients would benefit the model performance with less susceptibility of overfitting, though our network&#x2019;s loss function was validated to be converged. Gong et al. proposed to pre-train the DL network with simulation datasets and then fine-tuned it with a limited clinical dataset (<xref ref-type="bibr" rid="B32">32</xref>). Their results suggested that using simulation datasets with more realistic imaging conditions or with the use of a more accurate Monte Carlo simulation would generate a more robust pre-trained model.</p>
</sec>
<sec id="S5" sec-type="conclusion">
<title>5. Conclusion</title>
<p>In this work, we investigated the performance of AttGAN in denoising fast MP-SPECT images using clinical datasets. The proposed AttGAN provided superior denoising performance as compared to the conventional cGAN and Unet. Patient defect information could be useful parameter for further improving the AttGAN-based denoising performance.</p>
</sec>
<sec id="S6" sec-type="data-availability">
<title>Data availability statement</title>
<p>The original contributions presented in this study are included in this article/supplementary material, further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec id="S7" sec-type="ethics-statement">
<title>Ethics statement</title>
<p>The studies involving human participants were reviewed and approved by Institutional Review Board of Taipei Veterans General Hospital (IRB number 2022-11-002CC). The written informed consent was waived due to the retrospective nature of this study.</p>
</sec>
<sec id="S8" sec-type="author-contributions">
<title>Author contributions</title>
<p>JS, T-HW, B-HY, and GM were the primary authors of the manuscript. JS was mainly responsible for data processing and analysis. YD was mainly responsible for data analysis and manuscript revision. C-YL was mainly responsible for data collection. T-HW, B-HY, and Y-HL were responsible for clinical interpretation. GM and T-HW were responsible for data interpretation and study integration. All authors contributed to the article and approved the submitted version.</p>
</sec>
</body>
<back>
<sec id="S9" sec-type="funding-information">
<title>Funding</title>
<p>This work was supported by the Excellent Young Scientists Fund (Hong Kong and Macau, 81922080), Natural Science Foundation of China and Ministry of Science and Technology, Taiwan (MOST 109-2314-B-010-023-MY3).</p>
</sec>
<sec id="S10" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="S11" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><label>1.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Underwood</surname> <given-names>S</given-names></name> <name><surname>Anagnostopoulos</surname> <given-names>C</given-names></name> <name><surname>Cerqueira</surname> <given-names>M</given-names></name> <name><surname>Ell</surname> <given-names>P</given-names></name> <name><surname>Flint</surname> <given-names>E</given-names></name> <name><surname>Harbinson</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Myocardial perfusion scintigraphy: the evidence.</article-title> <source><italic>Eur J Nuclear Med Mol Imag.</italic></source> (<year>2004</year>) <volume>31</volume>:<fpage>261</fpage>&#x2013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.1007/s00259-003-1344-5</pub-id> <pub-id pub-id-type="pmid">15129710</pub-id></citation></ref>
<ref id="B2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Slomka</surname> <given-names>P</given-names></name> <name><surname>Patton</surname> <given-names>J</given-names></name> <name><surname>Berman</surname> <given-names>D</given-names></name> <name><surname>Germano</surname> <given-names>G</given-names></name></person-group>. <article-title>Advances in technical aspects of myocardial perfusion SPECT imaging.</article-title> <source><italic>J Nuclear Cardiol.</italic></source> (<year>2009</year>) <volume>16</volume>:<fpage>255</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-009-9052-6</pub-id></citation></ref>
<ref id="B3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>D</given-names></name> <name><surname>Pretorius</surname> <given-names>P</given-names></name> <name><surname>Ghaly</surname> <given-names>M</given-names></name> <name><surname>Zhang</surname> <given-names>Q</given-names></name> <name><surname>King</surname> <given-names>M</given-names></name> <name><surname>Mok</surname> <given-names>G</given-names></name></person-group>. <article-title>Evaluation of different respiratory gating schemes for cardiac SPECT.</article-title> <source><italic>J Nuclear Cardiol.</italic></source> (<year>2020</year>) <volume>27</volume>:<fpage>634</fpage>&#x2013;<lpage>47</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-018-1392-7</pub-id> <pub-id pub-id-type="pmid">30088195</pub-id></citation></ref>
<ref id="B4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J</given-names></name> <name><surname>Liu</surname> <given-names>C</given-names></name></person-group>. <article-title>Recent advances in cardiac SPECT instrumentation and imaging methods.</article-title> <source><italic>Phys Med Biol.</italic></source> (<year>2019</year>) <volume>64</volume>:<issue>06TR1</issue>. <pub-id pub-id-type="doi">10.1088/1361-6560/ab04de</pub-id> <pub-id pub-id-type="pmid">30726779</pub-id></citation></ref>
<ref id="B5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>D</given-names></name> <name><surname>Yang</surname> <given-names>B</given-names></name> <name><surname>Wu</surname> <given-names>N</given-names></name> <name><surname>Mok</surname> <given-names>G</given-names></name></person-group>. <article-title>Respiratory average CT for attenuation correction in myocardial perfusion SPECT/CT.</article-title> <source><italic>Ann Nuclear Med.</italic></source> (<year>2017</year>) <volume>31</volume>:<fpage>172</fpage>&#x2013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1007/s12149-016-1144-1</pub-id></citation></ref>
<ref id="B6"><label>6.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Q</given-names></name> <name><surname>Zhang</surname> <given-names>D</given-names></name> <name><surname>Mok</surname> <given-names>G</given-names></name></person-group>. <article-title>Comparison of different attenuation correction methods for dual gating myocardial perfusion SPECT/CT.</article-title> <source><italic>IEEE Trans Radiat Plasma Med Sci.</italic></source> (<year>2019</year>) <volume>3</volume>:<fpage>565</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1109/TRPMS.2019.2899066</pub-id></citation></ref>
<ref id="B7"><label>7.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kao</surname> <given-names>Y</given-names></name> <name><surname>Better</surname> <given-names>N</given-names></name></person-group>. <article-title>D-SPECT: new technology, old tricks.</article-title> <source><italic>J Nuclear Cardiol.</italic></source> (<year>2016</year>) <volume>23</volume>:<fpage>311</fpage>&#x2013;<lpage>2</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-015-0290-5</pub-id> <pub-id pub-id-type="pmid">26407951</pub-id></citation></ref>
<ref id="B8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ozsahin</surname> <given-names>I</given-names></name> <name><surname>Chen</surname> <given-names>L</given-names></name> <name><surname>K&#x00F6;nik</surname> <given-names>A</given-names></name> <name><surname>King</surname> <given-names>M</given-names></name> <name><surname>Beekman</surname> <given-names>F</given-names></name> <name><surname>Mok</surname> <given-names>G</given-names></name></person-group>. <article-title>The clinical utilities of multi-pinhole single photon emission computed tomography.</article-title> <source><italic>Quant Imag Med Surg.</italic></source> (<year>2020</year>) <volume>10</volume>:<fpage>2006</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.21037/qims-19-1036</pub-id> <pub-id pub-id-type="pmid">33014732</pub-id></citation></ref>
<ref id="B9"><label>9.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bocher</surname> <given-names>M</given-names></name> <name><surname>Blevis</surname> <given-names>I</given-names></name> <name><surname>Tsukerman</surname> <given-names>L</given-names></name> <name><surname>Shrem</surname> <given-names>Y</given-names></name> <name><surname>Kovalski</surname> <given-names>G</given-names></name> <name><surname>Volokh</surname> <given-names>L</given-names></name></person-group>. <article-title>A fast cardiac gamma camera with dynamic SPECT capabilities: design, system validation and future potential.</article-title> <source><italic>Eur J Nuclear Med Mol Imag.</italic></source> (<year>2010</year>) <volume>37</volume>:<fpage>1887</fpage>&#x2013;<lpage>902</lpage>. <pub-id pub-id-type="doi">10.1007/s00259-010-1488-z</pub-id> <pub-id pub-id-type="pmid">20585775</pub-id></citation></ref>
<ref id="B10"><label>10.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brambilla</surname> <given-names>M</given-names></name> <name><surname>Lecchi</surname> <given-names>M</given-names></name> <name><surname>Matheoud</surname> <given-names>R</given-names></name> <name><surname>Leva</surname> <given-names>L</given-names></name> <name><surname>Lucignani</surname> <given-names>G</given-names></name> <name><surname>Marcassa</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Comparative analysis of iterative reconstruction algorithms with resolution recovery and new solid state cameras dedicated to myocardial perfusion imaging.</article-title> <source><italic>Phys Med.</italic></source> (<year>2017</year>) <volume>41</volume>:<fpage>109</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1016/j.ejmp.2017.03.008</pub-id> <pub-id pub-id-type="pmid">28343906</pub-id></citation></ref>
<ref id="B11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Johnson</surname> <given-names>R</given-names></name> <name><surname>Bath</surname> <given-names>N</given-names></name> <name><surname>Rinker</surname> <given-names>J</given-names></name> <name><surname>Fong</surname> <given-names>S</given-names></name> <name><surname>James</surname> <given-names>S</given-names></name> <name><surname>Pampaloni</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Introduction to the D-SPECT for technologists: workflow using a dedicated digital cardiac camera.</article-title> <source><italic>J Nuclear Med Technol.</italic></source> (<year>2020</year>) <volume>48</volume>:<fpage>297</fpage>&#x2013;<lpage>303</lpage>. <pub-id pub-id-type="doi">10.2967/jnmt.120.254870</pub-id> <pub-id pub-id-type="pmid">33020236</pub-id></citation></ref>
<ref id="B12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Valenta</surname> <given-names>I</given-names></name> <name><surname>Treyer</surname> <given-names>V</given-names></name> <name><surname>Husmann</surname> <given-names>L</given-names></name> <name><surname>Gaemperli</surname> <given-names>O</given-names></name> <name><surname>Schindler</surname> <given-names>M</given-names></name> <name><surname>Herzog</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>New reconstruction algorithm allows shortened acquisition time for myocardial perfusion SPECT.</article-title> <source><italic>Eur J Nucl Med Mol Imag.</italic></source> (<year>2010</year>) <volume>37</volume>:<fpage>750</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1007/s00259-009-1300-0</pub-id> <pub-id pub-id-type="pmid">19921186</pub-id></citation></ref>
<ref id="B13"><label>13.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramon</surname> <given-names>A</given-names></name> <name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Pretorius</surname> <given-names>P</given-names></name> <name><surname>Johnson</surname> <given-names>K</given-names></name> <name><surname>King</surname> <given-names>M</given-names></name> <name><surname>Wernick</surname> <given-names>M</given-names></name></person-group>. <article-title>Improving diagnostic accuracy in low-dose SPECT myocardial perfusion imaging with convolutional denoising networks.</article-title> <source><italic>IEEE Trans Med Imag.</italic></source> (<year>2020</year>) <volume>39</volume>:<fpage>2893</fpage>&#x2013;<lpage>903</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2020.2979940</pub-id> <pub-id pub-id-type="pmid">32167887</pub-id></citation></ref>
<ref id="B14"><label>14.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>J</given-names></name> <name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Wernick</surname> <given-names>M</given-names></name> <name><surname>Pretorius</surname> <given-names>P</given-names></name> <name><surname>King</surname> <given-names>M</given-names></name></person-group>. <article-title>Deep learning with noise-to-noise training for denoising in SPECT myocardial perfusion imaging.</article-title> <source><italic>Med Phys.</italic></source> (<year>2021</year>) <volume>48</volume>:<fpage>156</fpage>&#x2013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.1002/mp.14577</pub-id> <pub-id pub-id-type="pmid">33145782</pub-id></citation></ref>
<ref id="B15"><label>15.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>J</given-names></name> <name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Wernick</surname> <given-names>M</given-names></name> <name><surname>Pretorius</surname> <given-names>P</given-names></name> <name><surname>Slomka</surname> <given-names>P</given-names></name> <name><surname>King</surname> <given-names>M</given-names></name></person-group>. <article-title>Improving detection accuracy of perfusion defect in standard dose SPECT-myocardial perfusion imaging by deep-learning denoising.</article-title> <source><italic>J Nucl Cardiol.</italic></source> (<year>2021</year>) <volume>2021</volume>:<fpage>1</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-021-02676-w</pub-id> <pub-id pub-id-type="pmid">34282538</pub-id></citation></ref>
<ref id="B16"><label>16.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aghakhan Olia</surname> <given-names>N</given-names></name> <name><surname>Kamali-Asl</surname> <given-names>A</given-names></name> <name><surname>Hariri Tabrizi</surname> <given-names>S</given-names></name> <name><surname>Geramifar</surname> <given-names>P</given-names></name> <name><surname>Sheikhzadeh</surname> <given-names>P</given-names></name> <name><surname>Farzanefar</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Deep learning&#x2013;based denoising of low-dose SPECT myocardial perfusion images: quantitative assessment and clinical performance.</article-title> <source><italic>Eur J Nucl Med Mol Imag.</italic></source> (<year>2022</year>) <volume>49</volume>:<fpage>1508</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1007/s00259-021-05614-7</pub-id> <pub-id pub-id-type="pmid">34778929</pub-id></citation></ref>
<ref id="B17"><label>17.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shiri</surname> <given-names>I</given-names></name> <name><surname>Sabet</surname> <given-names>K</given-names></name> <name><surname>Arabi</surname> <given-names>H</given-names></name> <name><surname>Pourkeshavarz</surname> <given-names>M</given-names></name> <name><surname>Teimourian</surname> <given-names>B</given-names></name> <name><surname>Ay</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Standard SPECT myocardial perfusion estimation from half-time acquisitions using deep convolutional residual neural networks.</article-title> <source><italic>J Nucl Cardiol.</italic></source> (<year>2021</year>) <volume>28</volume>:<fpage>2761</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-020-02119-y</pub-id> <pub-id pub-id-type="pmid">32347527</pub-id></citation></ref>
<ref id="B18"><label>18.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>J</given-names></name> <name><surname>Du</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>C</given-names></name> <name><surname>Wu</surname> <given-names>T</given-names></name> <name><surname>Yang</surname> <given-names>B</given-names></name> <name><surname>Mok</surname> <given-names>G</given-names></name></person-group>. <article-title>Pix2Pix generative adversarial network for low dose myocardial perfusion SPECT denoising.</article-title> <source><italic>Quant Imag Med Surg.</italic></source> (<year>2022</year>) <volume>12</volume>:<fpage>3539</fpage>&#x2013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.21037/qims-21-1042</pub-id> <pub-id pub-id-type="pmid">35782241</pub-id></citation></ref>
<ref id="B19"><label>19.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>J</given-names></name> <name><surname>Jiang</surname> <given-names>H</given-names></name> <name><surname>Du</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>C</given-names></name> <name><surname>Wu</surname> <given-names>T</given-names></name> <name><surname>Yang</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Deep learning-based denoising in projection-domain and reconstruction-domain for low dose myocardial perfusion SPECT.</article-title> <source><italic>J Nucl Cardiol.</italic></source> (<year>2022</year>) <volume>2022</volume>:<fpage>1</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-022-03045-x</pub-id> <pub-id pub-id-type="pmid">35982208</pub-id></citation></ref>
<ref id="B20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>H</given-names></name> <name><surname>Goodfellow</surname> <given-names>I</given-names></name> <name><surname>Metaxas</surname> <given-names>D</given-names></name> <name><surname>Odena</surname> <given-names>A</given-names></name></person-group>. <article-title>Self-attention generative adversarial networks.</article-title> <source><italic>Proceeding of the international conference on machine learning (ICML).</italic></source> <publisher-name>PMLR</publisher-name> (<year>2019</year>). <pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.105828</pub-id> <pub-id pub-id-type="pmid">35816855</pub-id></citation></ref>
<ref id="B21"><label>21.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schlemper</surname> <given-names>J</given-names></name> <name><surname>Oktay</surname> <given-names>O</given-names></name> <name><surname>Schaap</surname> <given-names>M</given-names></name> <name><surname>Heinrich</surname> <given-names>M</given-names></name> <name><surname>Kainz</surname> <given-names>B</given-names></name> <name><surname>Glocker</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Attention gated networks: learning to leverage salient regions in medical images.</article-title> <source><italic>Med Image Analy.</italic></source> (<year>2019</year>) <volume>53</volume>:<fpage>197</fpage>&#x2013;<lpage>207</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2019.01.012</pub-id> <pub-id pub-id-type="pmid">30802813</pub-id></citation></ref>
<ref id="B22"><label>22.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Lei</surname> <given-names>Y</given-names></name> <name><surname>Fu</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>T</given-names></name> <name><surname>Tang</surname> <given-names>X</given-names></name> <name><surname>Jiang</surname> <given-names>X</given-names></name><etal/></person-group> <article-title>CT-based multi-organ segmentation using a 3D self-attention U-net network for pancreatic radiotherapy.</article-title> <source><italic>Med Phys.</italic></source> (<year>2020</year>) <volume>47</volume>:<fpage>4316</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1002/mp.14386</pub-id> <pub-id pub-id-type="pmid">32654153</pub-id></citation></ref>
<ref id="B23"><label>23.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M</given-names></name> <name><surname>Hsu</surname> <given-names>W</given-names></name> <name><surname>Xie</surname> <given-names>X</given-names></name> <name><surname>Cong</surname> <given-names>J</given-names></name> <name><surname>Gao</surname> <given-names>W</given-names></name></person-group>. <article-title>SACNN: self-attention convolutional neural network for low-dose CT denoising with self-supervised perceptual loss network.</article-title> <source><italic>IEEE Trans Med Imag.</italic></source> (<year>2020</year>) <volume>39</volume>:<fpage>2289</fpage>&#x2013;<lpage>301</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2020.2968472</pub-id> <pub-id pub-id-type="pmid">31985412</pub-id></citation></ref>
<ref id="B24"><label>24.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Yu</surname> <given-names>B</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Zu</surname> <given-names>C</given-names></name> <name><surname>Lalush</surname> <given-names>D</given-names></name> <name><surname>Lin</surname> <given-names>W</given-names></name><etal/></person-group> <article-title>3D conditional generative adversarial networks for high-quality PET image estimation at low dose.</article-title> <source><italic>Neuroimage.</italic></source> (<year>2018</year>) <volume>174</volume>:<fpage>550</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2018.03.045</pub-id> <pub-id pub-id-type="pmid">29571715</pub-id></citation></ref>
<ref id="B25"><label>25.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ioffe</surname> <given-names>S</given-names></name> <name><surname>Szegedy</surname> <given-names>C</given-names></name></person-group>. <article-title>Batch normalization: accelerating deep network training by reducing internal covariate shift.</article-title> <source><italic>Proceeding of the international conference on machine learning (ICML).</italic></source> <publisher-name>PMLR</publisher-name> (<year>2015</year>). <pub-id pub-id-type="doi">10.1007/s11390-020-0679-8</pub-id> <pub-id pub-id-type="pmid">35496726</pub-id></citation></ref>
<ref id="B26"><label>26.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Isola</surname> <given-names>P</given-names></name> <name><surname>Zhu</surname> <given-names>J</given-names></name> <name><surname>Zhou</surname> <given-names>T</given-names></name> <name><surname>Efros</surname> <given-names>A</given-names></name></person-group>. <article-title>Image-to-image translation with conditional adversarial networks.</article-title> <source><italic>Proceeding of the IEEE conference on computer vision and pattern recognition.</italic></source> (<year>2017</year>). <pub-id pub-id-type="doi">10.1109/CVPR.2017.632</pub-id></citation></ref>
<ref id="B27"><label>27.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y</given-names></name></person-group>. <article-title>Quantification of nuclear cardiac images: the yale approach.</article-title> <source><italic>J Nucl Cardiol.</italic></source> (<year>2007</year>) <volume>14</volume>:<fpage>483</fpage>&#x2013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.1016/j.nuclcard.2007.06.005</pub-id> <pub-id pub-id-type="pmid">17679055</pub-id></citation></ref>
<ref id="B28"><label>28.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Buades</surname> <given-names>A</given-names></name> <name><surname>Coll</surname> <given-names>B</given-names></name> <name><surname>Morel</surname> <given-names>J</given-names></name></person-group>. <article-title>A non-local algorithm for image denoising.</article-title> <source><italic>Proceeding of the IEEE computer society conference on computer vision and pattern recognition (CVPR).</italic></source> <publisher-name>IEEE</publisher-name> (<year>2005</year>).</citation></ref>
<ref id="B29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gelfand</surname> <given-names>M</given-names></name> <name><surname>Lemen</surname> <given-names>L.</given-names></name></person-group> <source><italic>PET/CT and SPECT/CT dosimetry in children: the challenge to the pediatric imager seminars in nuclear medicine.</italic></source> <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Elsevier</publisher-name> (<year>2007</year>).</citation></ref>
<ref id="B30"><label>30.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>X</given-names></name> <name><surname>Zhou</surname> <given-names>B</given-names></name> <name><surname>Shi</surname> <given-names>L</given-names></name> <name><surname>Liu</surname> <given-names>H</given-names></name> <name><surname>Pang</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>CT-free attenuation correction for dedicated cardiac SPECT using a 3D dual squeeze-and-excitation residual dense network.</article-title> <source><italic>J Nucl Cardiol.</italic></source> (<year>2021</year>) <volume>2021</volume>:<fpage>1</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1007/s12350-021-02672-0</pub-id> <pub-id pub-id-type="pmid">34085168</pub-id></citation></ref>
<ref id="B31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Q</given-names></name> <name><surname>Liu</surname> <given-names>H</given-names></name> <name><surname>Mirian</surname> <given-names>N</given-names></name> <name><surname>Ren</surname> <given-names>S</given-names></name> <name><surname>Viswanath</surname> <given-names>V</given-names></name> <name><surname>Karp</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>A personalized deep learning denoising strategy for low-count PET images.</article-title> <source><italic>Phys Med Biol.</italic></source> (<year>2022</year>) <volume>67</volume>:<issue>145014</issue>. <pub-id pub-id-type="doi">10.1088/1361-6560/ac783d</pub-id> <pub-id pub-id-type="pmid">35697017</pub-id></citation></ref>
<ref id="B32"><label>32.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gong</surname> <given-names>K</given-names></name> <name><surname>Guan</surname> <given-names>J</given-names></name> <name><surname>Liu</surname> <given-names>C</given-names></name> <name><surname>Qi</surname> <given-names>J</given-names></name></person-group>. <article-title>PET image denoising using a deep neural network through fine tuning.</article-title> <source><italic>IEEE Trans Radiat Plasma Med Sci.</italic></source> (<year>2018</year>) <volume>3</volume>:<fpage>153</fpage>&#x2013;<lpage>61</lpage>.</citation></ref>
</ref-list>
</back>
</article>