<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Phys.</journal-id>
<journal-title>Frontiers in Physics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Phys.</abbrev-journal-title>
<issn pub-type="epub">2296-424X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1477253</article-id>
<article-id pub-id-type="doi">10.3389/fphy.2024.1477253</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Physics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A novel local contrast method based on third-order central moments for infrared small target detection</article-title>
<alt-title alt-title-type="left-running-head">Xu et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fphy.2024.1477253">10.3389/fphy.2024.1477253</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Xu</surname>
<given-names>Danyang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2815420/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Pang</surname>
<given-names>Dongdong</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Gou</surname>
<given-names>Shimiao</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2810545/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Zhaoyu</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2351791/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhuo</surname>
<given-names>Zhihai</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Information Engineering</institution>, <institution>Henan University of Science and Technology</institution>, <addr-line>Luoyang</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution> School of Information and Communication Engineering</institution>, <institution>Beijing Information Science and Technology University</institution>, <addr-line>Beijing</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Henan Key Laboratory of General Aviation Technology</institution>, <institution>School of Electronics and Information</institution>, <institution>Zhengzhou University of Aeronautics</institution>, <addr-line>Zhengzhou</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1184687/overview">Rajib Biswas</ext-link>, Tezpur University, India</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2045963/overview">Yongling Ren</ext-link>, University of Western Australia, Australia</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1894508/overview">Peng Zhenming</ext-link>, University of Electronic Science and Technology of China, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Dongdong Pang, <email>dongdong.pang@foxmail.com</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>10</day>
<month>04</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>12</volume>
<elocation-id>1477253</elocation-id>
<history>
<date date-type="received">
<day>07</day>
<month>08</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>22</day>
<month>10</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Xu, Pang, Gou, Liu and Zhuo.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Xu, Pang, Gou, Liu and Zhuo</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>As local contrast mechanisms are extensively utilized in infrared small target detection. However, the performance of existing local contrast-based methods is often compromised in complex backgrounds. This study presents a novel local contrast method based on third-order central moments to address the above challenges. Initially, the infrared image undergoes top-hat transformation to mitigate most background clutter and highlight potential target pixels. Then, a local contrast description operator based on third-order central moments is defined to characterize the grayscale changes in different regions of the preprocessed image, enhance the target and suppress the background. Finally, the target is extracted by using an adaptive threshold segmentation operation. The experimental results in six real-life scenarios demonstrate that the proposed method occupies the best detection index compared to other similar technologies.</p>
</abstract>
<kwd-group>
<kwd>small infrared (IR) target detection</kwd>
<kwd>local contrast</kwd>
<kwd>third-order central moments</kwd>
<kwd>image sequence</kwd>
<kwd>low-altitude moving background</kwd>
</kwd-group>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content>
</contract-sponsor>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Optics and Photonics</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Target detection is a pivotal technology in infrared search and track (IRST) systems [<xref ref-type="bibr" rid="B1">1</xref>]. In practical applications such as guidance, early warning, and surveillance/monitoring, whether airborne, space-borne or in anti-UAV (unmanned aerial vehicle) operations, which have garnered significant attention recently, detecting targets presents substantial challenges. These challenges stem from the inherent characteristics of the targets, which typically exhibit weak brightness and small size, while lacking unique shape, texture, and color information. Consequently, rapid and robust detection of small infrared targets has always been an unresolved issue in the field of object detection.</p>
<p>Currently, the relevant infrared small target detection technologies are mainly classified into two types: sequential detection approaches and single-frame detection approaches. Sequential detection technology typically capitalizes on the continuous motion of the target contrasted against the random motion of noise in sequential imagery. This type of technology facilitates small target detection and random noise elimination by identifying potential motion trajectories, albeit requiring substantial prior information. Conversely, the single-frame detection method relies on the characteristic feature information of dim and small targets within a single image to accomplish detection. This type of technology offers several advantages over sequential detection, including simpler computations, faster processing speeds, higher real-time performance, and broader applicability.</p>
<p>The current single-frame detection technology mainly includes the following two categories: non-local information-based methods and local information-based methods. Non-local information methods distinguish the target image from the entire original image by leveraging differences in frequency bands and data spaces between the target and background/noise images. These methods include frequency domain techniques, classifier methods, overcomplete sparse representation, and sparse low-rank decomposition. To illustrate, Gregoris et al. [<xref ref-type="bibr" rid="B2">2</xref>] integrated wavelet transform into infrared small target detection. Wu et al. [<xref ref-type="bibr" rid="B3">3</xref>] developed a method based on the contour transform that suppresses background frequencies while enhancing those of the target within the transform domain to bolster detection capabilities. Kong et al. [<xref ref-type="bibr" rid="B4">4</xref>] utilized diagonal detail information from Haar wavelet decomposition to aid in detecting weak infrared targets against sea-sky backgrounds. Wang et al. [<xref ref-type="bibr" rid="B5">5</xref>] applied dual-tree complex wavelets for decomposing the original image and employed a top-hat operator to filter the low-frequency sub-bands, subsequently using local image entropy to weight the reconstructed image for extracting infrared small targets. Zheng et al. [<xref ref-type="bibr" rid="B6">6</xref>] initially employed background estimation and differencing to pinpoint suspicious pixels, followed by clustering analysis to identify potential small targets. Dong et al. [<xref ref-type="bibr" rid="B7">7</xref>] extracted interest points from the image and introduced a novel R-mean clustering method to categorize these points into targets and backgrounds based on their associative patterns. Wu et al. [<xref ref-type="bibr" rid="B8">8</xref>] utilized support vector machines to identify the optimal hyperplane in kernel space that segregates targets from backgrounds, thus classifying pixels into these categories. Jiang et al. [<xref ref-type="bibr" rid="B9">9</xref>] merged the correlation between the observation matrix and the sparse matrix with the concept of gradient descent to devise an adaptive gradient descent method. Gao et al. [<xref ref-type="bibr" rid="B10">10</xref>] posited in their IPI model that the background in infrared images constitutes a low-rank component and the target a sparse component, transforming the detection of infrared small targets into a recovery problem for low-rank and sparse matrices, followed by decomposing image blocks to more effectively separate targets from backgrounds. Building on this framework, Zhang et al. [<xref ref-type="bibr" rid="B11">11</xref>] proposed a method that involves partial tensor nuclear norm sums and incorporates target edge information into the model to enhance its ability to suppress edge clutter.</p>
<p>The local information class method posits that the grey level of the background pixel point in an infrared image is typically similar to that of its local surrounding pixel points, whereas the target pixel point exhibits a more pronounced disparity in its grey level relative to its peripheral pixel points. By extracting the difference information between each pixel point in the image and its neighborhood reference pixel point, the target can be successfully filtered. For instance, Shao et al. [<xref ref-type="bibr" rid="B12">12</xref>] utilized a Laplacian of Gaussian (LoG) filter template characterized by a positive central coefficient and negative surrounding coefficients. During the convolution filtering of the original image with this template, the differences between each pixel and its neighbors are accentuated. From the standpoint of local image segmentation, Yao et al. [<xref ref-type="bibr" rid="B13">13</xref>] developed a small target detection model based on facet kernel and random walker (FKRW), which effectively mitigates edge noise. Chen et al. [<xref ref-type="bibr" rid="B14">14</xref>] introduced the Local Contrast Measure (LCM) method to enhance small target detection performance, although it suffered from a pronounced &#x201c;blocking effect.&#x201d; Building on LCM, a series of improved measures were subsequently proposed. Such as, Han et al. [<xref ref-type="bibr" rid="B15">15</xref>] introduced an Improved LCM (ILCM) method, incorporating the average value of sub-blocks as a parameter to better suppress random point noise, yet it tended to smooth out small targets when they were diminutive. Qin et al. [<xref ref-type="bibr" rid="B16">16</xref>] proposed a novel LCM (NLCM), that averages only the largest number of pixels in each sub-block, thus better addressing the issue of small targets being smoothed while still retaining noise suppression capabilities. Han et al. [<xref ref-type="bibr" rid="B17">17</xref>] initially defined Refined LCM (RLCM) and made significant enhancements in the design of the filtering template [<xref ref-type="bibr" rid="B18">18</xref>], the selection of background references [<xref ref-type="bibr" rid="B19">19</xref>], and the introduction of a weighting function [<xref ref-type="bibr" rid="B20">20</xref>]. This concept has been widely recognized and adopted in the field. For example, Wei et al. [<xref ref-type="bibr" rid="B21">21</xref>] proposed a target enhancement method based on multi-scale patch contrast measurement (MPCM) for infrared small target detection, although this method struggled with retaining target shape and edge information and resulted in numerous false alarms. Fu et al. [<xref ref-type="bibr" rid="B22">22</xref>] combined an adaptive filter with a probabilistic Hough transform to enhance local contrast, effectively distinguishing targets from backgrounds and accelerating the target detection process, albeit with reduced accuracy in highly complex backgrounds. Notably, recent studies have shown a trend where many researchers are combining local contrast with other types of algorithms to achieve superior detection results. Such as, Cui et al. [<xref ref-type="bibr" rid="B23">23</xref>, <xref ref-type="bibr" rid="B24">24</xref>] integrated local contrast with support vector machines. Deng et al. [<xref ref-type="bibr" rid="B25">25</xref>&#x2013;<xref ref-type="bibr" rid="B27">27</xref>] applied information entropy to weight local contrast, thoroughly analyzing the shape and size of the local information entropy window. Chen et al. [<xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B29">29</xref>] utilized the local signal-to-clutter ratio (SCR) to weight local contrast or combined the local contrast with frequency domain concepts. Du et al. [<xref ref-type="bibr" rid="B30">30</xref>] employed local smoothness to weight local contrast. Xiong et al. [<xref ref-type="bibr" rid="B31">31</xref>] initially calculated the local gradient of the original image and then assessed the local contrast of the gradient map. Han et al. [<xref ref-type="bibr" rid="B32">32</xref>] merged local contrast with TDLMS adaptive background estimation. Additionally, Dai et al. [<xref ref-type="bibr" rid="B33">33</xref>] integrated both local and non-local prior information to propose the RIPT model, which effectively suppresses interference factors and enhances the accuracy of target detection in specific scenarios. However, in complex environments, this method is prone to interference from background elements, resulting in lower detection precision. Pang et al. [<xref ref-type="bibr" rid="B34">34</xref>] proposed a low-rank and sparse decomposition method based on greedy bilateral decomposition for infrared dim and small target detection. This method can detect the target quickly and stably in complex scenes with a low signal-to-noise ratio, but the detection effect performs unsatisfactorily in the background with significant changes between different frames.</p>
<p>In general, local information class algorithms focus on a limited number of pixel points within a local area when calculating each pixel point. This results in a relatively small computational volume, which may have the potential for real-time processing when engineering optimisation techniques such as parallel acceleration and pipelined architecture are employed. The local contrast algorithm is a relatively simple and straightforward approach that aligns well with the infrared image model. By employing a carefully designed contrast formula, it is possible to enhance the visibility of a target while simultaneously suppressing complex background noise. However, the effectiveness of this method hinges on the premise that the target must be the most prominent locally. In practice, this may not always be the case, particularly in scenarios with highly complex backgrounds. In a real scene, if the background is highly complex, the target may be in close proximity to an extremely bright background, which may overwhelm it. This makes it challenging to detect the target using local information, which in turn leads to a degradation in the detection performance of local information-based algorithms in complex real backgrounds. The non-local information class makes full use of all the information in the frame to detect weak targets, irrespective of the prominence of the target in the local region. Non-local information algorithms leverage all available information within the frame to detect weak targets, independent of the target&#x2019;s prominence within the local region. Even when a target becomes less discernible due to its proximity to a highlighted background, it can still be successfully separated, presenting a significant advantage over local information-based classification algorithms. However, many non-local information-based algorithms proposed to date exhibit certain shortcomings that warrant further research and improvement. For instance, frequency domain methods posit that the frequency bands occupied by the target and background are distinct. However, in scenarios with more complex backgrounds, these frequency bands often overlap, complicating accurate differentiation in the frequency domain. Furthermore, most classifier methods require a substantial number of training samples, which can be challenging to obtain in the field of infrared weak target detection, especially in the presence of non-cooperative targets. The efficacy of the hyper-complete sparse representation method relies on the accuracy of the ultra-complete dictionary. Nevertheless, constructing an ultra-complete dictionary that encompasses all potential scenarios is impractical in practice. The sparse low-rank decomposition method assumes that the target is sparse while the background is low-rank. However, in complex backgrounds, sparse information may also be present at the edges of the background and at noise locations, which is susceptible to generating false alarms.</p>
<p>Due to the presence of various types of noise and complex background interference, the aforementioned methods are likely to result in false positives and missed detections. In practical applications, it is necessary to adjust and optimize these methods based on specific scenarios and data characteristics. Therefore, to address the issue of detecting small infrared targets in complex backgrounds, this paper proposes a novel target detection method based on the contrast of local third-order central moments. Specifically, the innovative aspects of the proposed method are as follows:<list list-type="simple">
<list-item>
<p>1. This study introduces the third-order central moment to characterize the fluctuation properties of image gray levels in different regions for the first time.</p>
</list-item>
<list-item>
<p>2. Utilizing the gray level fluctuation properties across different image regions, a local contrast descriptor based on the third-order central moment is designed to enhance targets and suppress backgrounds.</p>
</list-item>
<list-item>
<p>3. Extensive experiments have been conducted, and multiple evaluation metrics have been employed to validate the effectiveness and superiority of the proposed method.</p>
</list-item>
</list>
</p>
<p>The other sections of the paper are organized as follows: <xref ref-type="sec" rid="s2">Section 2</xref> details the method proposed in this paper; <xref ref-type="sec" rid="s3">Section 3</xref> presents comparative experiments with six baseline methods in real infrared scenes, provides experimental results, and uses a series of evaluation metrics to verify the effectiveness of the proposed method; finally, <xref ref-type="sec" rid="s4">Section 4</xref> concludes this article.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>2 Materials and methods</title>
<p>
<xref ref-type="fig" rid="F1">Figure 1</xref> illustrates the detection workflow of the proposed method. Initially, the infrared image is preprocessed using a top-hat operator to suppress most of the background clutter and extract candidate target pixels. Subsequently, the contrast of the third-order central moments in the local areas of each candidate target pixel is calculated to enhance the targets and suppress clutter, resulting in a saliency detection map. Finally, the targets are accurately segmented and extracted using an adaptive threshold.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Illustration of whole target detection flowchart of the proposed approach. The detection process includes top-hat filtering, calculation of the third-order central moment, and the extraction of targets by adaptive threshold segmentation.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g001.tif"/>
</fig>
<sec id="s2-1">
<title>2.1 Preprocessing</title>
<p>Due to the weak target signal and low target intensity, the target is easily masked under complex background or strong clutter interference. In order to enhance the target detection ability under various backgrounds, here the top-hat operator is used to do preprocessing on the whole image to suppress the background noise and improve the image signal-to-noise ratio. For the original image and the structure element two basic operations, namely dilation and erosion, are defined. The dilation operation makes the gray value of the image larger than the gray value of the input image due to the maxima operation, while the erosion operation makes the gray value of the image smaller than the gray value of the input image due to the minima operation. Thus, dilation results in increasing the size of the bright areas and decreasing the size of the dark areas. Erosion results in the opposite. They are denoted by and as <xref ref-type="disp-formula" rid="e1">Equations 1</xref>, <xref ref-type="disp-formula" rid="e2">2</xref>:<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2295;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mi mathvariant="bold-italic">max</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mtext>&#x200a;</mml:mtext>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
<disp-formula id="e2">
<mml:math id="m2">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2296;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mi mathvariant="bold-italic">min</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mtext>&#x200a;</mml:mtext>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf1">
<mml:math id="m3">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf2">
<mml:math id="m4">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> are the coordinates of the pixels in the image, <inline-formula id="inf3">
<mml:math id="m5">
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf4">
<mml:math id="m6">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> are the offsets of the coordinates of the pixels in the structure element with respect to <inline-formula id="inf5">
<mml:math id="m7">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Then, the open operation <inline-formula id="inf7">
<mml:math id="m9">
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2218;</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> and the closed operation <inline-formula id="inf8">
<mml:math id="m10">
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> can be expressed as <xref ref-type="disp-formula" rid="e3">Equations 3</xref>, <xref ref-type="disp-formula" rid="e4">4</xref>:<disp-formula id="e3">
<mml:math id="m11">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2218;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2299;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2295;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
<disp-formula id="e4">
<mml:math id="m12">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2295;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2296;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>By using the open operation, the background can be obtained after extracting the fore-ground. Subsequently, the background is subtracted from the original image to high-light the target. The above process is also described as top-hat transform and is defined as <xref ref-type="disp-formula" rid="e5">Equations 5</xref>, <xref ref-type="disp-formula" rid="e6">6</xref>:<disp-formula id="e5">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">T</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x2218;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>
<disp-formula id="e6">
<mml:math id="m14">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">B</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:mi mathvariant="bold-italic">b</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="bold-italic">f</mml:mi>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>where <inline-formula id="inf9">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> are referred to as the open top-hat and closed top-hat operations, respectively.</p>
<p>As illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref>, after preprocessing with the top-hat transformation operation, the SNR of infrared images under different backgrounds is improved, and the targets are significantly enhanced. Although some clutter remains, the background clutter in local areas tends to spread in a certain direction, exhibiting local directional consistency in grayscale values. The areas containing targets show drastic changes in grayscale values, with little correlation to the surrounding neighborhood.</p>
</sec>
<sec id="s2-2">
<title>2.2 Calculation of the third-order center moment</title>
<p>By employing a sliding window of size <inline-formula id="inf11">
<mml:math id="m17">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the image is traversed from top to bottom and left to right after preprocessing. The third-order central moment can reflect the intensity of pixel value changes within a certain spatial range, as well as whether the pixel value changes conform to a Gaussian distribution [<xref ref-type="bibr" rid="B35">35</xref>]. For image blocks with small grayscale changes or conforming to a Gaussian distribution, the third-order central moment is approximately zero. For an image patch with a size of <inline-formula id="inf12">
<mml:math id="m18">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, its third-order central moment, <inline-formula id="inf13">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, can be defined as <xref ref-type="disp-formula" rid="e7">Equations 7</xref>, <xref ref-type="disp-formula" rid="e8">8</xref>:<disp-formula id="e7">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">J</mml:mi>
<mml:mn mathvariant="bold">3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="bold-italic">M</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn mathvariant="bold">3</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
<disp-formula id="e8">
<mml:math id="m21">
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn mathvariant="bold">1</mml:mn>
<mml:mrow>
<mml:mi mathvariant="bold-italic">M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="bold-italic">M</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where, <inline-formula id="inf14">
<mml:math id="m22">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the grayscale value of the pixel at <inline-formula id="inf15">
<mml:math id="m23">
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf16">
<mml:math id="m24">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>f</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> denotes the average grayscale value of the pixel in the image patch.</p>
<p>The third-order central moment is calculated for each window area, and this value is assigned to the central pixel of the local area. This process generates a saliency map based on the third-order central moments, as shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. In this figure, the brighter pixels indicate higher grayscale values, corresponding to larger values of the third-order central moments [<xref ref-type="bibr" rid="B36">36</xref>]. It is evident from the figure that the target areas have significantly higher third-order central moment values compared to the surrounding neighborhood pixels, creating a stark contrast. The residual background clutter in the preprocessed image has lower third-order central moment values, resulting in low contrast with the surrounding areas. Therefore, it is considered to further suppress the residual clutter by applying the contrast of local third-order central moments in the image post top-hat preprocessing.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Third-order central moment saliency map and sliding window example.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g002.tif"/>
</fig>
<p>Initially, a nested sliding window structure is constructed, as illustrated in the enlarged area of the figure. This model includes a central block T and eight surrounding neighborhood blocks <inline-formula id="inf17">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, all of equal size [<xref ref-type="bibr" rid="B37">37</xref>]. The third-order central moments for the central block T and the eight surrounding blocks are calculated separately. The local third-order central moment contrast value for the area is determined and assigned to the central pixel within the sliding window.</p>
<p>The specific calculation process is as <xref ref-type="disp-formula" rid="e9">Equation 9</xref>:<disp-formula id="e9">
<mml:math id="m26">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">C</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">J</mml:mi>
<mml:mi mathvariant="bold-italic">T</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">J</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">B</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mn mathvariant="bold">8</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf18">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:mi>T</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf19">
<mml:math id="m28">
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the third-order central moments for the T area and <inline-formula id="inf20">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> areas, respectively. The minimum value of <inline-formula id="inf21">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is selected as the contrast gain coefficient <inline-formula id="inf22">
<mml:math id="m31">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> for the local area, calculated as <xref ref-type="disp-formula" rid="e10">Equation 10</xref>:<disp-formula id="e10">
<mml:math id="m32">
<mml:mrow>
<mml:mi mathvariant="bold-italic">C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold-italic">min</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">C</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mn mathvariant="bold">8</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>
</p>
<p>Then, the pixel value <inline-formula id="inf23">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> in the third-order central moment contrast saliency map, defined as <xref ref-type="disp-formula" rid="e11">Equation 11</xref>:<disp-formula id="e11">
<mml:math id="m34">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">C</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mi mathvariant="bold-italic">T</mml:mi>
</mml:msub>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="bold-italic">C</mml:mi>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where <inline-formula id="inf24">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>T</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the mean value of the 3 &#xd7; 3 region in the central of region T. The value range of the gain coefficient <inline-formula id="inf25">
<mml:math id="m36">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is influenced by the position of the window within the image and the information contained within the image patches at various locations [<xref ref-type="bibr" rid="B38">38</xref>]. The behavior of the gain coefficient <inline-formula id="inf26">
<mml:math id="m37">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> can be described in the following scenarios:<list list-type="simple">
<list-item>
<p>1. When the window is positioned in a stable background area, the background and surrounding pixels display structural similarity. Consequently, both the central and nearest neighbor blocks exhibit high third-order central moment values. Nevertheless, the discrepancy between these values is minimal, resulting in a gain coefficient <inline-formula id="inf27">
<mml:math id="m38">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that is approximately 1.</p>
</list-item>
<list-item>
<p>2. If the window is located in a target area and the central block encompasses complete target information, the difference in third-order central moment values between the central block and the neighboring blocks is substantial. This results in a gain coefficient <inline-formula id="inf28">
<mml:math id="m39">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that is greater than 1, effectively enhancing the target.</p>
</list-item>
<list-item>
<p>3. In cases where the window includes partial target information and the central block represents a background area, the third-order central moment values between the central block and the neighboring background blocks are comparable. However, they significantly differ from those of the neighboring target blocks, leading to a gain coefficient <inline-formula id="inf29">
<mml:math id="m40">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that is less than or equal to 1. This characteristic effectively suppresses the background while enhancing the target.</p>
</list-item>
</list>
</p>
<p>The image regions where the central block is the target have large contrast gain coefficients <inline-formula id="inf30">
<mml:math id="m41">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, while the contrast gain coefficients <inline-formula id="inf31">
<mml:math id="m42">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the gentle and undulating background regions are small. Although the gray value of the undulating background region changes greatly, it is strongly correlated with the gray value of the surrounding neighbourhood, and there is a similar structure in the local region, and the third-order central moments of the small blocks in the background region do not vary much, so the contrast gain coefficient of the value is much smaller compared to the target. The dynamic adjustment of the gain coefficient <inline-formula id="inf32">
<mml:math id="m43">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> based on local image characteristics allows for the selective enhancement of targets and the suppression of background noise, improving the overall detection performance. In summary, the local third-order central moment contrast-based detection algorithm has good target enhancement and background suppression capabilities, which is conducive to target detection.</p>
</sec>
<sec id="s2-3">
<title>2.3 Target extraction</title>
<p>After the above operation, the target and background are well enhanced and suppressed. Subsequently, the adaptive threshold segmentation operation is employed to extract targets, and the threshold calculation formula is defined as <xref ref-type="disp-formula" rid="e12">Equation 12</xref>:<disp-formula id="e12">
<mml:math id="m44">
<mml:mrow>
<mml:mi mathvariant="bold-italic">T</mml:mi>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold-italic">&#x3bc;</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold-italic">k</mml:mi>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>where <inline-formula id="inf33">
<mml:math id="m45">
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf34">
<mml:math id="m46">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> denote the mean and standard deviation of the processed image, respectively, and <inline-formula id="inf35">
<mml:math id="m47">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is an adjustable parameter that allows for threshold adjustment in different scenarios, and experimental results show that a setting of 5 is quite appropriate in our work. The framework of the proposed method in this paper is summarized in <xref ref-type="statement" rid="Algorithm_1">Algorithm 1</xref>.</p>
<p>
<statement content-type="algorithm" id="Algorithm_1">
<label>Algorithm 1.</label>
<p>Specific Target Detection Steps of The Proposed Method.<list list-type="simple">
<list-item>
<p>
<bold>Input:</bold>Infrared image <inline-formula id="inf36">
<mml:math id="m48">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, size of local window <inline-formula id="inf37">
<mml:math id="m49">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, size of large window <inline-formula id="inf38">
<mml:math id="m50">
<mml:mrow>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, contrast enhancement exponent <inline-formula id="inf39">
<mml:math id="m51">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, structural element for top-hat transformation <inline-formula id="inf40">
<mml:math id="m52">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and parameter <inline-formula id="inf41">
<mml:math id="m53">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</p>
</list-item>
<list-item>
<p>
<bold>Output:</bold>Detection result</p>
</list-item>
<list-item>
<p>1:&#x2003;Calculate the top-hat transformation map <inline-formula id="inf42">
<mml:math id="m54">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>using <xref ref-type="disp-formula" rid="e5">Equation 5</xref>;</p>
</list-item>
<list-item>
<p>2:&#x2003;<bold>for</bold>1 to <inline-formula id="inf43">
<mml:math id="m55">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> <bold>do</bold>
</p>
</list-item>
<list-item>
<p>3:&#x2003;Calculate the local mean <inline-formula id="inf44">
<mml:math id="m56">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>f</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula>using <xref ref-type="disp-formula" rid="e8">Equation 8</xref>;</p>
</list-item>
<list-item>
<p>4:&#x2003;Calculate the local contrast <inline-formula id="inf45">
<mml:math id="m57">
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>using <xref ref-type="disp-formula" rid="e7">Equation 7</xref>;</p>
</list-item>
<list-item>
<p>5:&#x2003;Calculate the absolute mean contrast;</p>
</list-item>
<list-item>
<p>6:&#x2003;<bold>end for</bold>
</p>
</list-item>
<list-item>
<p>7:&#x2003;Calculate the contrast metrics of the sub-blocks and normalize them;</p>
</list-item>
<list-item>
<p>8:&#x2003;Obtain the saliency map based on the third-order central moment;</p>
</list-item>
<list-item>
<p>9:&#x2003;Obtain the final detection result using <xref ref-type="disp-formula" rid="e12">Equation 12</xref>.</p>
</list-item>
</list>
</p>
</statement>
</p>
</sec>
</sec>
<sec id="s3">
<title>3 Experiment and analysis</title>
<p>Here, we first introduce the datasets used in the experiments and then analyses the performance of the proposed method from both qualitative and quantitative perspectives. Qualitatively, the performance of the method is described through the detection result images and their three-dimensional distributions. Quantitatively, the analysis is conducted based on several metrics: Signal-to-Noise Ratio (SNR), Signal-to-Noise Ratio Gain (SNRG), Background Suppression Factor (BSF), Receiver Operating Characteristic (ROC) curves, Area Under the Curve (AUC), and the average runtime of the algorithm.</p>
<sec id="s3-1">
<title>3.1 Datasets and baseline methods</title>
<p>To further validate the effectiveness and robustness of the proposed algorithm, six infrared image sequences and an infrared dataset containing mixed frames are select-ed as the experimental dataset. Among these, the background of scenes 1 and 6 contain strong cloud clutter, and the background of scene 2 includes high-brightness buildings. The single-frame mixed dataset contains high-brightness point-like noise similar to re-al targets, mountain-forest environments, high-brightness interference objects, and multi-target scenarios, significantly increasing the detection difficulty. <xref ref-type="table" rid="T1">Table 1</xref> provides a detailed feature description of the dataset.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Presentation of specific details for sequences 1&#x2013;6.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center"/>
<th align="center">Frames number</th>
<th align="center">Image size</th>
<th align="center">Specific characteristics of target and background</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Seq1</td>
<td align="center">30</td>
<td align="center">256 &#xd7; 200</td>
<td align="center">Gloomy sky background, strong edge cloud clutter<break/>The target intensity is weak and overlaps with the cloud</td>
</tr>
<tr>
<td align="center">Seq2</td>
<td align="center">40</td>
<td align="center">300 &#xd7; 300</td>
<td align="center">Gloomy sky background, high brightness tree and building clutter<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq3</td>
<td align="center">400</td>
<td align="center">330 &#xd7; 230</td>
<td align="center">Dim sky background, strong edge cloud clutter<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq4</td>
<td align="center">400</td>
<td align="center">198 &#xd7; 200</td>
<td align="center">Ground - sky background, railing<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq5</td>
<td align="center">100</td>
<td align="center">506 &#xd7; 404</td>
<td align="center">Sky background, buildings<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq6</td>
<td align="center">200</td>
<td align="center">254 &#xd7; 200</td>
<td align="center">Sky background, strong edge cloud clutter<break/>Target and cloud overlap</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To illustrate the efficacy and robustness of the proposed approach, six established methods were chosen as baseline comparisons. The IPI [<xref ref-type="bibr" rid="B10">10</xref>] model was selected to represent non-local information-based techniques. Local information-based methods comprised MPCM [<xref ref-type="bibr" rid="B21">21</xref>], LoG [<xref ref-type="bibr" rid="B12">12</xref>], AAGD [<xref ref-type="bibr" rid="B9">9</xref>] and FKRW [<xref ref-type="bibr" rid="B13">13</xref>]. Furthermore, the RIPT [<xref ref-type="bibr" rid="B33">33</xref>] method, which combines local and non-local <italic>a priori</italic> information, was selected as a comparison method. The parameters of different methods are adjusted to the best through experiments.</p>
</sec>
<sec id="s3-2">
<title>3.2 Evaluation indicators</title>
<p>The evaluation of infrared small target detection algorithms can be conducted from both qualitative and quantitative perspectives. Qualitative analysis involves subjective assessment based on the detection result images and their corresponding three-dimensional distributions, such as whether targets are detected, the number of false alarms, and the degree of background clutter suppression compared to the original image. Due to the influence of human subjective factors, it is essential to perform a quantitative analysis to objectively evaluate the experimental results. Common evaluation metrics include Signal-to-Noise Ratio (SNR), Signal-to-Noise Ratio Gain (SNRG), Background Suppression Factor (BSF), Probability of Detection (PD), False Alarm Rate (FA), Receiver Operating Characteristic (ROC) curve and Area Under the Curve (AUC). SNR is expressed as <xref ref-type="disp-formula" rid="e13">Equation 13</xref>:<disp-formula id="e13">
<mml:math id="m58">
<mml:mrow>
<mml:mi mathvariant="bold-italic">S</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mi mathvariant="bold-italic">R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:mi mathvariant="bold-italic">max</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mi mathvariant="bold-italic">e</mml:mi>
<mml:mi mathvariant="bold-italic">a</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(13)</label>
</disp-formula>where, <inline-formula id="inf46">
<mml:math id="m59">
<mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi mathvariant="italic">max</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf47">
<mml:math id="m60">
<mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represent the maximum grey value and the average grey value of the image respectively, and <inline-formula id="inf48">
<mml:math id="m61">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> signifies the standard deviation of the image. The SNRG is defined as <xref ref-type="disp-formula" rid="e14">Equation 14</xref>:<disp-formula id="e14">
<mml:math id="m62">
<mml:mrow>
<mml:mi mathvariant="bold-italic">S</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mi mathvariant="bold-italic">R</mml:mi>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">20</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">log</mml:mi>
<mml:mn mathvariant="bold">10</mml:mn>
</mml:msub>
<mml:mo>&#x2061;</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">S</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">R</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">o</mml:mi>
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">S</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">R</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(14)</label>
</disp-formula>where, <inline-formula id="inf49">
<mml:math id="m63">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>N</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf50">
<mml:math id="m64">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>N</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denote the SNR of the processed image and the original image, respectively. In general, the larger SNRG is, the better the target enhancement performance of the method.</p>
<p>BSF can be used to describe the background suppression ability of the corresponding method as <xref ref-type="disp-formula" rid="e15">Equation 15</xref>:<disp-formula id="e15">
<mml:math id="m65">
<mml:mrow>
<mml:mi mathvariant="bold-italic">B</mml:mi>
<mml:mi mathvariant="bold-italic">S</mml:mi>
<mml:mi mathvariant="bold-italic">F</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3c3;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">o</mml:mi>
<mml:mi mathvariant="bold-italic">u</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(15)</label>
</disp-formula>where, <inline-formula id="inf51">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf52">
<mml:math id="m67">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> indicate the standard deviation of the processed image and the standard deviation of the original image, respectively. Normally, the higher the BSF value, the better the method suppresses the infrared background.</p>
<p>The ROC curve is a critical metric for assessing target detection performance, composed of PD and FA rates, defined as <xref ref-type="disp-formula" rid="e16">Equation 16</xref>:<disp-formula id="e16">
<mml:math id="m68">
<mml:mrow>
<mml:mi mathvariant="bold-italic">P</mml:mi>
<mml:mi mathvariant="bold-italic">D</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">T</mml:mi>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mi mathvariant="bold-italic">T</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">F</mml:mi>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">F</mml:mi>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mi mathvariant="bold-italic">P</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(16)</label>
</disp-formula>where, <inline-formula id="inf53">
<mml:math id="m69">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the number of correctly detected targets, <inline-formula id="inf54">
<mml:math id="m70">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> denotes the actual number of true targets, <inline-formula id="inf55">
<mml:math id="m71">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf56">
<mml:math id="m72">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> respectively represent the number of pixels in false alarm regions and the total number of pixels in the test data.</p>
<p>When there is an overlap between the detected target pixels and the real target pixels and the central distance between them is less than 5 pixels, the detected target is considered to be a real target, and <italic>vice versa</italic>, it is considered to be a false target.</p>
<p>Additionally, the AUC can be calculated as a supplementary quantitative evaluation metric for the ROC curve. Generally, a larger AUC signifies better detection performance represented by the ROC curve. The area under the ROC curve, AUC, is calculated using a non-parametric method. The area under the curve is calculated as <xref ref-type="disp-formula" rid="e17">Equation 17</xref>:<disp-formula id="e17">
<mml:math id="m73">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mi mathvariant="bold-italic">U</mml:mi>
<mml:mi mathvariant="bold-italic">C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn mathvariant="bold">2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mtext>&#x200a;</mml:mtext>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(17)</label>
</disp-formula>where <inline-formula id="inf57">
<mml:math id="m74">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf58">
<mml:math id="m75">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represent the false alarm rate and the detection rate, and <inline-formula id="inf59">
<mml:math id="m76">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> means the total number of operating points on the ROC curve.</p>
</sec>
<sec id="s3-3">
<title>3.3 Analysis and discussion of experimental results</title>
<p>This section employs a series of evaluation metrics to assess the detection performance of different methods, thereby validating the effectiveness and robustness of the proposed approach. The experiments utilized six sequence datasets and one single-frame mixed dataset. The proposed method was compared with baseline methods to verify its effectiveness and robustness.</p>
<sec id="s3-3-1">
<title>3.3.1 Visual comparision</title>
<p>The detection results are presented in <xref ref-type="fig" rid="F3">Figures 3</xref>, <xref ref-type="fig" rid="F4">4</xref>. <xref ref-type="fig" rid="F3">Figure 3</xref> displays the out-comes for the sequence datasets, while <xref ref-type="fig" rid="F4">Figure 4</xref> shows the results for the single-frame mixed dataset. In Sequence 1, despite the significant cloud clutter in the background, the targets are distinct and brightly illuminated. The baseline methods succeed in detecting the targets, but the heightened background clutter leads to an elevated false alarm rate. In Sequence 2, where the targets are less bright and the background interference is more pronounced, the performance of the LoG and AAGD methods exhibit a marked decline in detection accuracy, failing to identify the actual targets. The remaining baseline methods manage to detect the actual targets but also flag numerous suspicious targets. For Seq 3 and 6, which feature more uniform scenes, all seven methods yield favorable detection results. However, Seq 4 and 5 present challenges due to the high-brightness interference clutter in the backgrounds, leading to suboptimal performance from the baseline methods. The MPCM method detects real targets in both sequences but also registers numerous background noises and suspicious targets. The AAGD, FKRW, IPI, and RIPT methods fail to detect real targets in Sequence 4 and retain false targets. In Sequence 5, although AAGD, IPI, and RIPT detect real targets, they also identify suspicious targets, and FKRW fails to detect real targets.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Visual comparison results of different approaches on sequences 1&#x2013;6. For better resolution, the red rectangle are used to mark the target, respectively.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g003.tif"/>
</fig>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Visual comparison of single-frame mixed datasets, the red rectangle are used to mark the target, respectively.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g004.tif"/>
</fig>
<p>As observed in <xref ref-type="fig" rid="F4">Figure 4</xref>, the proposed method performs excellently in various scenes, demonstrating strong target response capabilities and background suppression abilities, and is capable of detecting multiple targets.</p>
</sec>
<sec id="s3-3-2">
<title>3.3.2 Quantitative comparison</title>
<p>SNRG and BSF are used to evaluate the target enhancement and background suppression capabilities of the proposed method compared to baseline methods. Higher values of SNRG and BSF indicate superior performance of the respective methods. <xref ref-type="table" rid="T2">Tables 2</xref>, <xref ref-type="table" rid="T3">3</xref> display the SNRG and BSF values for different methods under six different backgrounds, where bold numbers represent the highest values of SNRG and BSF in each sequence, and underlined numbers indicate the second highest values. The proposed method achieved the highest SNRG and BSF in Sequences 2, 4, and 5. Notably, high SNRG and BSF values are primarily found in the RIPT, FKRW, and the proposed methods. Among these, the RIPT model showed the highest SNRG and BSF values in Scenes 1 and 3, although background clutter still existed in the detection results of Sequence 1.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Average SNRG values of different methods in six real scenes. The bold value is the maximum value.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="left">LoG [<xref ref-type="bibr" rid="B12">12</xref>]</th>
<th align="left">AAGD [<xref ref-type="bibr" rid="B9">9</xref>]</th>
<th align="left">MPCM [<xref ref-type="bibr" rid="B21">21</xref>]</th>
<th align="left">FKRW [<xref ref-type="bibr" rid="B13">13</xref>]</th>
<th align="left">IPI [<xref ref-type="bibr" rid="B10">10</xref>]</th>
<th align="left">RIPT [<xref ref-type="bibr" rid="B33">33</xref>]</th>
<th align="left">Proposed</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Seq 1</td>
<td align="left">42.9126</td>
<td align="left">37.2999</td>
<td align="left">18.1030</td>
<td align="left">46.5820</td>
<td align="left">45.7398</td>
<td align="left">
<bold>47.1608</bold>
</td>
<td align="left">46.5872</td>
</tr>
<tr>
<td align="left">Seq 2</td>
<td align="left">12.6842</td>
<td align="left">22.7342</td>
<td align="left">&#x2212;3.3249</td>
<td align="left">33.7752</td>
<td align="left">25.8948</td>
<td align="left">31.9745</td>
<td align="left">
<bold>35.1651</bold>
</td>
</tr>
<tr>
<td align="left">Seq 3</td>
<td align="left">38.6286</td>
<td align="left">39.0160</td>
<td align="left">25.0280</td>
<td align="left">40.0749</td>
<td align="left">39.6803</td>
<td align="left">
<bold>41.1559</bold>
</td>
<td align="left">38.9639</td>
</tr>
<tr>
<td align="left">Seq 4</td>
<td align="left">3.1224</td>
<td align="left">12.4164</td>
<td align="left">11.8490</td>
<td align="left">24.5923</td>
<td align="left">25.5748</td>
<td align="left">22.5606</td>
<td align="left">
<bold>26.3360</bold>
</td>
</tr>
<tr>
<td align="left">Seq 5</td>
<td align="left">15.2126</td>
<td align="left">20.7905</td>
<td align="left">19.9816</td>
<td align="left">37.1702</td>
<td align="left">16.5926</td>
<td align="left">29.0504</td>
<td align="left">
<bold>41.4241</bold>
</td>
</tr>
<tr>
<td align="left">Seq 6</td>
<td align="left">42.5041</td>
<td align="left">43.1355</td>
<td align="left">
<bold>44.4799</bold>
</td>
<td align="left">42.7732</td>
<td align="left">NaN</td>
<td align="left">42.6279</td>
<td align="left">41.4010</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Average BSF values of different methods in six real scenes. The bold value is the maximum value.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="left">LoG [<xref ref-type="bibr" rid="B12">12</xref>]</th>
<th align="left">AAGD [<xref ref-type="bibr" rid="B9">9</xref>]</th>
<th align="left">MPCM [<xref ref-type="bibr" rid="B21">21</xref>]</th>
<th align="left">FKRW [<xref ref-type="bibr" rid="B13">13</xref>]</th>
<th align="left">IPI [<xref ref-type="bibr" rid="B10">10</xref>]</th>
<th align="left">RIPT [<xref ref-type="bibr" rid="B33">33</xref>]</th>
<th align="left">Proposed</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Seq 1</td>
<td align="left">17.4953</td>
<td align="left">9.1209</td>
<td align="left">5.9284</td>
<td align="left">26.2413</td>
<td align="left">23.2279</td>
<td align="left">
<bold>27.5569</bold>
</td>
<td align="left">26.3840</td>
</tr>
<tr>
<td align="left">Seq 2</td>
<td align="left">3.7184</td>
<td align="left">11.4748</td>
<td align="left">2.4414</td>
<td align="left">40.9638</td>
<td align="left">17.4044</td>
<td align="left">33.9158</td>
<td align="left">
<bold>48.8688</bold>
</td>
</tr>
<tr>
<td align="left">Seq 3</td>
<td align="left">49.0199</td>
<td align="left">52.9625</td>
<td align="left">27.3042</td>
<td align="left">59.2640</td>
<td align="left">55.6245</td>
<td align="left">
<bold>66.0983</bold>
</td>
<td align="left">51.7224</td>
</tr>
<tr>
<td align="left">Seq 4</td>
<td align="left">1.3675</td>
<td align="left">3.8351</td>
<td align="left">3.5887</td>
<td align="left">16.6863</td>
<td align="left">17.6726</td>
<td align="left">12.2383</td>
<td align="left">
<bold>19.9535</bold>
</td>
</tr>
<tr>
<td align="left">Seq 5</td>
<td align="left">4.5858</td>
<td align="left">8.6565</td>
<td align="left">2.1983</td>
<td align="left">58.7325</td>
<td align="left">5.3834</td>
<td align="left">22.4054</td>
<td align="left">
<bold>92.9656</bold>
</td>
</tr>
<tr>
<td align="left">Seq 6</td>
<td align="left">42.1676</td>
<td align="left">45.3576</td>
<td align="left">
<bold>53.0407</bold>
</td>
<td align="left">43.6129</td>
<td align="left">Inf</td>
<td align="left">42.7872</td>
<td align="left">37.6230</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As shown in <xref ref-type="fig" rid="F5">Figure 5</xref>, the ROC curves indicate the relationship between Pd and Pf in Seq 1&#x2013;6 for all methods. Also, in order to visually evaluate the detection performance of different algorithms, we calculated the area under the curve (AUC) values of different methods, as shown in <xref ref-type="table" rid="T4">Table 4</xref>. Bold and underline indicate the maximum and second largest values, respectively. In general, the closer the ROC curve is to the upper left corner and the larger the AUC value is, the better the detection performance is. For seq 1, the target detection result of LoG and MPCM methods show the presence of obvious interfering targets, which fail to be detected accurately. Therefore, a lower Pd value corresponds to it when the pf is the same, causing a low AUC value. In seq 2 and 5, MPCM obtained the closest detection accuracy to the proposed method. The AUC values of the corresponding ROC curves were also second only to our method. In seq 3 and 4, the AUC values of LoG and RIPT were low. In sequence 6, AAGD accurately detects the target, closest to the detection accuracy of the proposed method. The AUC values of the corresponding ROC curves are also second only to our method. The ROC curves of the proposed method in this paper consistently maintain the optimal Pd value and the maximum AUC value in all six scenarios, indicating that our method has excellent target detection capability in all six scenarios.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>
<bold>(A&#x2013;F)</bold> Comparison of ROC curves obtained by all methods on Sequences 1-6.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g005.tif"/>
</fig>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>AUC values corresponding to ROC curves of all approaches on sequences 1.6 (&#xd7; <inline-formula id="inf60">
<mml:math id="m77">
<mml:mrow>
<mml:msup>
<mml:mn>10</mml:mn>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> ). The bold value is the maximum value.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="left">MPCM [<xref ref-type="bibr" rid="B21">21</xref>]</th>
<th align="left">LoG [<xref ref-type="bibr" rid="B12">12</xref>]</th>
<th align="left">AAGD [<xref ref-type="bibr" rid="B9">9</xref>]</th>
<th align="left">FKRW [<xref ref-type="bibr" rid="B13">13</xref>]</th>
<th align="left">IPI [<xref ref-type="bibr" rid="B10">10</xref>]</th>
<th align="left">RIPT [<xref ref-type="bibr" rid="B33">33</xref>]</th>
<th align="left">Proposed</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Seq 1</td>
<td align="left">758.7167</td>
<td align="left">717.5151</td>
<td align="left">947.0283</td>
<td align="left">805.2301</td>
<td align="left">806.2288</td>
<td align="left">790.8851</td>
<td align="left">
<bold>999.9911</bold>
</td>
</tr>
<tr>
<td align="left">Seq 2</td>
<td align="left">944.0244</td>
<td align="left">801.0041</td>
<td align="left">935.2437</td>
<td align="left">777.6925</td>
<td align="left">859.6449</td>
<td align="left">794.2826</td>
<td align="left">
<bold>999.9886</bold>
</td>
</tr>
<tr>
<td align="left">Seq 3</td>
<td align="left">857.5838</td>
<td align="left">702.3206</td>
<td align="left">993.0174</td>
<td align="left">780.9553</td>
<td align="left">852.5991</td>
<td align="left">678.0886</td>
<td align="left">
<bold>999.9926</bold>
</td>
</tr>
<tr>
<td align="left">Seq 4</td>
<td align="left">693.3668</td>
<td align="left">581.772</td>
<td align="left">718.3679</td>
<td align="left">607.5147</td>
<td align="left">726.6379</td>
<td align="left">601.7396</td>
<td align="left">
<bold>981.3734</bold>
</td>
</tr>
<tr>
<td align="left">Seq 5</td>
<td align="left">984.0468</td>
<td align="left">895.36</td>
<td align="left">945.2263</td>
<td align="left">799.9512</td>
<td align="left">946.2618</td>
<td align="left">667.6413</td>
<td align="left">
<bold>991.48</bold>
</td>
</tr>
<tr>
<td align="left">Seq 6</td>
<td align="left">999.4926</td>
<td align="left">990.079</td>
<td align="left">999.9209</td>
<td align="left">898.6716</td>
<td align="left">500</td>
<td align="left">847.0958</td>
<td align="left">
<bold>999.9969</bold>
</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>4 Discussion</title>
<sec id="s4-1">
<title>4.1 Discussion of detection performance</title>
<p>The detection of infrared small targets is rendered challenging by the complexity and variability of the infrared image environment, which frequently includes substantial background clutter and interference noise. Conventional detection techniques, such as LoG, MPCM, and AAGD, are characterized by high computational complexity and noise sensitivity, rendering them prone to image noise and resulting in false or missed detections. Methods like IPI and FKRW are highly responsive to variations in target gray intensity, leading to unstable detection outcomes when the target and background exhibit minimal grayscale differences or when influenced by factors such as illumination. The RIPT detection method, which integrates local and non-local in-formation, is effective in suppressing interference factors and enhancing target detection accuracy in specific scenarios, nonetheless, in complex scenes, it is more vulnerable to background interference, thereby decreasing detection accuracy.</p>
<p>This paper presents a novel methodology that commences with the preprocessing of infrared images using the top-hat operator. It proceeds to enhance target objects and suppress background clutter by calculating the third-order central moment contrast within the local region of each candidate target pixel. The extraction of targets is subsequently accomplished through adaptive threshold segmentation. The proposed method is evaluated against baseline techniques on six real sequence datasets and one single-frame hybrid dataset, demonstrating consistently superior performance across all datasets. Furthermore, the proposed method is marked by low computational complexity and is amenable to acceleration through GPU or Field-Programmable Gate Array (FPGA) technologies.</p>
</sec>
<sec id="s4-2">
<title>4.2 Discussion of the key parameter V</title>
<p>We briefly discuss in this section the selection of the key parameter <inline-formula id="inf61">
<mml:math id="m78">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> in the proposed method, and the value of <inline-formula id="inf62">
<mml:math id="m79">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> in <xref ref-type="table" rid="T5">Table 5</xref> will directly determine the quality of the generated target salient maps. To obtain the optimal value of <inline-formula id="inf63">
<mml:math id="m80">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, we select different sequences of infrared images for simulation experiments. In the experiments, we set <inline-formula id="inf64">
<mml:math id="m81">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> as 3, 5, 7, and 9 respectively. The details of different sequences of infrared images are described in <xref ref-type="table" rid="T6">Table 6</xref>. In addition, we use ROC curves to evaluate the detection performance of the proposed method at different <inline-formula id="inf65">
<mml:math id="m82">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> values. <xref ref-type="fig" rid="F6">Figure 6</xref> shows the ROC curves for different scenes.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Parameter settings for all approaches.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">No.</th>
<th align="center">Abbreviations</th>
<th align="center">Parameter settings</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td align="center">MPCM [<xref ref-type="bibr" rid="B21">21</xref>]</td>
<td align="center">Local window size: <inline-formula id="inf66">
<mml:math id="m83">
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 3,5,7,9. Mean filter size: 3 &#xd7; 3</td>
</tr>
<tr>
<td align="center">2</td>
<td align="center">LoG [<xref ref-type="bibr" rid="B12">12</xref>]</td>
<td align="center">
<inline-formula id="inf67">
<mml:math id="m84">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1.5, scale size: <inline-formula id="inf68">
<mml:math id="m85">
<mml:mrow>
<mml:mi mathvariant="normal">n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 5</td>
</tr>
<tr>
<td align="center">3</td>
<td align="center">AAGD [<xref ref-type="bibr" rid="B9">9</xref>]</td>
<td align="center">
<inline-formula id="inf69">
<mml:math id="m86">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mi mathvariant="italic">max</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 19, 19, 19, 19; <inline-formula id="inf70">
<mml:math id="m87">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mi mathvariant="italic">min</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 3, 5, 7, 9</td>
</tr>
<tr>
<td align="center">4</td>
<td align="center">FKRW [<xref ref-type="bibr" rid="B13">13</xref>]</td>
<td align="center">Local window size:11 &#xd7; 11</td>
</tr>
<tr>
<td align="center">5</td>
<td align="center">IPI [<xref ref-type="bibr" rid="B10">10</xref>]</td>
<td align="center">Patch size: 50 &#xd7; 50, sliding step:10, <inline-formula id="inf71">
<mml:math id="m88">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">min</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:msqrt>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mn>10</mml:mn>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">6</td>
<td align="center">RIPT [<xref ref-type="bibr" rid="B33">33</xref>]</td>
<td align="center">Patch size: 50 &#xd7; 50, sliding step: 10, <inline-formula id="inf72">
<mml:math id="m89">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">min</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>J</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</inline-formula> , L &#x3d; 1, h &#x3d; 10, <inline-formula id="inf73">
<mml:math id="m90">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.01, <inline-formula id="inf74">
<mml:math id="m91">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mn>10</mml:mn>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">7</td>
<td align="center">Proposed</td>
<td align="center">Local window size: <inline-formula id="inf75">
<mml:math id="m92">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 3,5,7,9 <inline-formula id="inf76">
<mml:math id="m93">
<mml:mrow>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 3&#xd7; <inline-formula id="inf77">
<mml:math id="m94">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Presentation of specific details for sequences 1&#x2013;4.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center"/>
<th align="center">Frames number</th>
<th align="center">Image size</th>
<th align="center">Specific characteristics of target and background</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Seq1</td>
<td align="center">30</td>
<td align="center">256 &#xd7; 200</td>
<td align="center">Gloomy sky background, strong edge cloud clutter<break/>The target intensity is weak and overlaps with the cloud</td>
</tr>
<tr>
<td align="center">Seq2</td>
<td align="center">30</td>
<td align="center">300 &#xd7; 300</td>
<td align="center">Gloomy sky background, high brightness building clutter<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq3</td>
<td align="center">30</td>
<td align="center">330 &#xd7; 230</td>
<td align="center">Dim sky background, strong edge cloud clutter<break/>Extremely weak target intensity and low contrast</td>
</tr>
<tr>
<td align="center">Seq4</td>
<td align="center">30</td>
<td align="center">198 &#xd7; 200</td>
<td align="center">Ground - sky background, railing<break/>Extremely weak target intensity and low contrast</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>
<bold>(A&#x2013;D)</bold> Comparison of ROC curves obtained by different v value on Sequences 1&#x2013;4.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g006.tif"/>
</fig>
<p>As shown in <xref ref-type="fig" rid="F6">Figure 6</xref>, we find that the setting of <inline-formula id="inf78">
<mml:math id="m95">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> value significantly affects the final target detection results. <inline-formula id="inf79">
<mml:math id="m96">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> value is set to 3, the detection results are the worst in all the scenes. <inline-formula id="inf80">
<mml:math id="m97">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> value is set to 7, the optimal detection results are obtained in all the scenes. Therefore, it is recommended to set the <inline-formula id="inf81">
<mml:math id="m98">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> value to 7.</p>
</sec>
<sec id="s4-3">
<title>4.3 Robustness to noisy scene</title>
<p>Furthermore, the algorithm must also overcome the challenge of detecting objects in a noisy scene. In order to verify the robustness of the proposed method in noisy environments, we add Gaussian white noise with different standard variances to the original infrared image. As shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. Gaussian white noise with standard deviation of 5, 10, and 15 was added to the original infrared image in the first, third, and fifth rows, respectively. From the detection results in the figure, it can be seen that the proposed method can effectively detect the target in different degrees of noise scenarios. Thus, the ability of the proposed method in combating noise scenarios is verified. It shows that the proposed method possesses strong robustness in noisy scenes.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Detection results obtained in a noisy scene using the proposed method. The standard deviation of Gaussian white noise in the first, second and third rows are 5, 10, and 15, respectively.</p>
</caption>
<graphic xlink:href="fphy-12-1477253-g007.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>5 Conclusion</title>
<p>In this paper, a novel target detection method based on the contrast of local third-order central moments is proposed. Initially, the top-hat operator is used to pro-cess infrared images to suppress most of the background clutter and extract candidate target pixels. Subsequently, the contrast of the third-order central moments in the local areas of each candidate target pixel is calculated to enhance targets and suppress clutter. Finally, we conducted extensive experiments and compared six SOTA approaches on a dataset containing six real IR videos with various scenes and a mixed dataset consisting of 18 single-frame infrared images with different backgrounds. Experimental results demonstrate that the proposed method can efficiently detect small infrared targets and shows significant advantages across a range of evaluation metrics. However, given the complexity and variability of real-world scenes, further in-depth research is warranted. This future work will primarily focus on the following aspects: (1) The small target detection method presented in this study is currently limited to single-frame infrared images and does not leverage the temporal information inherent in image sequences. Future investigations will integrate time-domain information to enhance target detection, considering the correlations across multiple frames. (2) The proposed method has been simulated in MATLAB, and we intend to explore its implementation on FPGA and other hardware platforms, in conjunction with existing laboratory equipment, to develop a real-time infrared small target detection system.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>DX: Conceptualization, Data curation, Formal Analysis, Methodology, Writing&#x2013;original draft, Writing&#x2013;review and editing. DP: Conceptualization, Validation, Writing&#x2013;original draft, Writing&#x2013;review and editing. SG: Data curation, Writing&#x2013;original draft, Writing&#x2013;review and editing. ZL: Conceptualization, Writing&#x2013;review and editing. ZZ: Conceptualization, Writing&#x2013;review and editing.</p>
</sec>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded by National Natural Science Foundation of China [grant number 62301036]; The Open Fund Projects of Henan Key Laboratory of General Aviation Technology [grant number ZHKF-240201].</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Dai</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Ni</surname>
<given-names>K</given-names>
</name>
</person-group>. <article-title>Graph-regularized laplace approximation for detecting small infrared target against complex backgrounds</article-title>. <source>IEEE Access</source> (<year>2019</year>) <volume>7</volume>:<fpage>85354</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1109/access.2019.2925563</pub-id>
</citation>
</ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gregoris</surname>
<given-names>DJ</given-names>
</name>
<name>
<surname>Simon</surname>
<given-names>KW</given-names>
</name>
<name>
<surname>Tritchew</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sevigny</surname>
<given-names>L</given-names>
</name>
</person-group>. <article-title>Wavelet transform-based filtering for the en-hancement of dim targets in FLIR images</article-title>. <source>Wavelet Appl</source> (<year>1994</year>) <volume>2242</volume>:<fpage>573</fpage>&#x2013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1117/12.170058</pub-id>
</citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>YQ</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>ZJ</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>WY</given-names>
</name>
</person-group>. <article-title>A method of small target detection in infrared image based on nonsubsampled contourlet transform</article-title>. <source>J Image Graphics</source> (<year>2009</year>) <volume>14</volume>(<issue>3</issue>):<fpage>477</fpage>&#x2013;<lpage>81</lpage>.</citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kong</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Qian</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Automatic detection of sea-sky horizon line and small targets in maritime infrared imagery</article-title>. <source>Infrared Phys and Technology</source> (<year>2016</year>) <volume>76</volume>:<fpage>185</fpage>&#x2013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1016/j.infrared.2016.01.016</pub-id>
</citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Yunhong</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Infrared small target detection algorithm based on double-tree complex wavelet transform</article-title>. <source>Laser and Infrared</source> (<year>2020</year>) <volume>50</volume>(<issue>9</issue>):<fpage>1145</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1109/aimsec.2011.6010000</pub-id>
</citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sheng</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Jian</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>T</given-names>
</name>
</person-group>. <article-title>Research of SVM-based infrared small object segmentation and clustering method</article-title>. <source>Signal Process.</source> (<year>2005</year>) <volume>21</volume>(<issue>5</issue>):<fpage>515</fpage>&#x2013;<lpage>9</lpage>.</citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>W</given-names>
</name>
</person-group>. <article-title>A novel infrared small moving target detection method based on tracking interest points under complicated background</article-title>. <source>Infrared Phys and Technology</source> (<year>2014</year>) <volume>65</volume>:<fpage>36</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1016/j.infrared.2014.03.007</pub-id>
</citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Long</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Small target detection in hyperspectral remote sensing image based on adaptive parameter SVM</article-title>. <source>&#x5149;&#x5b66;&#x5b66;&#x62a5;</source> (<year>2015</year>) <volume>35</volume>:<fpage>0928001</fpage>. <pub-id pub-id-type="doi">10.3788/aos201535.0928001</pub-id>
</citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yilin</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Rongbing</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Haiyan</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Qingbo</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Adaptive gradient descent optimization algorithm for measurement matrix</article-title>. <source>Application research of computers</source> (<year>2017</year>).</citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gao</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Hauptmann</surname>
<given-names>AG</given-names>
</name>
</person-group>. <article-title>Infrared patch-image model for small target de-tection in a single image</article-title>. <source>IEEE Trans Image Process</source> (<year>2013</year>) <volume>22</volume>(<issue>12</issue>):<fpage>4996</fpage>&#x2013;<lpage>5009</lpage>. <pub-id pub-id-type="doi">10.1109/tip.2013.2281420</pub-id>
</citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>Z</given-names>
</name>
</person-group>. <article-title>Infrared small target detection based on partial sum of the tensor nuclear norm</article-title>. <source>Remote Sensing</source> (<year>2019</year>) <volume>11</volume>(<issue>4</issue>):<fpage>382</fpage>. <pub-id pub-id-type="doi">10.3390/rs11040382</pub-id>
</citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shao</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>An improved infrared dim and small target detection algorithm based on the contrast mechanism of human visual system</article-title>. <source>Infrared Phys and Technology</source> (<year>2012</year>) <volume>55</volume>(<issue>5</issue>):<fpage>403</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1016/j.infrared.2012.06.001</pub-id>
</citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qin</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Bruzzone</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>B</given-names>
</name>
</person-group>. <article-title>Infrared small target detection based on facet kernel and random walker</article-title>. <source>IEEE Trans Geosci Remote Sensing</source> (<year>2019</year>) <volume>57</volume>(<issue>9</issue>):<fpage>7104</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2019.2911513</pub-id>
</citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>CP</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Xia</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>YY</given-names>
</name>
</person-group>. <article-title>A local contrast method for small infrared target detection</article-title>. <source>IEEE Trans Geosci remote sensing</source> (<year>2013</year>) <volume>52</volume>(<issue>1</issue>):<fpage>574</fpage>&#x2013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2013.2242477</pub-id>
</citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Fang</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>A robust infrared small target detection algorithm based on human visual system</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2014</year>) <volume>11</volume>(<issue>12</issue>):<fpage>2168</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2014.2323236</pub-id>
</citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qin</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>B</given-names>
</name>
</person-group>. <article-title>Effective infrared small target detection utilizing a novel local contrast method</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2016</year>) <volume>13</volume>(<issue>12</issue>):<fpage>1890</fpage>&#x2013;<lpage>4</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2016.2616416</pub-id>
</citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>L</given-names>
</name>
</person-group>. <article-title>Infrared small target detection utilizing the multiscale relative local contrast measure</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2018</year>) <volume>15</volume>(<issue>4</issue>):<fpage>612</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2018.2790909</pub-id>
</citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Moradi</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Faramarzi</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Q</given-names>
</name>
</person-group>. <article-title>A local contrast method for infrared small-target detection utilizing a tri-layer window</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2019</year>) <volume>17</volume>(<issue>10</issue>):<fpage>1822</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2019.2954578</pub-id>
</citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Niu</surname>
<given-names>Q</given-names>
</name>
</person-group>. <article-title>Infrared small target detection utilizing the enhanced clos-est-mean background estimation</article-title>. <source>IEEE J Selected Top Appl Earth Observations Remote Sensing</source> (<year>2020</year>) <volume>14</volume>:<fpage>645</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1109/jstars.2020.3038442</pub-id>
</citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Moradi</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Faramarzi</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X</given-names>
</name>
<etal/>
</person-group> <article-title>Infrared small target detection based on the weighted strengthened local contrast measure</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2020</year>) <volume>18</volume>(<issue>9</issue>):<fpage>1670</fpage>&#x2013;<lpage>4</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2020.3004978</pub-id>
</citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>You</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H</given-names>
</name>
</person-group>. <article-title>Multiscale patch-based contrast measure for small infrared target detection</article-title>. <source>Pattern Recognition</source> (<year>2016</year>) <volume>58</volume>:<fpage>216</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2016.04.002</pub-id>
</citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>F</given-names>
</name>
</person-group>. <article-title>Infrared sea-sky line detection utilizing self-adaptive Laplacian of Gaussian filter and vis-ual-saliency-based probabilistic Hough transform</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2021</year>) <volume>19</volume>:<fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2021.3111099</pub-id>
</citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>An infrared small target detection algorithm based on high-speed local contrast method</article-title>. <source>Infrared Phys and Technol</source> (<year>2016</year>) <volume>76</volume>:<fpage>474</fpage>&#x2013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.1016/j.infrared.2016.03.023</pub-id>
</citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>An infrared small target detection framework based on local contrast method</article-title>. <source>Measurement</source> (<year>2016</year>) <volume>91</volume>:<fpage>405</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1016/j.measurement.2016.05.071</pub-id>
</citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Small infrared target detection based on weighted local difference measure</article-title>. <source>IEEE Trans Geosci Remote Sensing</source> (<year>2016</year>) <volume>54</volume>(<issue>7</issue>):<fpage>4204</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2016.2538295</pub-id>
</citation>
</ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Infrared small-target detection using multiscale gray difference weighted image entropy</article-title>. <source>IEEE Trans Aerospace Electron Syst</source> (<year>2016</year>) <volume>52</volume>(<issue>1</issue>):<fpage>60</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1109/taes.2015.140878</pub-id>
</citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Entropy-based window selection for detecting dim and small infrared targets</article-title>. <source>Pattern Recognition</source> (<year>2017</year>) <volume>61</volume>:<fpage>66</fpage>&#x2013;<lpage>77</lpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2016.07.036</pub-id>
</citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>L</given-names>
</name>
</person-group>. <article-title>An effective infrared small target detection method based on the human visual attention</article-title>. <source>Infrared Phys and Technology</source> (<year>2018</year>) <volume>95</volume>:<fpage>128</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1016/j.infrared.2018.10.033</pub-id>
</citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Guizani</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Infrared small target detection through multiple feature analysis based on visual saliency</article-title>. <source>IEEE Access</source> (<year>2019</year>) <volume>7</volume>:<fpage>38996</fpage>&#x2013;<lpage>9004</lpage>. <pub-id pub-id-type="doi">10.1109/access.2019.2906076</pub-id>
</citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Du</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Hamdulla</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>Infrared small target detection using homogeneity-weighted local contrast measure</article-title>. <source>IEEE Geosci remote sensing Lett</source> (<year>2019</year>) <volume>17</volume>(<issue>3</issue>):<fpage>514</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2019.2922347</pub-id>
</citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Local gradient field feature contrast measure for infrared small target detection</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2020</year>) <volume>18</volume>(<issue>3</issue>):<fpage>553</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2020.2976208</pub-id>
</citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>N</given-names>
</name>
</person-group>. <article-title>A local contrast method combined with adaptive background estimation for infrared small target detection</article-title>. <source>IEEE Geosci Remote Sensing Lett</source> (<year>2019</year>) <volume>16</volume>(<issue>9</issue>):<fpage>1442</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2019.2898893</pub-id>
</citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dai</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Reweighted infrared patch-tensor model with both nonlocal and local priors for single-frame small target detection</article-title>. <source>IEEE J selected Top Appl earth observations remote sensing</source> (<year>2017</year>) <volume>10</volume>(<issue>8</issue>):<fpage>3752</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1109/jstars.2017.2700023</pub-id>
</citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>R</given-names>
</name>
</person-group>. <article-title>Infrared dim and small target detection based on greedy bilateral factorization in image sequences</article-title>. <source>IEEE J Selected Top Appl Earth Observations Remote Sensing</source> (<year>2020</year>) <volume>13</volume>:<fpage>3394</fpage>&#x2013;<lpage>408</lpage>. <pub-id pub-id-type="doi">10.1109/jstars.2020.2998822</pub-id>
</citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Jin</surname>
<given-names>Q</given-names>
</name>
</person-group>. <article-title>Tensor spectral k-support norm minimization for detecting infrared dim and small target against urban backgrounds</article-title>. <source>IEEE Trans Geosci Remote Sensing</source> (<year>2023</year>) <volume>61</volume>:<fpage>1</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2023.3277848</pub-id>
</citation>
</ref>
<ref id="B36">
<label>36.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>R</given-names>
</name>
</person-group>. <article-title>A novel spatiotemporal saliency method for low-altitude slow small infrared target detection</article-title>. <source>IEEE Geosci remote sensing Lett</source> (<year>2021</year>) <volume>19</volume>:<fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2020.3048199</pub-id>
</citation>
</ref>
<ref id="B37">
<label>37.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>Y</given-names>
</name>
<etal/>
</person-group> <article-title>STTM-SFR: spatial&#x2013;temporal tensor modeling with saliency filter regularization for infrared small target detection</article-title>. <source>IEEE Trans Geosci Remote Sensing</source> (<year>2022</year>) <volume>60</volume>:<fpage>1</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2022.3172745</pub-id>
</citation>
</ref>
<ref id="B38">
<label>38.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Facet derivative-based multidirectional edge awareness and spatial&#x2013;temporal tensor model for infrared small target detection</article-title>. <source>IEEE Trans Geosci Remote Sensing</source> (<year>2022</year>) <volume>60</volume>:<fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2021.3098969</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>