<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="research-article" dtd-version="1.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Earth Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Earth Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Earth Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-6463</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1790572</article-id>
<article-id pub-id-type="doi">10.3389/feart.2026.1790572</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>WTTnet: a network combining wavelet transform and transformer for denoising microseismic signal</article-title>
<alt-title alt-title-type="left-running-head">Sun et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/feart.2026.1790572">10.3389/feart.2026.1790572</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Sun</surname>
<given-names>Wei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3353383"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing - original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing - review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yu</surname>
<given-names>Shengbao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing - review and editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wang</surname>
<given-names>Junqiu</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing - review and editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<institution>Key Laboratory of Geophysical Exploration Equipment, Ministry of Education of China</institution>, <city>Changchun</city>, <country country="CN">China</country>
</aff>
<aff id="aff2">
<label>2</label>
<institution>College of Instrumentation and Electrical Engineering Department, Jilin University</institution>, <city>Changchun</city>, <country country="CN">China</country>
</aff>
<author-notes>
<corresp id="c001">
<label>&#x2a;</label>Correspondence: Junqiu Wang, <email xlink:href="mailto:wjq@jlu.edu.cn">wjq@jlu.edu.cn</email>
</corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-19">
<day>19</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>14</volume>
<elocation-id>1790572</elocation-id>
<history>
<date date-type="received">
<day>18</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="rev-recd">
<day>03</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Sun, Yu and Wang.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Sun, Yu and Wang</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-19">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Noise suppression is a key component in microseismic monitoring technology. Accurate denoising of microseismic signals is crucial for ensuring reliable data for locating mining-related seismic events and analyzing the state of rock mass during mining operations. This paper proposes a network combining wavelet transform and Transformer for denoising microseismic signals (WTTnet). WTTnet leverages discrete wavelet transform to separate the high- and low-frequency components of the input signal. These components are concatenated to form full-frequency features, which are then used as query and value vectors in the Transformer, while the high-frequency features serve as keys. The multi-head self-attention mechanism captures cross-scale correlations. Finally, inverse discrete wavelet transform converts the frequency-domain output back to the time domain. The primary strength of this model is its ability to identify and distinguish noise components across varying frequencies. The proposed method is tested on synthetic data contaminated with various noise types and on field data. Its denoising performance is evaluated using appropriate metrics and compared with other denoising methods. Experimental results show that this method outperforms traditional denoising methods in terms of overall denoising performance across diverse noise conditions.</p>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>denoising</kwd>
<kwd>seismic</kwd>
<kwd>signal processing</kwd>
<kwd>transformer</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="13"/>
<table-count count="0"/>
<equation-count count="9"/>
<ref-count count="30"/>
<page-count count="00"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Solid Earth Geophysics</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Microseismic monitoring technology is widely applied in the development of unconventional oil and gas, monitoring of mining disasters, and <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>O</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> capture, utilization, and storage (CCUS) fields. In the microseismic monitoring process, denoising is a critical data processing step. Due to the small amplitude of microseismic signals and the frequent interference from noise in the measured data, their signal-to-noise ratio (SNR) is much lower than that of active-source seismic events. This makes it challenging to directly extract arrival times and locate events from the raw observed data (<xref ref-type="bibr" rid="B30">Zhang et al., 2022</xref>; <xref ref-type="bibr" rid="B25">Waheed et al., 2024</xref>).</p>
<p>To suppress noise in microseismic data, extensive research has been conducted, leading to various effective denoising methods (<xref ref-type="bibr" rid="B8">Dong et al., 2025</xref>; <xref ref-type="bibr" rid="B9">Gao et al., 2021</xref>). Traditional microseismic noise suppression methods mainly rely on domain transformation algorithms and thresholding techniques, including wavelet transform (WT) (<xref ref-type="bibr" rid="B20">Mousavi et al., 2016</xref>), continuous wavelet transform (CWT) (<xref ref-type="bibr" rid="B19">Mousavi and Langston, 2016</xref>), short-time Fourier transform (STFT) (<xref ref-type="bibr" rid="B1">Allen and Rabiner, 1977</xref>), curvelet transform (<xref ref-type="bibr" rid="B17">Liu et al., 2016</xref>), and empirical mode decomposition (EMD) (<xref ref-type="bibr" rid="B3">Chen and Fomel, 2018</xref>), among others. These traditional methods have relatively weak capabilities in capturing local features, resulting in poor denoising performance in complex noise environments.</p>
<p>With the continuous development of deep learning techniques and the accumulation of seismic data, deep learning has become a powerful tool in seismology (<xref ref-type="bibr" rid="B22">Saad et al., 2023</xref>; <xref ref-type="bibr" rid="B7">Dong et al., 2024</xref>; <xref ref-type="bibr" rid="B18">Mousavi and Beroza, 2022</xref>). Its ability to process large volumes of data and extract complex non-linear features has led to widespread applications across various subfields, including earthquake detection, phase picking, and seismic signal analysis (<xref ref-type="bibr" rid="B5">Cui et al., 2025</xref>; <xref ref-type="bibr" rid="B24">Tang et al., 2023</xref>; <xref ref-type="bibr" rid="B4">Chen et al., 2025</xref>). In particular, many researchers have focused on applying deep learning to microseismic signal denoising. <xref ref-type="bibr" rid="B15">Lin et al. (2023)</xref> proposed a microseismic denoising method based on a Convolutional Blind Denoising Network (CBDNet), which combines noise estimation and denoising sub-networks to directly learn and remove complex background noise from noisy microseismic data. Its limitation lies in the fact that both CNN sub-networks are built upon standard convolutions with limited and fixed receptive fields. When dealing with multi-scale noise whose energy is distributed across a very broad frequency band, the network struggles to simultaneously model both low-frequency and high-frequency noise, which may lead to suboptimal suppression of noise at specific scales. <xref ref-type="bibr" rid="B13">Iqbal (2023)</xref> proposed a new denoising framework based on an intelligent deep convolutional neural network, which simultaneously learns sparse representations of data in the time-frequency domain and adaptively captures microseismic signals corrupted by noise. Although this method incorporates sparse representation in the time-frequency domain, its core feature learning and adaptive capture are still accomplished through a deep CNN. The translation invariance of CNN treats all time-frequency coefficients equally, making it difficult to precisely eliminate noise that appears in specific frequency bands. <xref ref-type="bibr" rid="B21">Saad et al. (2021)</xref> developed an unsupervised deep learning method based on variational autoencoder (VAE) and squeeze-and-excitation (SE) networks for enhancing microseismic signals and suppressing noise. The core issue of this method is that the CNN encoder compresses and mixes noise and signal information from different frequencies into a single latent vector, resulting in entanglement of frequency-domain information and ineffective removal of specific frequency noise components. <xref ref-type="bibr" rid="B12">Hu et al. (2023)</xref> employed the Short-Time Fourier Transform to convert signals into the time-frequency domain and used deep convolutional autoencoders to automatically learn features and separate effective microseismic signals from noise. The shortcoming of this approach is that the window length of STFT is fixed, unable to adaptively match the optimal analysis scale for both signal and noise. Moreover, the subsequent convolutional autoencoder treats the frequency axis as an ordinary spatial dimension when processing the two-dimensional time-frequency map, disregarding the continuous physical meaning and scale structure inherent to the frequency axis. <xref ref-type="bibr" rid="B2">Cai et al. (2024)</xref> captured noise features across different scales through multi-scale dilated convolution layers and used an improved attention mechanism to adaptively enhance key signal channels, achieving efficient microseismic denoising. This method uses dilated convolutions to simulate multi-scale analysis by expanding the receptive field. However, its drawback is that branches with different dilation rates operate in parallel or series, and the features they extract remain a mixture of information from different scales.</p>
<p>Most of the methods mentioned above rely on CNN as the backbone for feature extraction. Due to its inherent local connectivity, weight sharing, and hierarchical feature mixing characteristics, CNN excels in capturing local patterns, but it is less capable of performing explicit frequency band decomposition or modeling global dependencies across scales. Consequently, when dealing with multi-scale noise that exhibits distinct frequency-band distributions, these methods often perform only coarse filtering in the time domain or mixed time-frequency feature spaces, failing to achieve precise, frequency-band-level noise identification and separation.</p>
<p>The core advantage of Transformer lies in its self-attention mechanism, which can model global dependencies between any positions in the input sequence, thereby effectively capturing complex long-term correlation patterns between signal and noise in the time-frequency domain (<xref ref-type="bibr" rid="B10">Gao et al., 2025</xref>). Unlike CNNs, which primarily rely on local receptive fields, this global modeling capability enables Transformer to adaptively distinguish structured effective signals from unstructured background noise, while dynamically allocating attention weights to enhance weak signal segments and suppress noise-dominated regions (<xref ref-type="bibr" rid="B29">Zhang et al., 2019</xref>). Consequently, when processing sequential data like microseismic signals, where noise and signal are highly intertwined on a global scale, Transformer offers more refined and adaptive denoising capabilities compared to traditional CNN-based methods (<xref ref-type="bibr" rid="B27">Wen et al., 2025</xref>). Although the Transformer architecture has not yet been widely applied in microseismic signal denoising, it has demonstrated significant advantages and achieved successful applications in denoising tasks within other branches of seismology. For example, <xref ref-type="bibr" rid="B26">Wang et al. (2023)</xref> further developed a multi-scale interactive Transformer network to effectively handle complex noise in DAS seismic data; <xref ref-type="bibr" rid="B6">Dong et al. (2021)</xref> first introduced Transformer for random noise suppression in seismic data, with its global modeling capability significantly outperforming traditional local filtering methods; <xref ref-type="bibr" rid="B28">Xiang et al. (2025)</xref> proposed a five-dimensional intelligent denoising framework via Transformer, extending its application to high-dimensional seismic data denoising and achieving efficient noise suppression in complex spatiotemporal domains. However, existing research has primarily focused on conventional seismic exploration data or natural earthquake signals, and has not yet systematically explored the denoising potential of Transformer for microseismic signals, which are characterized by lower signal-to-noise ratios and more complex noise spectra.</p>
<p>To address this issue, this study proposes a network combining wavelet transform and Transformer for denoising microseismic signals (<xref ref-type="bibr" rid="B16">Liu and Yang, 2025</xref>). This model combines the excellent time-frequency localization capability of wavelet transform with the advantages of the Transformer in capturing long-range correlations. WTTnet first applies a shallow feature extraction module, combining a convolutional layer with depthwise separable convolutions, to initially extract signal features. It then performs deeper feature extraction through depthwise separable convolutions and a block that integrates wavelet transform with the Transformer. Specifically, DWT decomposes the input signal into high-frequency and low-frequency components, which are concatenated to form full-frequency domain features. During feature fusion, these full-frequency features serve as the query and value vectors for the Transformer module, while the high-frequency components act as the key vector. The frequency-domain results are then transformed back to the time domain using IDWT. Finally, a reconstruction module, consisting of two convolutional layers and activation functions, combines the deep feature-extracted output with the original input to reconstruct the denoised signal. The core advantage of this approach lies in its ability to enable the model to recognize and differentiate between noise components of different frequencies. Experimental results on both synthetic and field data demonstrate that the proposed method effectively suppresses noise while preserving the signal&#x2019;s detailed features.</p>
</sec>
<sec sec-type="methods" id="s2">
<label>2</label>
<title>Methods</title>
<p>Microseismic signals can be considered as the superposition of effective signals and noise, represented as <xref ref-type="disp-formula" rid="e1">Equation 1</xref>:<disp-formula id="e1">
<mml:math id="m2">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents the clean signal, <inline-formula id="inf3">
<mml:math id="m4">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents seismic noise, and <inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the noisy signal.</p>
<p>WTTnet learns to separate the clean signal <inline-formula id="inf5">
<mml:math id="m6">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> from seismic noise <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> by establishing a nonlinear mapping between the noisy signal <inline-formula id="inf7">
<mml:math id="m8">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and the clean signal <inline-formula id="inf8">
<mml:math id="m9">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. Based on this, by feeding the noisy signal <inline-formula id="inf9">
<mml:math id="m10">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> into the network, inference prediction can provide an estimated signal:<disp-formula id="e2">
<mml:math id="m11">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>W</mml:mi>
<mml:mi>T</mml:mi>
<mml:mi>T</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>t</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e2">Equation 2</xref> <inline-formula id="inf10">
<mml:math id="m12">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf11">
<mml:math id="m13">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> denote the denoised signal and training parameters, respectively.</p>
<sec id="s2-1">
<label>2.1</label>
<title>Wavelet Transform-Transformer Block</title>
<p>The wavelet transform converts microseismic signals from the time domain to the frequency domain, offering excellent time-frequency localization capabilities. This ability enables multi-scale refinement analysis and represents the signal as different frequency components, aiding in the extraction of fine features of the microseismic signal. The Transformer model, with its advantages in capturing long-range dependencies, further improves the signal&#x2019;s representational capability. This study proposes a Wavelet Transform-Transformer Block (WTTB), with the specific structure shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The block first performs DWT using Haar wavelets, decomposing the signal into low-frequency components (Low) containing the main information and high-frequency components (High) containing detailed information. Both components then undergo feature transformation through convolution layers with a kernel size of <inline-formula id="inf12">
<mml:math id="m14">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, where the transformed low-frequency and high-frequency components form a Full feature map containing the entire frequency-domain information. In the self-attention mechanism, the Full feature map is used as both the query and value vectors, while the high-frequency components are used as the key vectors. The rationale is that high-frequency components often contain more noise and fine details, while low-frequency components carry the main signal structure. Using high-frequency components as keys allows the Transformer to focus on distinguishing between noise and fine signal features. The full-frequency features as queries and values provide a comprehensive context for this discrimination process. The model generates a new feature map that integrates both global and local information through multi-head self-attention. This feature map is then processed with IDWT to reconstruct the full-frequency domain information back into the time domain. Finally, the feature map is further refined through a convolution layer with a kernel size of <inline-formula id="inf13">
<mml:math id="m15">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, generating the final output feature map. This block effectively combines the excellent time-frequency localization and multi-scale analysis capabilities of wavelet transform with the powerful long-range dependency capture of the Transformer, thereby significantly enhancing the feature representation and denoising performance for microseismic signals.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Structure of the wavelet transform-transformer block.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g001.tif">
<alt-text content-type="machine-generated">Flowchart diagram showing a neural network architecture starting with DWT, splitting into Low and High branches processed by Conv1d, merging as Full, forming Query, Value, and Key for Multi-head Attention, followed by IDWT and Conv1d.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2-2">
<label>2.2</label>
<title>Neural network</title>
<p>As shown in <xref ref-type="fig" rid="F2">Figure 2</xref>, the proposed model is composed of three key components: a shallow feature extraction module (SFEM), a deep feature extraction module (DFEM), and a reconstruction module.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Architecture of WTTnet. <bold>(a)</bold> DSConv. <bold>(b)</bold> DEFB.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g002.tif">
<alt-text content-type="machine-generated">Block diagram illustrating a neural network architecture for waveform processing, comprising three main modules: SFEM with Conv 1D and DSConv, DFEM with repeated DEFB blocks, and a Reconstruction section with Conv 1D and GELU layers. Insets detail DSConv containing DConv, Batch Norm, GELU, Conv 1D, and DEFB containing DSConv and WTTB components, showing internal layer structure and connections.</alt-text>
</graphic>
</fig>
<p>The SFEM begins with a standard convolutional layer, followed by a depthwise separable convolution (DSConv). Initially, a <inline-formula id="inf14">
<mml:math id="m16">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolution projects the input from a single channel to 64 channels. The DSConv structure, depicted in <xref ref-type="fig" rid="F2">Figure 2a</xref>, includes a depthwise convolution, a pointwise convolution, and a non-linear activation. The depthwise convolution, using a <inline-formula id="inf15">
<mml:math id="m17">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> kernel, performs spatial convolutions independently on each input channel. The pointwise convolution (<inline-formula id="inf16">
<mml:math id="m18">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> kernel) then integrates the outputs along the channel dimension. To accelerate convergence, stabilize training, improve non-linearity, and reduce overfitting, Batch Normalization and GELU activation are applied after both convolution stages. The shallow feature extraction process is represented as <xref ref-type="disp-formula" rid="e3">Equation 3</xref>:<disp-formula id="e3">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>where <inline-formula id="inf17">
<mml:math id="m20">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the noisy signal and <inline-formula id="inf18">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the extracted shallow features.</p>
<p>The DFEM consists of four deep feature extraction blocks (DFEB), as illustrated in <xref ref-type="fig" rid="F2">Figure 2b</xref>. Each DFEB contains four DSConvs, a WTTB, and residual connections. The DSConvs are configured identically to those in the shallow feature extraction module. Each DSConv layer processes the input data through depthwise spatial convolutions, effectively extracting local temporal information. By stacking multiple DSConv layers, the model incrementally abstracts features, allowing it to capture increasingly complex signal characteristics. The WTTB further strengthens the model&#x2019;s capacity to extract informative features through a multi-scale attention mechanism. Additionally, the multi-level residual architecture promotes stable training of the deep network. The deep feature extraction process is defined as <xref ref-type="disp-formula" rid="e4">Equation 4</xref>:<disp-formula id="e4">
<mml:math id="m22">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>F</mml:mi>
<mml:mi>E</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mo>&#x22ef;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>F</mml:mi>
<mml:mi>E</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf19">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the learned deep features.</p>
<p>The reconstruction module comprises two <inline-formula id="inf20">
<mml:math id="m24">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolutional layers followed by GELU activations. This module integrates the deep features with the original input signal, preserving key characteristics while reconstructing the denoised signal <inline-formula id="inf21">
<mml:math id="m25">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula>. The reconstruction process is outlined as <xref ref-type="disp-formula" rid="e5">Equation 5</xref>:<disp-formula id="e5">
<mml:math id="m26">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mi>E</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>v</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>The mean absolute error (MAE) is used as the loss function:<disp-formula id="e6">
<mml:math id="m27">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e6">Equation 6</xref> <inline-formula id="inf22">
<mml:math id="m28">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> denotes the denoised signal and <inline-formula id="inf23">
<mml:math id="m29">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the clean signal.</p>
</sec>
<sec id="s2-3">
<label>2.3</label>
<title>Parameters selection</title>
<p>The number of DWT levels in WTTnet was set to five, a choice grounded in the dominant frequency distribution of typical microseismic signals and the spectral characteristics of noise. Too few decomposition levels would fail to adequately separate low-frequency useful signals from low-frequency noise, whereas too many levels would produce excessively sparse high-frequency sub-bands, impairing the preservation of fine details. A five-level decomposition sufficiently covers the main frequency range of microseismic signals while providing a structurally clear multi-scale representation for the subsequent attention mechanism. The model adopts an eight head self attention mechanism to balance computational efficiency with sufficient feature interaction. Each attention head can be regarded as an independent feature discriminator that simultaneously examines different frequency band subspaces or distinct time frequency regions, thereby modeling signal noise relationships. For a microseismic sequence length of 2000 samples, eight heads effectively capture long-range dependencies across scales while avoiding the parameter redundancy and overfitting risks associated with an excessive number of attention heads. The dimensionality of each attention head in the Transformer was set to 64. When combined with eight attention heads, the resulting total hidden dimension achieves an optimal trade-off between modeling the full signal length and maintaining computational efficiency.</p>
<p>The proposed WTTnet framework exhibits strong potential for cross platform generalization to other geophysical data modalities, owing to its core design that addresses universal challenges in geophysical signal processing: multi scale decomposition, non stationary feature representation, and noise signal separation in the time frequency domain. By adjusting the core parameters of WTTnet based on the dominant frequency distribution, spectral concentration of noise, sequence length, and other characteristics of the target data, the model can be adapted to meet the demands of various tasks ranging from natural earthquake monitoring to exploration seismology. For instance, in natural earthquake monitoring, the same architecture could enhance weak phase arrivals obscured by cultural or oceanic noise. In exploration seismology, the model could be retrained to suppress coherent noise in reflection seismic data, where the wavelet based frequency separation and attention based spatial temporal modeling would help distinguish between primary reflections and noise trains.</p>
</sec>
<sec id="s2-4">
<label>2.4</label>
<title>Experiment details</title>
<p>The model is optimized using the Adam optimizer. The initial learning rate is set to <inline-formula id="inf24">
<mml:math id="m30">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and is adaptively reduced by a factor of <inline-formula id="inf25">
<mml:math id="m31">
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mn>0.1</mml:mn>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:math>
</inline-formula> whenever the validation loss fails to decrease after 10 epochs, until it reaches a minimum value of <inline-formula id="inf26">
<mml:math id="m32">
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. The total number of training epochs is set to 200. All experiments are conducted on an NVIDIA RTX 4090 GPU (24 GB) and a Xeon Gold 6430R CPU.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Results &#x26; discussion</title>
<sec id="s3-1">
<label>3.1</label>
<title>Synthetic data</title>
<p>To assess the effectiveness of the proposed method, we first conduct evaluations on synthetic seismic data. A portion of the Marmousi velocity model is chosen to simulate a microseismic monitoring region, as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. The model spans <inline-formula id="inf27">
<mml:math id="m33">
<mml:mrow>
<mml:mn>1280</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mi>m</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1280</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> with a grid resolution of 10 m. Synthetic events are generated using a 50 Hz source frequency, with 500 randomly selected grid points serving as event origins. A total of 128 receivers are uniformly distributed across the surface, spaced 10 m apart, and the data is sampled at 1 m intervals for 2000 time steps per trace. The data are normalized to the range of [-1, 1], and a microseismic dataset is created by randomly adding Gaussian noise, colored noise, electromagnetic interference (EMI) noise, and mixed noise to the signals. Gaussian noise obeys a zero mean normal distribution, where the standard deviation is determined by a randomly drawn SNR value within [-10,5] dB. Colored noise is generated by passing Gaussian white noise through a 1/f spectral shaping filter. The fundamental frequency of 50 Hz is adopted as the EMI noise source. Its amplitude varies randomly between 0.1 and 0.3 times the amplitude of the clean signal to simulate the fluctuating intensity of interference in real-world environments. Mixed noise is produced by randomly selecting two or three of the above noise types and superimposing them with random relative energy ratios, where each constituent noise contributes (20%)&#x2013;(60%) of the total noise energy. The resulting dataset consists of 64,000 one-channel waveform samples, each 2000 time steps in length. The 64,000 samples were evenly distributed across the four noisy signal categories. The selection of 2000 time steps per trace was determined based on a trade-off between capturing sufficient waveform duration for microseismic events and maintaining computational efficiency during training. This length not only covers the complete event window in our synthetic setup but also helps reduce the model&#x2019;s computational complexity and improve overall efficiency. The data are randomly divided into training (75%), validation (15%), and test (10%) sets.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Experimental velocity model.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g003.tif">
<alt-text content-type="machine-generated">Contour plot showing seismic velocity distribution with depth on the vertical axis and horizontal distance on the horizontal axis. Color scale ranges from purple to yellow, representing velocities from 2.0 to over 4.0 kilometers per second.</alt-text>
</graphic>
</fig>
<p>We input the test set into the trained model to obtain experimental results. <xref ref-type="fig" rid="F4">Figure 4a</xref> presents a microseismic signal contaminated with Gaussian random noise, and <xref ref-type="fig" rid="F4">Figure 4b</xref> shows its corresponding frequency spectrum. As illustrated, the original noisy signal exhibits significant overlap between signal and noise in both domains, with the frequency spectrum completely dominated by the flat Gaussian noise. After processing with WTTnet, the time-domain waveform becomes clear and coherent, with virtually no residual noise visible. In the frequency domain, the signal spectrum is well-recovered, closely matching the clean signal&#x2019;s spectral characteristics. Gaussian noise is stationary and exhibits a flat spectrum, which allows wavelet transforms to effectively separate the signal from the noise in the frequency domain. The self-attention mechanism of the Transformer can capture global dependencies, further suppressing noise. By decomposing the signal into different frequency bands using wavelet transform, WTTnet can then model the relationships between these frequency bands using the Transformer, effectively removing Gaussian noise.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Denoising results with Gaussian random noise. <bold>(A)</bold> Time domain and <bold>(B)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WTTnet.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g004.tif">
<alt-text content-type="machine-generated">Panel A shows four line plots of normalized amplitude versus sampling point for clean signal, Gaussian noise, noisy signal, and denoised signal. Panel B presents corresponding spectrograms with frequency in hertz, illustrating spectral differences among the signals.</alt-text>
</graphic>
</fig>
<p>
<xref ref-type="fig" rid="F5">Figure 5</xref> illustrates a microseismic signal contaminated by colored noise and its corresponding frequency spectrum. The introduced colored noise is concentrated below 50 Hz, directly overlapping with the dominant frequency band of the effective microseismic signal. This presents a particular challenge as the noise shares spectral content with the signal of interest. The original noisy signal displays obvious contamination in both time and frequency domains, with the low-frequency colored noise obscuring the fundamental components of the microseismic signal. WTTnet effectively restores the signal&#x2019;s original waveform, producing a clean time-domain trace with minimal distortion. The processed frequency spectrum shows successful separation, with the low-frequency noise components significantly attenuated while preserving the signal&#x2019;s spectral integrity. Colored noise typically exhibits specific frequency distribution characteristics, which wavelet transform can separate into low- and high-frequency components. WTTnet processes the low- and high-frequency components separately and uses the self-attention mechanism to learn the correlations between them. This enables the model to specifically reduce the impact of noise at certain frequencies while retaining the useful frequency components of the signal.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Denoising results with colored noise. <bold>(A)</bold> Time domain and <bold>(B)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WTTnet.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g005.tif">
<alt-text content-type="machine-generated">Figure with two panels labeled A and B. Panel A shows four line graphs of normalized amplitude versus sampling point for clean signal, colored noise, noisy signal, and denoised signal from top to bottom. Panel B shows corresponding spectrograms, with frequency in hertz versus sampling point, illustrating signal clarity and denoising effects.</alt-text>
</graphic>
</fig>
<p>
<xref ref-type="fig" rid="F6">Figure 6</xref> shows a microseismic signal with EMI noise and its corresponding frequency spectrum. EMI noise typically manifests as narrow-band interference or periodic spikes, often originating from electrical equipment in field environments. The original signal is heavily corrupted, with the EMI components distorting both the amplitude and phase characteristics of the underlying seismic event. After WTTnet processing, the time-domain waveform appears clean with the oscillatory interference removed, and the signal&#x2019;s natural morphology is well-preserved. The frequency spectrum becomes organized, with the narrow-band EMI peaks effectively suppressed. EMI noise is challenging due to its structured nature and potential similarity to genuine signal features. Wavelet transform provides multi-resolution analysis to localize these frequency-specific interferences, while the attention mechanism of the Transformer helps to distinguish between persistent EMI patterns and transient seismic signals. Through this combined approach, structured EMI noise can be effectively identified and removed.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Denoising results with EMI noise. <bold>(A)</bold> Time domain and <bold>(B)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WTTnet.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g006.tif">
<alt-text content-type="machine-generated">Figure showing four time series plots in panel A (clean signal, EMI noise, noisy signal, denoised signal) and corresponding spectrograms in panel B, illustrating changes in signal clarity and frequency content before and after denoising.</alt-text>
</graphic>
</fig>
<p>In <xref ref-type="fig" rid="F7">Figure 7</xref>, a microseismic signal is contaminated with a mixture of the three aforementioned types of noise. <xref ref-type="fig" rid="F7">Figure 7b</xref> displays its corresponding frequency spectrum. The complex noise mixture creates severe signal distortion in both domains, making it difficult to distinguish the original microseismic event. WTTnet successfully recovers the signal&#x2019;s essential features, producing a clean time-domain waveform with excellent fidelity. The denoised frequency spectrum exhibits a clear, organized structure that facilitates accurate event analysis. Mixed noise contains various noise characteristics, and WTTnet&#x2019;s multi-scale processing capabilities and self-attention mechanism enable it to handle different types of noise simultaneously. The wavelet decomposition separates the signal into different frequency bands, allowing the model to process noise in each band independently, while the Transformer integrates information from multiple bands, providing a more comprehensive denoising solution.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Denoising results with mixed noise. <bold>(A)</bold> Time domain and <bold>(B)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WTTnet.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g007.tif">
<alt-text content-type="machine-generated">Panel A shows four line graphs of signals labeled clean, mixed noise, noisy, and denoised, with normalized amplitude versus sampling point. Panel B displays corresponding time-frequency spectrograms for each signal, showing frequency in hertz against sampling point.</alt-text>
</graphic>
</fig>
<p>The algorithm restores the denoised signal with high accuracy, effectively suppressing common types of noise encountered in field operations, such as Gaussian random noise and various environmental noises. The shape and amplitude characteristics of the predicted signal exhibit a high degree of similarity to the original signal, with the overall frequency spectrum remaining clean and clear, ensuring excellent fidelity. This high fidelity in signal reconstruction is crucial for ensuring that the restored signal retains the essential features of the original, which is vital for accurate analysis and interpretation of microseismic data in practical applications.</p>
<p>Subsequently, the SNR of the original noisy data and the denoised data is calculated. The formula for calculating the SNR is as <xref ref-type="disp-formula" rid="e7">Equation 7</xref>:<disp-formula id="e7">
<mml:math id="m34">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>10</mml:mn>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>g</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>10</mml:mn>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>g</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msubsup>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>where <inline-formula id="inf28">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf29">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denote the signal and noise power, respectively. <inline-formula id="inf30">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the clean signal, and <inline-formula id="inf31">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the denoised signal.</p>
<p>As shown in <xref ref-type="fig" rid="F8">Figure 8</xref>, the noisy signals are primarily distributed in the &#x2212;7 to 4 dB range, reflecting signals buried at the noise threshold. In contrast, the denoised outputs predominantly lie between 5 and 17 dB. The proposed method boosts the average SNR from &#x2212;0.225 dB to 11.556 dB, highlighting its strong denoising capability.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>SNR of noisy signals and signals denoised using the proposed method.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g008.tif">
<alt-text content-type="machine-generated">Histogram comparing SNR in decibels for noisy signal and denoised signal using the proposed method. The noisy signal distribution peaks at lower SNR values near negative five, while the denoised version shifts right, peaking above ten SNR.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-2">
<label>3.2</label>
<title>Comparison with other methods</title>
<p>The proposed method is compared to WT and Swin Transformer (<xref ref-type="bibr" rid="B23">Sun et al., 2025</xref>; <xref ref-type="bibr" rid="B14">Li et al., 2024</xref>). For WT, we employ the Haar wavelet with 5 decomposition levels. Swin Transformer was implemented as a 1D version comprising 4 stages, each containing 2 Transformer blocks, with a window size of 8, 8 attention heads, and a hidden dimension of 64. To ensure a fair comparison under identical conditions, Swin Transformer was trained using the same experimental setup as WTTnet. The Adam optimizer was employed, and the MAE was adopted as the loss function. The model was trained for 200 epochs on the same hardware platform as WTTnet, using the same learning rate schedule and batch size. The results are shown in <xref ref-type="fig" rid="F9">Figure 9</xref>. For the original noisy signal, the time-domain waveform clearly displays significant distortion and confusion, with a heavy overlap between the signal and noise. In the frequency domain, this overlap is also quite evident in the spectrum. When comparing the denoising performance of the three methods, it becomes apparent that the WT method performs the worst. While it does reduce noise to some extent, the improvement in SNR is relatively modest. The Swin Transformer demonstrates improved performance; its denoised time-domain waveform shows minimal visible noise. However, residual high-frequency noise is still present in the frequency domain. In contrast, WTTnet shows a clear advantage over the other methods in both signal recovery and noise reduction. The time-domain waveform is notably clean and coherent, with almost no remaining noise. The frequency spectrum is also organized and well-defined, with high-frequency noise completely removed, facilitating more accurate event detection. From the frequency domain comparison, it is evident that the Swin Transformer&#x2019;s self-attention mechanism, which struggles to effectively distinguish different frequency components, leaves some high-frequency noise in the denoised signal. On the other hand, WTTnet leverages DWT to decompose the input signal into both high-frequency and low-frequency components. It then applies a multi-head self-attention mechanism to model the interdependencies between cross-scale features, allowing for more effective extraction of low-frequency signals while efficiently removing high-frequency noise. This significantly enhances the overall denoising performance.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Denoising results of synthetic data. <bold>(A)</bold> Time domain and <bold>(B)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WT, Transformer, and the proposed method.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g009.tif">
<alt-text content-type="machine-generated">Panel A displays five line graphs of normalized amplitude versus sampling point, comparing clean, noisy, and denoised signals using WT, Swin Transformer, and WTTnet. Panel B shows corresponding spectrograms for each signal, with frequency on the y-axis and sampling point on the x-axis, illustrating noise reduction effectiveness.</alt-text>
</graphic>
</fig>
<p>To further evaluate denoising performance, we employ three metrics: SNR, MAE, and the structural similarity index (SSIM). While the SNR formula was introduced in the previous section, the MAE and SSIM formulations are provided <xref ref-type="disp-formula" rid="e8">Equations 8</xref>, <xref ref-type="disp-formula" rid="e9">9</xref>:<disp-formula id="e8">
<mml:math id="m39">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>
<disp-formula id="e9">
<mml:math id="m40">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>M</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf32">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf33">
<mml:math id="m42">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the means of the clean and denoised signals, <inline-formula id="inf34">
<mml:math id="m43">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf35">
<mml:math id="m44">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> are their variances, <inline-formula id="inf36">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the covariance, and <inline-formula id="inf37">
<mml:math id="m46">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf38">
<mml:math id="m47">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are stabilization constants.</p>
<p>As shown in <xref ref-type="fig" rid="F10">Figure 10</xref>, the proposed denoising method outperforms both the Swin Transformer and WT methods. Compared to the other two approaches, WTTnet achieves the greatest improvement in SNR. The average improvement is 7.604 dB over traditional wavelet denoising and approximately 4.382 dB over the Swin Transformer method. The higher SSIM values in <xref ref-type="fig" rid="F10">Figure 10b</xref> further indicate that the signal processed by WTTnet maintains greater similarity to the original clean signal, suggesting that the method introduces minimal waveform distortion during the denoising process. The lowest average MAE value in <xref ref-type="fig" rid="F10">Figure 10c</xref> demonstrates that WTTnet produces the least overall deviation during signal reconstruction. This not only reflects the effectiveness of the denoising but also highlights its exceptional performance in signal reconstruction accuracy. The lower error indicates that the denoised signal is closer to the true seismic waveform, providing a reliable data foundation for subsequent precise analysis and interpretation.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Comparison of the denoising performance of WT, Transformer and WTTnet. <bold>(a)</bold> SNR, <bold>(b)</bold> MAE and <bold>(c)</bold> SSIM.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g010.tif">
<alt-text content-type="machine-generated">Three line charts comparing WTTnet, Swin Transformer, and WT on denoising performance across SNR before denoising values. Chart a shows SNR after denoising, with WTTnet highest. Chart b shows MAE, with WTTnet lowest. Chart c shows SSIM, with WTTnet highest.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-3">
<label>3.3</label>
<title>Application to earthquake detection</title>
<p>Improving signal quality has a profound and widespread impact on subsequent microseismic data processing. The accurate determination of the P-wave arrival time is a fundamental step in the localization of microseismic events and the calculation of source parameters. The precision of this key parameter directly determines the reliability of all subsequent analyses. Additionally, the presence of non-seismic signals significantly increases the false positive detection rate. These interfering signals, which may stem from mining equipment operation, vehicle vibrations, electrical interference, or other human activities, often share similar characteristics with microseismic signals, making it difficult for traditional detection methods to effectively distinguish them.</p>
<p>In <xref ref-type="fig" rid="F11">Figure 11</xref>, we showcase two typical examples demonstrating the superior performance of the WTTnet model in enhancing the STA/LTA characteristic function. Traditional STA/LTA methods face significant challenges when processing noisy signals, as noise interference can create false peaks in the characteristic function, making accurate identification of the true P-wave arrival time difficult. Meanwhile, weak signals hidden under noise are prone to being missed, with this issue being particularly evident in low SNR environments. In the first example, a weak useful signal was originally submerged in strong background noise, and the STA/LTA algorithm failed to produce a clear triggering signal. However, after enhancement with WTTnet, the SNR of the signal was significantly improved, and the characteristic function showed a distinct upward slope at the P-wave arrival time, providing a reliable basis for automatic detection. In the second example, noise in the signal caused fluctuations in the STA/LTA ratio, generating multiple false peaks that severely disrupted accurate identification of the P-wave. After processing with WTTnet, these non-seismic noise components were effectively suppressed, and the characteristic function curve became smooth and clear, with the peak corresponding to the true P-wave arrival time standing out and being easily identifiable.</p>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Improvement of microseismic detection after denoising. <bold>(a)</bold> and <bold>(b)</bold> two examples of earthquake detection. (i) Noisy signals. (ii) Denoised signals. (iii) and (iv) Corresponding STA/LTA characteristic functions.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g011.tif">
<alt-text content-type="machine-generated">Two panels labeled &#x22;a&#x22; and &#x22;b&#x22; each display four graphs comparing noisy and denoised seismic signals, showing amplitude versus sampling point and STA/LTA ratio versus sampling point. Denoised signals exhibit clearer peaks and reduced background noise.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-4">
<label>3.4</label>
<title>Field data</title>
<p>To further evaluate the method&#x2019;s generalization and field applicability, we test it on field microseismic data from the 2018 hydraulic fracturing experiment at the Grimsel Test Site, as reported by <xref ref-type="bibr" rid="B11">Gischig et al. (2018)</xref>. The monitoring system includes 28 piezoelectric sensors and 4 accelerometers, with 20 sensors installed along the tunnel wall and four accelerometers co-located. An 8-channel borehole array with 1 m spacing is placed in borehole SBH-1, which consistently records all events due to its low-noise environment. The shortest sensor-to-source distance is approximately 9 m (from HF1), while the furthest sensors (S1-S5) are 55&#x2013;72 m away. Data acquisition uses a 32-channel digitizer, with band-pass filters at 1000 Hz and 50 Hz. A representative trace containing a microseismic event is preprocessed, and the model trained on synthetic data is directly applied.</p>
<p>The trained model was directly applied to field data for denoising. As shown in <xref ref-type="fig" rid="F12">Figures 12</xref>, <xref ref-type="fig" rid="F13">13</xref>, the original microseismic recordings are heavily contaminated by various types of complex noise, with the effective signals largely buried, making it difficult to directly identify microseismic events. The WT method exhibits relatively poor denoising performance, failing to effectively separate signal from noise. The denoised waveform in the time domain still contains significant interference components, and the frequency domain features show no notable improvement. The Swin Transformer demonstrates better performance in the denoising task. Its denoised waveform in the time domain shows minimal visible noise components. However, frequency domain analysis reveals that some low-amplitude residual noise remains in the waveform following the end of the microseismic event. WTTnet delivers the most impressive results in both signal recovery and noise suppression. It not only effectively restores the original waveform characteristics, maintaining continuity and clarity in the time domain, but also demonstrates strong noise suppression capabilities in the frequency domain. The resulting spectral distribution becomes more focused and structured, significantly enhancing the feasibility of subsequent event extraction and precise analysis.</p>
<fig id="F12" position="float">
<label>FIGURE 12</label>
<caption>
<p>Denoising results of field data 1. <bold>(a)</bold> Time domain and <bold>(b)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WT, Transformer, and the proposed method.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g012.tif">
<alt-text content-type="machine-generated">Panel a displays four line charts comparing normalized amplitude versus sampling point for a field signal and signals denoised using WT, Swin Transformer, and WTTnet methods. Panel b presents corresponding spectrograms, showing changes in frequency content for each signal, illustrating noise reduction effectiveness.</alt-text>
</graphic>
</fig>
<fig id="F13" position="float">
<label>FIGURE 13</label>
<caption>
<p>Denoising results of field data 2. <bold>(a)</bold> Time domain and <bold>(b)</bold> Time-frequency domain of clean signal, noisy signal, and denoised signal using WT, Transformer, and the proposed method.</p>
</caption>
<graphic xlink:href="feart-14-1790572-g013.tif">
<alt-text content-type="machine-generated">Panel a shows four line graphs of normalized amplitude versus sampling point for field signal denoising methods, while panel b shows corresponding heatmaps of signal frequency content, with labels for each method including WT, Swin Transformer, and WTTnet.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-5">
<label>3.5</label>
<title>Limitations and future considerations</title>
<p>In terms of computational efficiency, the Transformer&#x2019;s self-attention mechanism has a time complexity of <inline-formula id="inf39">
<mml:math id="m48">
<mml:mrow>
<mml:mi mathvariant="bold">O</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn mathvariant="bold">2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, which results in significantly higher computational costs compared to CNN-based parallel sliding window methods during training. However, compared with traditional methods, deep learning models can perform rapid inference once trained, which offers a significant advantage for real-time microseismic monitoring applications that require low-latency response. Nevertheless, optimizing real-time processing latency on edge computing devices, such as downhole microseismic acquisition instruments, remains a key direction for future research. Another important consideration relates to the model&#x2019;s generalization capability across diverse geological settings.</p>
<p>Although the proposed method performs well on both synthetic data and the Grimsel test site dataset, its generalizability across diverse geological and acquisition conditions requires further validation. The training data in this study are primarily based on synthetic velocity models and a single field dataset, which possess specific noise characteristics, signal frequency content, and geological responses. In practice, factors such as varying lithology, fracture density, depth, and acquisition systems can significantly alter microseismic signal patterns and noise structures, potentially leading to degraded model performance. Future work should test and fine-tune the model on broader and more diverse field datasets, or explore the integration of physics-based constraints to enhance its robustness and applicability in unseen geological and noisy environments.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s4">
<label>4</label>
<title>Conclusion</title>
<p>This study proposes WTTnet, a microseismic denoising method that combines wavelet transform and Transformer. WTTnet separates high- and low-frequency components of the signal via DWT and leverages the self-attention mechanism to recognize and differentiate between noise components of different frequencies. The high- and low-frequency components are combined to form full-frequency domain features. During feature fusion, full-frequency features serve as query and value vectors for the Transformer, with high-frequency components acting as key vectors. The multi-head self-attention mechanism models cross-scale relationships, and the frequency-domain results are converted back to the time domain using IDWT. This architecture enables the model to effectively recognize and distinguish noise components of different frequencies. Experiments on both synthetic and field data demonstrate that this method outperforms traditional denoising techniques in terms of denoising performance.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>WS: Visualization, Validation, Data curation, Conceptualization, Writing &#x2013; original draft, Methodology, Writing &#x2013; review and editing. SY: Funding acquisition, Writing &#x2013; review and editing, Project administration, Supervision. JW: Project administration, Writing &#x2013; review and editing, Funding acquisition, Supervision.</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Allen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Rabiner</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>1977</year>). <article-title>A unified approach to short-time fourier analysis and synthesis</article-title>. <source>Proc. IEEE</source> <volume>65</volume>, <fpage>1558</fpage>&#x2013;<lpage>1564</lpage>. <pub-id pub-id-type="doi">10.1109/PROC.1977.10770</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Duan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Multiscale dilated denoising convolution with channel attention mechanism for micro-seismic signal denoising</article-title>. <source>J. Pet. Explor. Prod. Technol.</source> <volume>14</volume>, <fpage>883</fpage>&#x2013;<lpage>908</lpage>. <pub-id pub-id-type="doi">10.1007/s13202-024-01752-4</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fomel</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Emd-seislet transform</article-title>. <source>Geophysics</source> <volume>83</volume>, <fpage>A27</fpage>&#x2013;<lpage>A32</lpage>. <pub-id pub-id-type="doi">10.1190/geo2017-0554.1</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Saad</surname>
<given-names>O. M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Savvaidis</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Deep learning for seismic data compression in distributed acoustic sensing</article-title>. <source>IEEE Trans. Geosci. Remote Sens.</source> <volume>63</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/TGRS.2025.3526933</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cui</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bin Waheed</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Unsupervised deep learning for das-vsp denoising using attention-based deep image prior</article-title>. <source>IEEE Trans. Geosci. Remote Sens.</source> <volume>63</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/TGRS.2025.3533597</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhong</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A deep-learning-based denoising method for multiarea surface seismic data</article-title>. <source>IEEE Geosci. Remote Sens. Lett.</source> <volume>18</volume>, <fpage>925</fpage>&#x2013;<lpage>929</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2020.2989450</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Cong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhong</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Global-feature-fusion and multiscale network for low-frequency extrapolation</article-title>. <source>IEEE Trans. Geoscience Remote Sens.</source> <volume>62</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/TGRS.2024.3408949</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Tong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Bidirectional physics-constrained full-waveform inversion: reducing seismic data dependence in velocity model building</article-title>. <source>Geophys. J. Int.</source> <volume>244</volume>, <fpage>ggaf466</fpage>. <pub-id pub-id-type="doi">10.1093/gji/ggaf466</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cai</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hong</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Research on deep convolutional neural network time-frequency domain seismic signal denoising combined with residual dense blocks</article-title>. <source>Front. Earth Sci.</source> <volume>9</volume>, <fpage>681869</fpage>. <pub-id pub-id-type="doi">10.3389/feart.2021.681869</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gao</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2025</year>). <article-title>Mmgpt4lf: leveraging an optimized pre-trained gpt-2 model with multi-modal cross-attention for load forecasting</article-title>. <source>Appl. Energy</source> <volume>392</volume>, <fpage>125965</fpage>. <pub-id pub-id-type="doi">10.1016/j.apenergy.2025.125965</pub-id>
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gischig</surname>
<given-names>V. S.</given-names>
</name>
<name>
<surname>Doetsch</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Maurer</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Krietsch</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Amann</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Evans</surname>
<given-names>K. F.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>On the link between stress field and small-scale hydraulic fracture growth in anisotropic rock derived from microseismicity</article-title>. <source>Solid Earth.</source> <volume>9</volume>, <fpage>39</fpage>&#x2013;<lpage>61</lpage>. <pub-id pub-id-type="doi">10.5194/se-9-39-2018</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wan</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Mine microseismic signal denoising based on a deep convolutional autoencoder</article-title>. <source>Shock Vib.</source> <volume>2023</volume>, <fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1155/2023/6225923</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Iqbal</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Deepseg: deep segmental denoising neural network for seismic data</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst.</source> <volume>34</volume>, <fpage>3397</fpage>&#x2013;<lpage>3404</lpage>. <pub-id pub-id-type="doi">10.1109/TNNLS.2022.3205421</pub-id>
<pub-id pub-id-type="pmid">36150003</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Swin transformer for seismic denoising</article-title>. <source>IEEE Geosci. Remote Sens. Lett.</source> <volume>21</volume>, <fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2024.3358234</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Research on microseismic denoising method based on cbdnet</article-title>. <source>Artif. Intell. Geosci.</source> <volume>4</volume>, <fpage>28</fpage>&#x2013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1016/j.aiig.2023.02.002</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Wtt: combining wavelet transform with transformer for remote sensing image super-resolution</article-title>. <source>Mach. Vis. Appl.</source> <volume>36</volume>, <fpage>35</fpage>. <pub-id pub-id-type="doi">10.1007/s00138-024-01655-8</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zu</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>An effective approach to attenuate random noise based on compressive sensing and curvelet transform</article-title>. <source>J. Geophys. Eng.</source> <volume>13</volume>, <fpage>135</fpage>&#x2013;<lpage>145</lpage>. <pub-id pub-id-type="doi">10.1088/1742-2132/13/2/135</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mousavi</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Beroza</surname>
<given-names>G. C.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Deep-learning seismology</article-title>. <source>Science</source> <volume>377</volume>, <fpage>eabm4470</fpage>. <pub-id pub-id-type="doi">10.1126/science.abm4470</pub-id>
<pub-id pub-id-type="pmid">35951699</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mousavi</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Langston</surname>
<given-names>C. A.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Hybrid seismic denoising using higher-order statistics and improved wavelet block thresholding</article-title>. <source>Bull. Seismol. Soc. Amer.</source> <volume>106</volume>, <fpage>1380</fpage>&#x2013;<lpage>1393</lpage>. <pub-id pub-id-type="doi">10.1785/0120150345</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mousavi</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Langston</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Horton</surname>
<given-names>S. P.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Automatic microseismic denoising and onset detection using the synchrosqueezed continuous wavelet transform</article-title>. <source>Geophysics</source> <volume>81</volume>, <fpage>V341</fpage>&#x2013;<lpage>V355</lpage>. <pub-id pub-id-type="doi">10.1190/geo2015-0598.1</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Saad</surname>
<given-names>O. M.</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Uncovering the microseismic signals from noisy data for high-fidelity 3d source-location imaging using deep learning</article-title>. <source>Geophysics</source> <volume>86</volume>, <fpage>KS161</fpage>&#x2013;<lpage>KS173</lpage>. <pub-id pub-id-type="doi">10.1190/geo2021-0021.1</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Saad</surname>
<given-names>O. M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Siervo</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Savvaidis</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>G.-c. D.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>Eqcct: a production-ready earthquake detection and phase-picking method using the compact convolutional transformer</article-title>. <source>IEEE Trans. Geosci. Remote Sens.</source> <volume>61</volume>, <fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1109/TGRS.2023.3319440</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sun</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Gong</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Removing random noise and improving the resolution of seismic data using deep-learning transformers</article-title>. <source>Geophys. Prospect.</source> <volume>73</volume>, <fpage>611</fpage>&#x2013;<lpage>627</lpage>. <pub-id pub-id-type="doi">10.1111/1365-2478.13617</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Mao</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Simultaneous reconstruction and denoising for DAS-VSP seismic data by RRU-net</article-title>. <source>Front. Earth Sci.</source> <volume>10</volume>, <fpage>993465</fpage>. <pub-id pub-id-type="doi">10.3389/feart.2022.993465</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waheed</surname>
<given-names>U. B.</given-names>
</name>
<name>
<surname>Di</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Angus</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Introduction to special issue on machine learning applications in geophysical exploration and monitoring</article-title>. <source>Geophys. Prospect.</source> <volume>72</volume>, <fpage>3</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1111/1365-2478.13424</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Shao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Multi-scale interactive network in the application of das seismic data processing</article-title>. <source>Front. Earth Sci.</source> <volume>10</volume>, <fpage>991860</fpage>. <pub-id pub-id-type="doi">10.3389/feart.2022.991860</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>All-in-one weather-degraded image restoration via adaptive degradation-aware self-prompting model</article-title>. <source>IEEE Trans. Multimedia</source> <volume>27</volume>, <fpage>3343</fpage>&#x2013;<lpage>3355</lpage>. <pub-id pub-id-type="doi">10.1109/TMM.2025.3535316</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Five-dimensional intelligent denoising via transformer</article-title>. <source>IEEE Trans. Geoscience Remote Sens.</source> <volume>63</volume>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1109/TGRS.2025.3601354</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Adversarial spatio-temporal learning for video deblurring</article-title>. <source>IEEE Trans. Image Process.</source> <volume>28</volume>, <fpage>291</fpage>&#x2013;<lpage>301</lpage>. <pub-id pub-id-type="doi">10.1109/TIP.2018.2867733</pub-id>
<pub-id pub-id-type="pmid">30176588</pub-id>
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Investigation of microseismic signal denoising using an improved wavelet adaptive thresholding method</article-title>. <source>Sci. Rep.</source> <volume>12</volume>, <fpage>22186</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-26576-2</pub-id>
<pub-id pub-id-type="pmid">36564455</pub-id>
</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3176234/overview">Federico Pichi</ext-link>, International School for Advanced Studies (SISSA), Italy</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1544669/overview">Shiqi Dong</ext-link>, Northeast Electric Power University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3000100/overview">Wenkun Yang</ext-link>, Hohai University, China</p>
</fn>
</fn-group>
</back>
</article>