<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Earth Sci.</journal-id>
<journal-title>Frontiers in Earth Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Earth Sci.</abbrev-journal-title>
<issn pub-type="epub">2296-6463</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1081998</article-id>
<article-id pub-id-type="doi">10.3389/feart.2022.1081998</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Earth Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Deep learning based on self-supervised pre-training: Application on sandstone content prediction</article-title>
<alt-title alt-title-type="left-running-head">Wang et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/feart.2022.1081998">10.3389/feart.2022.1081998</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wang</surname>
<given-names>Chong Ming</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2071603/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wang</surname>
<given-names>Xing Jian</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1916770/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Chen</surname>
<given-names>Yang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wen</surname>
<given-names>Xue Mei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Yong Heng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Qing Wu</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>College of Geophysics</institution>, <institution>Chengdu University of Technology</institution>, <addr-line>Chengdu</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>State Key Laboratory of Oil and Gas Reservoir Geology and Exploitation</institution>, <institution>Chengdu University of Technology</institution>, <addr-line>Chengdu</addr-line>, <addr-line>Sichuan</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1564152/overview">Tariq Alkhalifah</ext-link>, King Abdullah University of Science and Technology, Saudi Arabia</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2092042/overview">Jiefu Chen</ext-link>, University of Houston, United States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1812844/overview">Motaz Alfarraj</ext-link>, King Fahd University of Petroleum and Minerals, Saudi Arabia</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Chong Ming Wang, <email>1352036380@qq.com</email>; Xing Jian Wang, <email>wangxj@cdut.edu.cn</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Solid Earth Geophysics, a section of the journal Frontiers in Earth Science</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>10</day>
<month>01</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>10</volume>
<elocation-id>1081998</elocation-id>
<history>
<date date-type="received">
<day>27</day>
<month>10</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>28</day>
<month>12</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2023 Wang, Wang, Chen, Wen, Zhang and Li.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Wang, Wang, Chen, Wen, Zhang and Li</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Deep learning has been widely used in various fields and showed promise in recent years. Therefore, deep learning is the future trend to realize seismic data&#x2019;s intelligent and automatic interpretation. However, traditional deep learning only uses labeled data to train the model, and thus, does not utilize a large amount of unlabeled data. Self-supervised learning, widely used in Natural Language Processing (NLP) and computer vision, is an effective method of learning information from unlabeled data. Thus, a pretext task is designed with reference to Masked Autoencoders (MAE) to realize self-supervised pre-training of unlabeled seismic data. After pre-training, we fine-tune the model to the downstream task. Experiments show that the model can effectively extract information from unlabeled data through the pretext task, and the pre-trained model has better performance in downstream tasks.</p>
</abstract>
<kwd-group>
<kwd>RNN-recurrent neural network</kwd>
<kwd>self-supervised</kwd>
<kwd>pre-train</kwd>
<kwd>seismic signal</kwd>
<kwd>sandstone content</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>With the development of deep learning, the model&#x2019;s capability and capacity have exploded. Aided by the rapid gains in hardware, models today can easily overfit millions of data and begin to demand more&#x2014;often publicly inaccessible&#x2014;labeled data. However, labeling data could be extremely costly. So, researchers have recently focused on how to use unlabeled data. In Natural Language Processing (NLP), researchers successfully archive the use of unlabeled data through self-supervised pre-training. The solutions based on autoregressive language modeling in GPT (Generative Pre-Training) (<xref ref-type="bibr" rid="B17">Radford et al., 2018a</xref>; <xref ref-type="bibr" rid="B18">Radford et al., 2018b</xref>; <xref ref-type="bibr" rid="B5">Brown et al., 2020</xref>) and masked autoencoding in BERT (Bidirectional Encoder Representations from Transformer) (<xref ref-type="bibr" rid="B6">Devlin et al., 2019</xref>) are conceptually simple: they remove a portion of the data and learn to predict the removed parts (<xref ref-type="bibr" rid="B8">He et al., 2021</xref>). In computer vision, researchers achieve self-supervised training through contrast learning (<xref ref-type="bibr" rid="B2">Bachman et al., 2019</xref>; <xref ref-type="bibr" rid="B10">Hjelm et al., 2019</xref>; <xref ref-type="bibr" rid="B9">He et al., 2020</xref>). After being Inspired by NLP, researchers utilized masked autoencoder with the idea of target reconstruction (<xref ref-type="bibr" rid="B8">He et al., 2021</xref>).</p>
<p>There are already some works based on self-supervised learning in seismic processing and interpretation. Such as seismic denoising, some researchers achieve random noise suppression based on Noise2Void (<xref ref-type="bibr" rid="B13">Krull et al., 2019</xref>), a kind of self-supervised method (<xref ref-type="bibr" rid="B4">Birnie et al., 2021</xref>; <xref ref-type="bibr" rid="B3">Birnie and Alkhalifah, 2022</xref>), some researchers achieve denoising through self-supervised seismic reconstruction (<xref ref-type="bibr" rid="B15">Meng et al., 2022</xref>), and some researchers achieve it by adding additive signal-dependent noise to the original seismic data and learn to predict the original data (<xref ref-type="bibr" rid="B20">Wu et al., 2022</xref>). What&#x2019;s more, self-supervised learning can also be used to reconstruct seismic data with consecutively missing traces (<xref ref-type="bibr" rid="B11">Huang et al., 2022</xref>), reconstruct the low-frequency components of seismic data (<xref ref-type="bibr" rid="B19">Wang et al., 2020</xref>), predict facies and other properties (<xref ref-type="bibr" rid="B24">Zhangdong and Alkhalifah, 2020</xref>), and perform other seismic processing tasks, like velocity estimation, first arrival picking, and NMO (normal moveout) (<xref ref-type="bibr" rid="B7">Harsuko and Alkhalifah, 2022</xref>). These works use this method for the same reason: to solve the problem of lacking labeled data, and good results have been achieved.</p>
<p>However, in seismic lithology prediction, self-supervised pre-training is not widely used. Seismic lithology prediction is mainly based on supervised learning at present. Current researches are basically about the improvement of model architecture, model input, and model output. For example, using time-frequency maps of seismic data to train a convolutional neural network to predict sandstone (<xref ref-type="bibr" rid="B26">Zhang et al., 2018</xref>; <xref ref-type="bibr" rid="B27">Zhang et al., 2020</xref>), using dozens of seismic attributes to train a convolutional recurrent network to predict lithology (<xref ref-type="bibr" rid="B14">Li et al., 2021</xref>), and using seismic data, 90&#xb0; phase-shift data, reservoir discontinuous boundary attribute to train a Multilayer Perceptron (MLP) to indirectly predict lithology by predicting cross-well natural gamma (<xref ref-type="bibr" rid="B23">Xin et al., 2020</xref>).</p>
<p>Our research is based on masked autoencoding, which is a form of more general denoising autoencoding (<xref ref-type="bibr" rid="B8">He et al., 2021</xref>). According to the idea of target reconstruction, the core of masked autoencoding, we design a masked autoencoder for seismic data with reference to BERT and MAE. Our model randomly masks a part of the seismic signal and reconstructs the masked part. It is based on bidirectional Long Short-Term Memory (Bi-LSTM) because the seismic signal is sequential and Bi-LSTM is good at extracting information from the sequential signal. In addition, it has an asymmetric encoder-decoder design, our decoder is lightweight compared to our encoder. We also try different masking ratios to find the best setting.</p>
<p>We experimented with synthetic seismic signals. After pre-training, we fine-tuned our model to the downstream task&#x2014;sandstone context predicting. The results show that the pre-trained model performs better. In the downstream task, the accuracy of the pre-trained model is at best 5% higher than that of the random initialization model.</p>
<p>There is a lot of unlabeled seismic data in oil and gas exploration, we hope that the application of unlabeled seismic data can make deep learning better applied in seismic interpretation.</p>
</sec>
<sec id="s2">
<title>Approach</title>
<sec id="s2-1">
<title>Datasets</title>
<p>Training and validating a Recurrent Neural Network (RNN) model often require a lot of signals and corresponding labels. Manually labeling or interpreting seismic data could be extremely time-consuming and highly subjective. In addition, inaccurate manual interpretation, including mislabeled and unlabeled faults, may mislead the learning process (<xref ref-type="bibr" rid="B16">Pham et al., 2019</xref>; <xref ref-type="bibr" rid="B21">Wu et al., 2019</xref>; <xref ref-type="bibr" rid="B22">Wu et al., 2020</xref>; <xref ref-type="bibr" rid="B1">Alkhalifah et al., 2021</xref>; <xref ref-type="bibr" rid="B25">Zhang et al., 2022</xref>). To avoid these problems, we create synthetic seismic signals and corresponding labels based on the convolution model for training and validating our RNN model.</p>
</sec>
<sec id="s2-2">
<title>Synthetic seismic signals</title>
<p>According to the convolution model, we know that seismic data are formed of a seismic wavelet and reflectivity:<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2a;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf1">
<mml:math id="m2">
<mml:mrow>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the wavelet, <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the reflectivity and <inline-formula id="inf3">
<mml:math id="m4">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the synthetic seismic signal, and <inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> means time.</p>
<p>A ricker wavelet is a good representation of seismic wavelets, so we chose it to generate our signals:<disp-formula id="e2">
<mml:math id="m6">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="|">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>&#x3c0;</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>&#x3c0;</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf5">
<mml:math id="m7">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the ricker wavelet and <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the frequency of the ricker wavelet.</p>
<p>The reflectivity is decided by the lithology of the stratum. We can first calculate the impedance sequence <inline-formula id="inf7">
<mml:math id="m9">
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> by density <inline-formula id="inf8">
<mml:math id="m10">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and velocity <inline-formula id="inf9">
<mml:math id="m11">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the stratum. Thus, we can calculate reflectivity from <inline-formula id="inf10">
<mml:math id="m12">
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>:<disp-formula id="e3">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
<disp-formula id="e4">
<mml:math id="m14">
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x3c1;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf11">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the impedance, <inline-formula id="inf12">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the reflectivity, <inline-formula id="inf13">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c1;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf14">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the density and velocity of the <inline-formula id="inf15">
<mml:math id="m19">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-th layer of the stratum.</p>
<p>So, our workflow of creating synthetic seismic signals, see <xref ref-type="fig" rid="F1">Figure 1</xref>, is as follows: we first establish a stratum of randomly distributed sandstone and mudstone, and then we calculate the impedance sequence <inline-formula id="inf16">
<mml:math id="m20">
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> using <inline-formula id="inf17">
<mml:math id="m21">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf18">
<mml:math id="m22">
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the sandstone and mudstone. At last, we obtain the synthetic seismic signal using Eq. <xref ref-type="disp-formula" rid="e1">1</xref>, and the label is the corresponding stratum.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The workflow of creating synthetic seismic signals according to the convolution model for training and validation.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g001.tif"/>
</fig>
<p>The sampling interval of the Ricker wavelet and impedance sequence is 1&#xa0;ms, so the sampling interval of the synthetic seismic signal is 1&#xa0;ms too.</p>
</sec>
<sec id="s2-3">
<title>Self-supervised pre-training of seismic signals</title>
<p>Our self-supervised pre-training is based on a masked autoencoder, which performs well in NLP and computer vision. A masked autoencoder masks random parts from the input signal and reconstructs the missing part. It has an encoder that maps the input signal to a latent representation and a decoder that reconstructs the original signal from the latent representation. The workflow is shown in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Overview of the network architecture and self-supervised pre-training.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g002.tif"/>
</fig>
</sec>
<sec id="s2-4">
<title>Target reconstruction</title>
<p>As mentioned, a masked autoencoder masks random parts from the input signal and reconstructs them. It is important to choose how to mask random parts. In this paper, we design a method for the seismic signal with reference to Masked Language Model (MLM) in BERT (<xref ref-type="bibr" rid="B6">Devlin et al., 2019</xref>). It is as follows.</p>
<p>We randomly choose a part of the signal, and we replace the chosen part with (1) zeros 80% of the time, (2) a random part of this signal 10% of the time, and (3) keep unchanged 10% of the time.</p>
<p>In target reconstruction task design we follow the MLM in BERT (<xref ref-type="bibr" rid="B6">Devlin et al., 2019</xref>) closely. According to the ablation over different masking strategies in BERT, the best strategy of masking rates of zero masking, random masking, and no masking are 80%, 10%, and 10%, respectively. The purpose of no masking is to reduce the mismatch between pre-training and fine-tuning as the signal will not be masked during the fine-tuning stage. The mixed strategy of random masking and zero masking will help the model learn better.</p>
</sec>
<sec id="s2-5">
<title>Masking ratio</title>
<p>Another point of the masked autoencoder is the masking ratio. It is different between different signals, when the signals are highly semantic and information-dense, the masking ratio need could be small (e.g., 10%), but on the contrary, when the signals are low in semantic, it needs to be high (e.g., 75%). For low semantic signals, a missing part can be recovered from neighboring parts without high-level understanding of signals. A high masking ratio can create a challenging self-supervisory task that requires holistic understanding beyond low-level signal statistics. We tried different masking ratios to find out the best option: about 20%.</p>
</sec>
<sec id="s2-6">
<title>Masked autoencoder</title>
<p>Masked autoencoders have an encoder and a decoder.</p>
<p>Encoder: An encoder maps the input signal to a latent representation. We know that a seismic signal is a sequence signal, and the RNN model is a kind of sequential model, that is good at extracting information from a sequence signal. So, our encoder is based on Bi-LSTM, a kind of bidirectional RNN. We use a bidirectional RNN because the wavelet has a width, which makes one point of the seismic signal is affected by the above and below points.</p>
<p>Decoder: A decoder reconstructs the original signal from the latent representation. It plays a different role in reconstructing different signals. When the decoder&#x2019;s output is of a lower semantic level such as pixels, the decoder needs to be designed carefully. This is in contrast to language, where the decoder predicts missing words that contain rich semantic information. So, our decoder is based on Bi-LSTM and MLP. In addition, the decoder is only used during pre-training to perform the signal reconstruction task. Therefore, it is more lightweight than the encoder to make sure effectiveness of the encoder and reduce computation.</p>
</sec>
</sec>
<sec id="s3">
<title>Experiments</title>
<p>Our experiment is based on synthetic seismic data. Specifically, we pre-train our model on synthetic seismic data and we fine-tune our model on synthetic seismic data and corresponding labels. The specific method of creating synthetic seismic data is described above. In our experiments, the size of the dataset is shown in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Size of training and validating datasets.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center"/>
<th align="center">Pre-train</th>
<th align="center">Fine-tune</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Train</td>
<td align="center">100,000</td>
<td align="center">20,000</td>
</tr>
<tr>
<td align="center">Valid</td>
<td align="center">25,000</td>
<td align="center">5,000</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s3-1">
<title>Self-supervised pre-training</title>
<p>The overview of pre-training is shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. The N means the number of layers of the encoder&#x2019;s Bi-LSTM, and the M means the number of layers of the decoder&#x2019;s Bi-LSTM. The details of the pre-training model are shown in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Details of the pre-training model.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Encoder Bi-LSTM layers N</th>
<th align="center">Decoder Bi-LSTM layers M</th>
<th align="center">Hidden size D</th>
<th align="center">MLP size</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">
<bold>4</bold>
</td>
<td align="center">1</td>
<td align="center">256</td>
<td align="center">512</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We pre-train our model using an Adam optimizer (<xref ref-type="bibr" rid="B12">Kingma and Ba, 2017</xref>) with <inline-formula id="inf19">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>9</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf20">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>999</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and a batch size of 1,600. We use a step learning rate decay, see <xref ref-type="fig" rid="F3">Figure 3</xref>. We pre-trained for 200 epochs at 4 masking ratios: .1, .2, .3, and .4. To get a qualitative sense of our reconstruction task, see <xref ref-type="fig" rid="F4">Figure 4</xref>. Our loss function computes the Mean Squared Error (MSE) between the reconstructed and original seismic signals. We compute the loss only on masked parts, similar to BERT (<xref ref-type="bibr" rid="B6">Devlin et al., 2019</xref>) and MEA (<xref ref-type="bibr" rid="B8">He et al., 2021</xref>), so the model output on unmasked parts is qualitatively worse. So, we overlay the output with the unmasked parts to improve visual quality.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Schedule of the learning rate in pre-training.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g003.tif"/>
</fig>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Reconstructions of the test synthetic seismic signals.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g004.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>Fine-tuning</title>
<p>We can fine-tune our pre-trained model to downstream tasks, and, a new decoder for the downstream task is necessary. In this paper, the downstream task is a classification task.</p>
<p>Downstream task: We choose a simple task as our downstream task, so we can pay more attention to the effect of self-supervised pre-training. In the downstream task, our model predicts the sandstone content from the input seismic signal. This is a simple task but also meaningful, we can get an overall understanding of a stratum from the prediction results which can be used as a reference for further work.</p>
<p>The output of our downstream task is the percentage of the Sandstone content, which is suitable for regression. However, we still choose classification because a regression task is harder than a classification task. As mentioned above, we need a simple task to validate our method. In addition, we can calculate the accuracy which can intuitively show the performance of the model.</p>
<p>Our model predicts the sandstone content from the input seismic signal and its decoder is a simple MLP, see <xref ref-type="fig" rid="F5">Figure 5</xref>. The details of the fine-tuning model are shown in <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Overview of the fine-tuning network.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g005.tif"/>
</fig>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Details of the fine-tuning model.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Encoder Bi-LSTM layers N</th>
<th align="center">Hidden size D</th>
<th align="center">MLP size</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">4</td>
<td align="center">256</td>
<td align="center">1,024</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We fine-tune our model using again an Adam optimizer (<xref ref-type="bibr" rid="B12">Kingma and Ba, 2017</xref>) with <inline-formula id="inf21">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>9</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf22">
<mml:math id="m26">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>999</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, a batch size of 2,048 and apply a weight decay of .001. We use a step learning rate decay, see <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Schedule of learning rate of pre-training.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g006.tif"/>
</fig>
<p>In fine-tuning we try several pre-training models with different masking ratios to find out the best choice of masking ratio. We train our model 300 epochs and record the losses during training, the losses of the last 50 epochs are shown in <xref ref-type="fig" rid="F7">Figures 7</xref>, <xref ref-type="fig" rid="F8">8</xref>. We find that there is an abrupt behavior for the training and validation losses between epochs 270 and 280. We think this is caused by the an excessively high learning rate for that stage of the training. So, when the learning rate decreases, the problem disappears.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Training loss on the training dataset.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g007.tif"/>
</fig>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Training loss on validation dataset.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g008.tif"/>
</fig>
<p>After training, we evaluate the accuracy of the models on the test synthetic dataset, see <xref ref-type="fig" rid="F9">Figure 9</xref>.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Comparing the accuracy of different masking ratios and non-pre-training.</p>
</caption>
<graphic xlink:href="feart-10-1081998-g009.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>In this paper, we do successfully implement a self-supervised pre-training of seismic signals through a reconstruction task. We have tried different masking ratios and found that, in our experiment, the best choice of masking ratio is about 20%. Seismic signals are not highly semantic, a higher masking ratio should be suitable theoretically, but the truth is the exact opposite. We think it may be caused by masked input signals. We input our masked signals to the model, including the masked and unmasked parts. So, the RNN extracts information from potential noise (unmasked) and masked parts. With the rising masking ratio, the model is seriously disturbed by noise. We think a model which can input the unmasked part may solve this problem such as Transformer.</p>
<p>In the downstream tasks, the pre-trained model is better than the random initialization model. This proves the effectiveness of our self-supervised pre-training method for seismic signals. So, more tasks can be optimized by self-supervised pre-training in seismic deep learning.</p>
</sec>
<sec sec-type="conclusion" id="s5">
<title>Conclusion</title>
<p>With the development of deep learning, unlabeled data will be used more. On the one hand, the appetite for large-scale models is hard to address with the labeled data, because labeled data are very limited. On the other hand, extracting information from unlabeled data is an important target in artificial intelligence.</p>
<p>For now, self-supervised pre-training performs well in taking advantage of unlabeled data in NLP and computer vision. In our work, we designed a masked autoencoder for seismic self-supervised pre-training. When we fine-tune our model for a downstream task, the pre-trained model always performs well, which proves the effectiveness of our method. We also explore how the masking ratio influences accuracy in the downstream task.</p>
<p>Our model is based on RNN, which is quite costly. We think a low cost model is necessary so that we can train on bigger datasets. We think a bigger model and dataset may achieve higher accuracy. We hope that by using the self-supervised pre-training, we can better apply deep learning to seismic signals.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="s7">
<title>Author contributions</title>
<p>CW done experiment and writing. XJW guides this work. YZ, XMW, QL and YC help CW with work.</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Alkhalifah</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ovcharenko</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>2021</year>). <publisher-name>European Association of Geoscientists &#x26; Engineers</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>5</lpage>.<article-title>MLReal: Bridging the gap between training on synthetic data and real data applications in machine learning</article-title>, <conf-name>Proceedings of the 82nd EAGE Annual Conference &#x26; Exhibition</conf-name>, <conf-loc>Amsterdam, Netherlands</conf-loc>, <conf-date>October 2021</conf-date>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bachman</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Hjelm</surname>
<given-names>R. D.</given-names>
</name>
<name>
<surname>Buchwalter</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Learning representations by maximizing mutual information across views</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>32</volume>. <pub-id pub-id-type="doi">10.48550/arXiv.1906.00910</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Birnie</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Alkhalifah</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Transfer learning for self-supervised, blind-spot seismic denoising</article-title>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2209.12210">https://arxiv.org/abs/2209.12210</ext-link>.</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Birnie</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ravasi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Alkhalifah</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>The potential of self-supervised networks for random noise suppression in seismic data</article-title>. <source>Artif. Intell. Geosciences</source> <volume>2</volume>, <fpage>47</fpage>&#x2013;<lpage>59</lpage>. <pub-id pub-id-type="doi">10.1016/j.aiig.2021.11.001</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Brown</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Mann</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Ryder</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Subbiah</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kaplan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dhariwal</surname>
<given-names>P.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Language models are few-shot learners</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>33</volume>, <fpage>1877</fpage>&#x2013;<lpage>1901</lpage>. <pub-id pub-id-type="doi">10.48550/arXiv.2005.14165</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Devlin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>M. W.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Toutanova</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Bert: Pre-Training of deep bidirectional transformers for language understanding</article-title>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1810.04805">https://arxiv.org/abs/1810.04805</ext-link>.</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Harsuko</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Alkhalifah</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>StorSeismic: A new paradigm in deep learning for seismic processing</article-title>. <source>IEEE Trans. Geoscience Remote Sens.</source> <volume>60</volume>, <fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2022.3216660</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Dollar</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>Masked autoencoders are scalable vision learners</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>, <conf-loc>Nashville, TN, USA</conf-loc>, <conf-date>September 2021</conf-date>.</citation>
</ref>
<ref id="B9">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Momentum contrast for unsupervised visual representation learning</article-title>, in <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>, <conf-loc>Seattle, WA, USA</conf-loc>, <conf-date>June 2020</conf-date>, <fpage>9729</fpage>&#x2013;<lpage>9738</lpage>.</citation>
</ref>
<ref id="B10">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Hjelm</surname>
<given-names>R. D.</given-names>
</name>
<name>
<surname>Fedorov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Lavoie-Marchildon</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Grewal</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Bachman</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Trischler</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Learning deep representations by mutual information estimation and maximization</article-title>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1808.06670">https://arxiv.org/abs/1808.06670</ext-link>.</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Geng</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Self-supervised deep learning to reconstruct seismic data with consecutively missing traces</article-title>. <source>IEEE Trans. Geoscience Remote Sens.</source> <volume>60</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2022.3148994</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Adam: A method for stochastic optimization</article-title>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1412.6980">https://arxiv.org/abs/1412.6980</ext-link>.</citation>
</ref>
<ref id="B13">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Krull</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Buchholz</surname>
<given-names>T. O.</given-names>
</name>
<name>
<surname>Jug</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Noise2Void - learning denoising from single noisy images</article-title>,&#x201d; in <conf-name>Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Long Beach, CA, USA</conf-loc>, <conf-date>June 2019</conf-date>, <fpage>2124</fpage>&#x2013;<lpage>2132</lpage>.</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Xi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Research on reservoir lithology prediction method based on convolutional recurrent neural network</article-title>. <source>Comput. Electr. Eng.</source> <volume>95</volume>, <fpage>107404</fpage>. <pub-id pub-id-type="doi">10.1016/j.compeleceng.2021.107404</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meng</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Self-supervised learning for seismic data reconstruction and denoising</article-title>. <source>IEEE Geoscience Remote Sens. Lett.</source> <volume>19</volume>, <fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2021.3068132</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pham</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Fomel</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Dunlap</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Automatic channel detection using deep learning</article-title>. <source>Interpretation</source> <volume>7</volume> (<issue>3</issue>), <fpage>SE43</fpage>&#x2013;<lpage>SE50</lpage>. <pub-id pub-id-type="doi">10.1190/int-2018-0202.1</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Radford</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Narasimhan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Salimans</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Improving Language understanding by generative pre-training</source>. <publisher-loc>San Francisco, California</publisher-loc>: <publisher-name>OpenAI blog</publisher-name>.</citation>
</ref>
<ref id="B18">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Radford</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Child</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>David</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Dario</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Language models are unsupervised multitask learners</source>. <publisher-loc>San Francisco, California</publisher-loc>: <publisher-name>OpenAI blog</publisher-name>.</citation>
</ref>
<ref id="B19">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Self-supervised learning for low frequency extension of seismic data</article-title>,&#x201d; in <conf-name>Proceedings of the SEG Technical Program Expanded Abstracts 2020</conf-name>, <conf-loc>Virtual</conf-loc>, <conf-date>May 2020</conf-date>, <fpage>1501</fpage>&#x2013;<lpage>1505</lpage>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Self-adaptive denoising net: Self-supervised learning for seismic migration artifacts and random noise attenuation</article-title>. <source>J. Petroleum Sci. Eng.</source> <volume>214</volume>, <fpage>110431</fpage>. <pub-id pub-id-type="doi">10.1016/j.petrol.2022.110431</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Shi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fomel</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>FaultSeg3D: Using synthetic data sets to train an end-to-end convolutional neural network for 3D seismic fault segmentation</article-title>. <source>GEOPHYSICS</source> <volume>84</volume> (<issue>3</issue>), <fpage>IM35</fpage>&#x2013;<lpage>IM45</lpage>. <pub-id pub-id-type="doi">10.1190/geo2018-0646.1</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Deep learning for characterizing paleokarst collapse features in 3-D seismic images</article-title>. <source>J. Geophys. Res. Solid Earth</source> <volume>125</volume> (<issue>9</issue>), <fpage>e2020JB019685</fpage>. <pub-id pub-id-type="doi">10.1029/2020jb019685</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xin</surname>
<given-names>D. U.</given-names>
</name>
<name>
<surname>Tingen</surname>
<given-names>F. A. N.</given-names>
</name>
<name>
<surname>Jianhua</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Characterization of thin sand reservoirs based on a multi-layer perceptron deep neural network</article-title>. <source>Oil Geophys. Prospect.</source> <volume>55</volume> (<issue>6</issue>), <fpage>1178</fpage>&#x2013;<lpage>1187</lpage>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhangdong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Alkhalifah</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>High-resolution reservoir characterization using deep learning-aided elastic full-waveform inversion: The North Sea field data example</article-title>. <source>GEOPHYSICS</source> <volume>85</volume> (<issue>4</issue>), <fpage>WA137</fpage>&#x2013;<lpage>WA146</lpage>. <pub-id pub-id-type="doi">10.1190/geo2019-0340.1</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Seismic characterization of deeply buried paleocaves based on Bayesian deep learning</article-title>. <source>J. Nat. Gas Sci. Eng.</source> <volume>97</volume>, <fpage>104340</fpage>. <pub-id pub-id-type="doi">10.1016/j.jngse.2021.104340</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deep learning for seismic lithology prediction</article-title>. <source>Geophys. J. Int.</source> <volume>215</volume>. <pub-id pub-id-type="doi">10.1093/gji/ggy344</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>G. Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z. Z.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>C. Y.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Seismic reservoir prediction method based on wavelet transform and convolutional neural network and its application</article-title>. <source>J. China Univ. Petroleum Ed. Nat. Sci. Chin.</source> <volume>44</volume> (<issue>04</issue>), <fpage>83</fpage>&#x2013;<lpage>93</lpage>.</citation>
</ref>
</ref-list>
</back>
</article>