<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="research-article" dtd-version="1.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Energy Res.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Energy Research</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Energy Res.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-598X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1533878</article-id>
<article-id pub-id-type="doi">10.3389/fenrg.2025.1533878</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A self-attention enhanced GRU network for predicting life cycle cost of substation GIS equipment</article-title>
<alt-title alt-title-type="left-running-head">Xue et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fenrg.2025.1533878">10.3389/fenrg.2025.1533878</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Xue</surname>
<given-names>Wenjie</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2384508"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>An</surname>
<given-names>Chaoyin</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sheng</surname>
<given-names>Tengfei</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Dong</surname>
<given-names>Pingxian</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal Analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhai</surname>
<given-names>Yuxin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Jinfeng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal Analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<institution>State Grid Henan Electric Power Company Economic Research Institute</institution>, <city>Zhengzhou</city>, <state>Henan</state>, <country country="CN">China</country>
</aff>
<aff id="aff2">
<label>2</label>
<institution>State Grid Henan Electric Power Company</institution>, <city>Zhengzhou</city>, <state>Henan</state>, <country country="CN">China</country>
</aff>
<author-notes>
<corresp id="c001">
<label>&#x2a;</label>Correspondence: Wenjie Xue, <email xlink:href="mailto:15138656738@163.com">15138656738@163.com</email>
</corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-16">
<day>16</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>13</volume>
<elocation-id>1533878</elocation-id>
<history>
<date date-type="received">
<day>25</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="rev-recd">
<day>03</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>17</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Xue, An, Sheng, Dong, Zhai and Zhang.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Xue, An, Sheng, Dong, Zhai and Zhang</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-16">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The challenges of limited sample size and anomalies in the life cycle cost (LCC) data of substation GIS equipment make it difficult to achieve accurate LCC estimation. To address these issues, this paper introduces similar substation GIS equipment LCC data and incorporates a self-attention mechanism to expand the sample size. Additionally, a GRU algorithm is applied to predict time series data and mitigate the impact of anomalies on the prediction. A novel SA-GRU-based model for substation GIS equipment LCC data prediction is proposed. Considering the discrepancies in data characteristics between similar substation GIS equipment LCC data and existing sample data, the self-attention mechanism is utilized to enhance data features. The influence weights between similar and sample data are quantified based on these enhanced features to ensure consistency in data characteristics. Furthermore, the GRU algorithm is employed to alleviate gradient vanishing and exploding issues during training, ultimately enabling LCC data prediction over the entire life cycle. Validation using LCC data from a 110 kV GIS equipment substation in Henan Province demonstrates that the proposed method significantly improves prediction accuracy compared to other prediction approaches.</p>
</abstract>
<kwd-group>
<kwd>substation GIS equipment</kwd>
<kwd>LCC data</kwd>
<kwd>forecasting model</kwd>
<kwd>self-attention mechanism</kwd>
<kwd>GRU model</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the State Grid Technology Project (B717L028K136).</funding-statement>
</funding-group>
<counts>
<fig-count count="7"/>
<table-count count="5"/>
<equation-count count="13"/>
<ref-count count="26"/>
<page-count count="00"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Smart Grids</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>The results of Life Cycle Cost (LCC) estimation directly determine the selection of GIS equipment in substations (<xref ref-type="bibr" rid="B1">Cai et al., 2011</xref>; <xref ref-type="bibr" rid="B12">Su et al., 2012</xref>). Traditional LCC estimation methods, which rely on manual calculations, have gradually been replaced by predictive methods based on LCC data. However, the accuracy of LCC data prediction depends heavily on the continuous collection of substantial economic data generated by GIS equipment throughout its lifecycle. Currently, the collection of LCC data for GIS equipment in substations in China started relatively late, and the data typically suffers from issues such as anomalies and insufficient sample size. These challenges hinder accurate LCC estimation (<xref ref-type="bibr" rid="B11">Shi et al., 2020</xref>; <xref ref-type="bibr" rid="B6">Li et al., 2023</xref>). Therefore, developing a prediction method suitable for LCC data of substation GIS equipment is critical for addressing the economic evaluation and selection of GIS equipment under existing conditions (<xref ref-type="bibr" rid="B3">Du, 2013</xref>).</p>
<p>In recent years, extensive research has focused on substation LCC prediction. <xref ref-type="bibr" rid="B17">Xiong Z. et al. (2021)</xref> developed a substation LCC prediction model by optimizing a least-squares support vector machine (LS-SVM) with quantum-behaved particle swarm optimization (QPSO), effectively improving accuracy under small-sample conditions. <xref ref-type="bibr" rid="B10">Qiao et al. (2015)</xref> used a genetic algorithm (GA) to tune LS-SVM hyperparameters, further enhancing generalization and predictive stability. Addressing the unique environment of high-latitude frigid regions, <xref ref-type="bibr" rid="B7">Liu L. et al. (2020)</xref> constructed an LCC prediction model in which LS-SVM is optimized by the salp swarm algorithm (SSA), achieving a marked gain in region-specific accuracy. To handle data uncertainty, <xref ref-type="bibr" rid="B8">Liu S. et al. (2020)</xref> incorporated fuzzy theory into the LCC estimation process, improving the reliability of the estimates. <xref ref-type="bibr" rid="B15">Wang et al. (2021)</xref> established an LCC mathematical model for a specific substation type and conducted sensitivity analysis to identify key drivers.</p>
<p>Despite these advances in traditional model optimization and uncertainty handling, challenges remain for GIS equipment LCC in substations, a task characterized by small samples and nonstationary time series. Early studies (<xref ref-type="bibr" rid="B14">Wang et al., 2020</xref>) based on optimized least squares are susceptible to missing values and outliers, leading to substantial errors (<xref ref-type="bibr" rid="B4">Kong et al., 2020</xref>). Although fuzzy neural networks (<xref ref-type="bibr" rid="B20">Yang et al., 2017</xref>; <xref ref-type="bibr" rid="B26">Zhu et al., 2024</xref>) can mitigate the impact of anomalies, directly applying neural networks to GIS LCC time series often results in parameter instability during training, making vanishing or exploding gradients more likely (<xref ref-type="bibr" rid="B18">Xiong Y. et al., 2021</xref>).</p>
<p>In the related domain of wind power forecasting&#x2014;where data exhibit similar characteristics&#x2014;research is comparatively mature (<xref ref-type="bibr" rid="B9">Peng et al., 2016</xref>; <xref ref-type="bibr" rid="B13">Tao et al., 2018</xref>). For instance (<xref ref-type="bibr" rid="B19">Xue et al., 2019</xref>), introduced the gated recurrent unit (GRU), whose gating mechanism strengthens temporal dependencies. However, GRU models require sizable training datasets, and their accuracy is difficult to guarantee in small-sample settings (<xref ref-type="bibr" rid="B25">Zhang et al., 2022</xref>). To address this (<xref ref-type="bibr" rid="B16">Wang et al., 2022</xref>), integrated an attention mechanism with GRU and enlarged the training set using data from similar wind farms. Nevertheless, the method relies on manual feature extraction; when data distributions are complex, feature design becomes challenging and subjective, and accuracy remains insufficient (<xref ref-type="bibr" rid="B21">Yang et al., 2023</xref>).</p>
<p>Contributions. This paper makes the following contributions to LCC forecasting for substation GIS equipment:<list list-type="order">
<list-item>
<p>Model design for data scarcity and anomalies. We propose SA-GRU, a self-attention&#x2013;enhanced GRU tailored to LCC time series with scarce samples and outliers.</p>
</list-item>
<list-item>
<p>Attention-weighted similar-station integration. The self-attention module automatically learns feature representations and assigns data-driven weights to existing and similar-station samples, yielding effective sample augmentation and improved robustness.</p>
</list-item>
<list-item>
<p>Real-world validation and deployment. We evaluate the approach on real LCC data and report a successful deployment in a technical retrofit and major-overhaul project at a substation in Henan Province, demonstrating practical utility.</p>
</list-item>
</list>
</p>
<p>For clarity, our forecasting target is the annualized life-cycle cost (LCC), which serves as the decision-level quantity for substation equipment selection and budgeting. As LCC aggregates capital expenditure, routine/corrective maintenance, overhaul, outage-related loss, and end-of-life components, modeling it directly matches operational practice. The remainder of this paper is organized as follows. <xref ref-type="sec" rid="s2">Section 2</xref> describes the dataset and formalizes the problem; <xref ref-type="sec" rid="s3">Section 3</xref> details the SA-GRU methodology; <xref ref-type="sec" rid="s4">Section 4</xref> reports the experimental setup and comparative results; <xref ref-type="sec" rid="s5">Section 5</xref> concludes and discusses future work.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>LCC data of GIS equipment in substations and data prediction methods</title>
<p>This chapter establishes the foundational groundwork for GIS equipment LCC prediction by addressing two key prerequisites: first, analyzing the inherent characteristics of LCC data to support the design of effective methods; second, examining the limitations of current data prediction techniques and suggesting targeted improvements.</p>
<sec id="s2-1">
<label>2.1</label>
<title>LCC data of GIS equipment in substations</title>
<p>The economic performance of GIS equipment is typically assessed based on its Life Cycle Cost (LCC). As shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, the LCC of GIS equipment consists of several types of costs: initial investment cost, operation cost, maintenance cost, failure cost, and decommissioning cost. Among these, the initial investment cost and decommissioning cost can be largely determined when the substation is completed, making them fixed costs. However, the operation, maintenance, and failure costs require annual statistics and are considered variable costs.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>LCC data and issues of GIS equipment in substations.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g001.tif">
<alt-text content-type="machine-generated">Flowchart illustrating the life cycle cost of GIS equipment in a substation. It includes five main cost categories: initial investment, decommissioning, running, overhaul, and failure. Each category has specific expenses, such as equipment procurement and maintenance fees. The chart highlights issues like short commissioning time, lack of cost materials, and manual collection errors. It also notes data characteristics like few samples and exceptions, concluding with the difficulty in predicting LCC data accurately.</alt-text>
</graphic>
</fig>
<p>From the perspective of actual data collection on variable annual costs, GIS equipment in substations only began to be widely used in China in the 21st century, and related data collection activities started relatively late. As a result, the quantity of LCC data collected across different years is limited, leading to a sparse overall LCC data sample. Additionally, in the early stages of LCC data collection and transmission, manual recording methods were commonly relied upon. Consequently, issues such as insufficient cost materials, data loss, and recording errors have occurred, resulting in anomalies within the collected LCC sample data.</p>
<p>In summary, these data issues have created significant challenges for accurately calculating LCC data. Therefore, it is necessary to identify an appropriate data prediction method based on the characteristics of the LCC data for GIS equipment in substations. This method should aim to expand the data sample size while minimizing the impact of anomalous data on the predictions.</p>
</sec>
<sec id="s2-2">
<label>2.2</label>
<title>Limitations of existing data prediction methods and proposed improvements</title>
<p>As discussed in Section 1.1, to achieve accurate predictions of LCC data for GIS equipment in substations, it is essential first to expand the sample size as much as possible to address the issue of limited data samples. Then, the characteristics of the data samples should be analyzed to resolve the issue of data anomalies.</p>
<p>Currently, under the condition that sample data exhibit certain regular variations, an effective approach to address the issue of limited sample data for prediction is to introduce similar data of the same scale as the existing sample data, thereby expanding the original dataset (<xref ref-type="bibr" rid="B25">Zhang et al., 2022</xref>; <xref ref-type="bibr" rid="B16">Wang et al., 2022</xref>; <xref ref-type="bibr" rid="B21">Yang et al., 2023</xref>; <xref ref-type="bibr" rid="B2">Cheng et al , 2022</xref>). However, considering that the data characteristics of the existing samples and the similar data may not be entirely consistent, an attention mechanism with focused characteristics is employed to quantify the influence weights between the existing sample data and similar data. The weighted data is then used as the expanded sample dataset (<xref ref-type="bibr" rid="B13">Tao et al., 2018</xref>). However, the attention mechanism requires manually assigned data feature vectors to process and analyze the data around these vectors. Given that LCC data for GIS equipment in substations belongs to cost data, which is significantly influenced by the actual operational conditions of the equipment, the data distribution pattern is complex, and the data features are not prominent. Therefore, manual feature extraction lacks a solid basis and accuracy. As a result, directly using the attention mechanism to obtain influence weights between data may lead to low accuracy. If a method could calculate vector weights between data to automatically extract feature vectors, enhancing the reliability of feature extraction results, it would avoid the subjectivity associated with manual feature vector extraction. This would provide an objective basis for quantifying the influence weights between existing sample data and similar data, thus improving the accuracy of data predictions.</p>
<p>Regarding the issue of anomalous data, after performing data preprocessing based on data distribution characteristics, it is important to consider that GIS equipment LCC data is collected and arranged in chronological order, making it sequential data. Therefore, a Recurrent Neural Network (RNN), which can capture temporal relationships, is employed for time-series data prediction (<xref ref-type="bibr" rid="B22">Ye et al., 2023</xref>; <xref ref-type="bibr" rid="B5">Li and Lin, 2018</xref>). In processing time-series data such as LCC data, RNN adjusts neural network parameters by calculating gradients according to the backpropagation algorithm and the chain rule (<xref ref-type="bibr" rid="B4">Kong et al., 2020</xref>) to minimize the loss function (<xref ref-type="bibr" rid="B4">Kong et al., 2020</xref>), thereby reducing the overall error in the prediction results. However, according to the chain rule, each parameter&#x2019;s gradient expression in RNN is a product of terms, which often causes the calculated values to converge exponentially toward zero or infinity as the period increases. This leads to the gradient vanishing or exploding problem in the LCC data, raising concerns about the accuracy of the predictions. The fundamental cause of gradient vanishing or exploding in RNN lies in the linear nature of its information transfer structure. Therefore, by optimizing the internal structure of the network to selectively retain long-span information, the non-linear relationship between data points can be enhanced, suppressing the occurrence of gradient vanishing or exploding during the network training and updating process.</p>
<p>In summary, to achieve accurate LCC data predictions for GIS equipment in substations, we can first introduce a self-attention mechanism to enhance the feature representation of existing sample data, quantifying the influence weights between existing sample data and similar data, thereby expanding the sample size. Then, a GRU model can be employed, leveraging its gated structure to suppress the gradient vanishing or exploding issues, further improving the accuracy of the prediction results and enabling precise LCC data predictions.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>SA-GRU model for LCC data prediction</title>
<p>This chapter details the development of the SA-GRU model for LCC prediction, targeting two core challenges in substation GIS equipment LCC forecasting: the insufficiency of sample sizes and the presence of gradient anomalies in time-series data. The former challenge is addressed by quantifying the influence weights of data from similar substations through the self-attention mechanism, thereby expanding the sample dataset. The latter challenge is alleviated via the gated structure of the GRU, which suppresses gradient vanishing and exploding phenomena during model training.</p>
<sec id="s3-1">
<label>3.1</label>
<title>Sample expansion based on self-attention mechanism</title>
<p>Given the limited number of existing sample data, it is feasible to introduce LCC-similar data from GIS equipment in substations of the same voltage level and similar environments to expand the sample size. However, since similar data cannot fully represent the characteristics of the existing sample data, directly adding similar data to the existing samples for cost prediction would fail to achieve precise LCC calculations. Therefore, it is necessary to apply weighting to similar data based on the characteristics of the existing samples. This approach addresses the issue of reduced prediction accuracy caused by the incomplete consistency between similar data and existing sample data characteristics.</p>
<p>Due to the complex distribution patterns of LCC data for GIS equipment in substations, identifying correlations between data for weighting purposes requires transforming both existing sample data and similar data into matrix form. Different years are used as the row labels of the data, and various types are used as the column labels of the matrix. Based on this, a matrix representation of the existing sample data is constructed as follows:<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:mtable columnalign="center">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>11</mml:mn>
</mml:msub>
</mml:mtd>
<mml:mtd>
<mml:mo>&#x22ef;</mml:mo>
</mml:mtd>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mo>&#x22ee;</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd>
<mml:mo>&#x22ee;</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:msub>
<mml:mi mathvariant="normal">t</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:msub>
</mml:mtd>
<mml:mtd>
<mml:mo>&#x22ef;</mml:mo>
</mml:mtd>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mtext>tn</mml:mtext>
</mml:msub>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub> represents the constructed matrix of existing sample data; <italic>h</italic>
<sub>
<italic>tn</italic>
</sub> denotes the LCC data of the n-th type in the t-th year for GIS equipment; <inline-formula id="inf1">
<mml:math id="m2">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>;</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>t</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Considering that the similar data includes LCC data from multiple GIS equipment in substations of the same scale, the matrix is divided into sub-blocks by different substations as column references. Each sub-matrix is constructed in a manner consistent with the matrix representation of the existing sample data. To facilitate correlation comparisons between different data sets, the existing sample data is included as a reference value within the similar data, so that the matrix form of the similar data can be represented as follows:<disp-formula id="e2">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#x22ef;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>L</italic>
</bold>
<sub>LCC</sub> In this expression; <bold>
<italic>H</italic>
</bold>
<italic>m</italic> LCCdenotes the similarity data sub-matrix composed of the LCC data from the m-th GIS equipment in substations of the same scale. The number of rows and columns in <bold>
<italic>H</italic>
</bold>
<italic>m</italic> LCC matches those of the existing sample data matrix.</p>
<p>Based on the matrix constructions in <xref ref-type="disp-formula" rid="e1">Equations 1</xref> and <xref ref-type="disp-formula" rid="e2">2</xref>, it can be seen that the task of identifying correlations between existing sample data and similar data has been transformed into a problem of determining correlations between matrix column vectors (<xref ref-type="bibr" rid="B24">Zhang, 2021</xref>). According to the attention mechanism&#x2019;s solving algorithm (<xref ref-type="bibr" rid="B23">Yin et al., 2021</xref>), the correlation between matrix column vectors can be obtained using an attention scoring function, specifically:<disp-formula id="e3">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>Soft</mml:mtext>
<mml:mi>max</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>In this expression,<bold>
<italic>&#x3b1;</italic>
</bold>
<sub>
<italic>m</italic>
</sub> represents the weight coefficient matrix between the <italic>m-th</italic> similar data sub-matrix and the existing sample data matrix, indicating the correlation between matrices; Softmax is a normalization function; and <italic>s</italic>(&#xb7;) is the attention scoring function, Considering the long time sequence of LCC data and the suitability of the dot-product model for calculating vector weight coefficients between matrices, the attention scoring function in <xref ref-type="disp-formula" rid="e3">Equation 3</xref> employs a scaled dot-product model.</p>
<p>In solving matrix correlations using the scoring function, the attention scoring function performs calculations directly based on the matrix elements. Given that the LCC data of GIS equipment has inconspicuous data features, the scoring results of the attention function may lack accuracy. Therefore, an algorithm based on the self-attention mechanism (<xref ref-type="bibr" rid="B23">Yin et al., 2021</xref>) is introduced to determine the weight relationships among elements within the <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub> matrix, thereby enhancing data features. When solving for weight relationships with the self-attention mechanism, to better capture the correlations between any two elements in the matrix, the <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub> matrix needs to be transformed into three matrices through linear mapping: the LCC data query vector matrix <bold>
<italic>Q</italic>
</bold>
<sub>LCC</sub>, the LCC data key vector matrix <bold>
<italic>K</italic>
</bold>
<sub>LCC</sub>, and the LCC data value vector matrix <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub>. Here, matrices <bold>
<italic>Q</italic>
</bold>
<sub>LCC</sub> and <bold>
<italic>K</italic>
</bold>
<sub>LCC</sub> share the same dimensions and are used to calculate the weight coefficients between column vectors of matrix <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub>, while matrix <bold>
<italic>V</italic>
</bold>
<sub>LCC</sub> is used to compute the output by combining them with the weight coefficients. At this point, the weight coefficients between <bold>
<italic>Q</italic>
</bold>
<sub>LCC</sub> and <bold>
<italic>K</italic>
</bold>
<sub>LCC</sub> represent the weight coefficients between elements of the <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub> matrix, and these coefficients can also be obtained using a scoring function:<disp-formula id="e4">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>Softmax</mml:mtext>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">Q</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>A</italic>
</bold>
<sub>LCC</sub> represents the weight coefficient matrix of the existing sample data; Softmax is a normalization function, and <italic>s</italic>(&#xb7;) is the attention scoring function. Since both matrices <bold>
<italic>Q</italic>
</bold>
<sub>LCC</sub> and <bold>
<italic>K</italic>
</bold>
<sub>LCC</sub> are also sequential data and involve calculating vector weight coefficients between matrices, the scaled dot-product formula is similarly used in <xref ref-type="disp-formula" rid="e4">Equation 4</xref>.</p>
<p>Since the weight coefficient matrix <bold>
<italic>A</italic>
</bold>
<sub>LCC</sub> obtained from <xref ref-type="disp-formula" rid="e4">Equation 4</xref> only represents the correlations between matrix elements, to obtain a matrix with enhanced data features, it is necessary to perform matrix multiplication between <bold>
<italic>A</italic>
</bold>
<sub>LCC</sub> and <bold>
<italic>V</italic>
</bold>
<sub>LCC</sub>. This results in the data feature matrix:</p>
<p>Since the weight coefficient matrix <bold>
<italic>A</italic>
</bold>
<sub>LCC</sub> obtained from <xref ref-type="disp-formula" rid="e4">Equation 4</xref> only represents the correlations between matrix elements, to obtain a matrix with enhanced data features, it is necessary to perform matrix multiplication between <bold>
<italic>A</italic>
</bold>
<sub>LCC</sub> and <bold>
<italic>V</italic>
</bold>
<sub>LCC</sub>, resulting in the data feature matrix:<disp-formula id="e5">
<mml:math id="m6">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">V</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">A</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>H</italic>
</bold>&#x2b9;<sub>LCC</sub> is the data feature matrix, with the same rows and columns as matrix <bold>
<italic>H</italic>
</bold>
<sub>LCC</sub>.</p>
<p>Furthermore, by substituting the enhanced data feature matrix <bold>
<italic>H</italic>
</bold>&#x2b9;<sub>LCC</sub> of <xref ref-type="disp-formula" rid="e5">Equation 5</xref> into <xref ref-type="disp-formula" rid="e3">Equation 3</xref> for calculation, the weight coefficient matrix <bold>
<italic>&#x3b1;</italic>
</bold>&#x2b9;<sub>
<italic>m</italic>
</sub> between the m-th similar data sub-matrix and the data feature matrix can be obtained. On this basis, to reduce the impact of the partial inconsistency between the data features of matrices <bold>
<italic>H</italic>
</bold>&#x2b9;<sub>LCC</sub> and <bold>
<italic>H</italic>
</bold>
<italic>m</italic> LC on prediction accuracy, matrix <bold>
<italic>&#x3b1;</italic>
</bold>&#x2b9;<sub>
<italic>m</italic>
</sub> is multiplied by matrix <bold>
<italic>H</italic>
</bold>
<italic>m</italic> LCC. This yields the LCC data matrix for GIS equipment in substations after expanding the sample data in <xref ref-type="disp-formula" rid="e6">Equation 6</xref>.<disp-formula id="e6">
<mml:math id="m7">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mn>1</mml:mn>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#x22ef;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi>m</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">H</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>Y</italic>
</bold>
<sub>LCC</sub> represents the expanded sample data matrix.</p>
<p>Thus, the final LCC sample data expansion model based on the self-attention mechanism is constructed, as shown in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>LCC sample data expansion model based on self-attention mechanism.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g002.tif">
<alt-text content-type="machine-generated">Diagram of a self-attention mechanism in a neural network. It includes matrices \(H_{LCC}\), \(Q_{LCC}\), \(K_{LCC}\), \(V_{LCC}\), and operations like softmax, multiplication, and transformation to \(Y_{LCC}\). Arrows indicate data flow, with attention on matrix transformations and softmax applications. Components are labeled, showing complex interactions.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-2">
<label>3.2</label>
<title>Data prediction based on GRU</title>
<p>Since there are anomalies in the LCC data, directly using the expanded sample data for prediction would still affect the accuracy of data predictions. Therefore, to fundamentally improve prediction accuracy, it is necessary to handle the anomalies in the LCC data. Given that anomalies in LCC data only appear at specific points in the time series, classifying them as point anomalies (<xref ref-type="bibr" rid="B6">Li et al., 2023</xref>), a weighted moving average method is employed to process the expanded sample data to mitigate the impact of point anomalies in the original time series:<disp-formula id="e7">
<mml:math id="m8">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="{" close="" separators="&#x7c;">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mstyle>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mi>i</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mstyle>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#x22ef;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">Y</mml:mi>
<mml:mtext>LCC</mml:mtext>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>Y</italic>
</bold>
<italic>i</italic> LCC represents the expanded sample data value for the <italic>i-th</italic> year; <italic>w</italic>
<sub>
<italic>i</italic>
</sub> denotes the data weight for the <italic>i-th</italic> year.; <italic>i</italic>&#x2208;[1, T];<italic>t</italic> is the year for which the weighted moving average method is used for data processing; and Y&#x2b9; LCC is the expanded sample data matrix after processing.</p>
<p>On this basis, considering that the recurrent structure of the RNN algorithm can effectively capture the overall trend of time-series data, the RNN algorithm is introduced to predict the processed expanded sample data (<xref ref-type="bibr" rid="B21">Yang et al., 2023</xref>). Matrix <bold>
<italic>Y</italic>
</bold>&#x2b9; LCC is used as the data input, and the RNN algorithm performs network training and prediction on the input data through its gradient calculation process.<disp-formula id="e8">
<mml:math id="m9">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:mi>V</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x220f;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:mi>V</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mspace width="1.9em"/>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x220f;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mi>tan</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">h</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mo>&#xb7;</mml:mo>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:mi>V</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<p>In this expression, <italic>L</italic>
<sub>
<italic>t</italic>
</sub> represents the loss function of the RNN network at time t; <italic>y</italic>
<sub>
<italic>k</italic>
</sub> is the output of the RNN network at time k; <italic>h</italic>
<sub>
<italic>j</italic>
</sub> is the hidden state of the RNN network&#x2019;s hidden layer at time <italic>j</italic>; tanh is the activation function of the RNN hidden layer; <italic>V</italic> is the weight parameter from the RNN hidden layer to other layers; and <italic>U</italic> is the weight parameter between hidden layers of the RNN.</p>
<p>From the linear multiplicative relationship in <xref ref-type="disp-formula" rid="e8">Equation 8</xref>, it can be seen that if tanh<sup>&#x2b9;</sup>&#xb7;<italic>U</italic> is less than 1, the gradient will converge exponentially to zero as the time span of the input LCC data increases. Conversely, if tanh<sup>&#x2b9;</sup>&#xb7;<italic>U</italic> is greater than 1, the gradient will tend towards infinity as the time span increases. Therefore, during gradient calculations over long time spans, gradient vanishing or exploding issues may arise, resulting in decreased prediction accuracy. To address this issue, considering that the LCC data for GIS equipment in substations does not belong to a large dataset, and to facilitate computation, a GRU gating structure is added to the RNN algorithm&#x2019;s architecture. This modification mitigates the gradient vanishing and exploding problems caused by the linear multiplicative relationship.</p>
<p>The network structure with the added GRU gating structure can be expressed as follows in <xref ref-type="disp-formula" rid="e9">Equation 9</xref>:<disp-formula id="e9">
<mml:math id="m10">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="{" close="" separators="&#x7c;">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">r</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">&#x3c3;</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">U</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mo>1</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">b</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">z</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">&#x3c3;</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="italic">z</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">U</mml:mi>
<mml:mi mathvariant="italic">z</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mo>1</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">b</mml:mi>
<mml:mi mathvariant="italic">z</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>tanh</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="italic">h</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">U</mml:mi>
<mml:mi mathvariant="italic">h</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mo>1</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2299;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">r</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">b</mml:mi>
<mml:mi mathvariant="italic">h</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mo>1</mml:mo>
<mml:mo>&#x2010;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2299;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2010;</mml:mo>
<mml:mo>1</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:msub>
<mml:mo>&#x2299;</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msubsup>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>
</p>
<p>In this expression, <bold>
<italic>r</italic>
</bold>
<sub>
<italic>t</italic>
</sub> is the reset gate at the current time step; <bold>
<italic>z</italic>
</bold>
<sub>
<italic>t</italic>
</sub> is the update gate at the current time step; <bold>
<italic>h</italic>
</bold>
<sup>&#x2b9;</sup>
<sub>
<italic>t</italic>
</sub> represents the candidate state information at the current time step; <bold>
<italic>x</italic>
</bold>
<sub>
<italic>t</italic>
</sub> is the input information at the current time step; <bold>
<italic>h</italic>
</bold>
<sub>
<italic>t</italic>
</sub> is the hidden state information passed to the next time step; <bold>
<italic>h</italic>
</bold>
<sub>
<italic>t</italic>-1</sub> is the hidden state information from the previous time step; <italic>&#x3c3;</italic> is the sigmoid activation function; tanh is the activation function; &#x2299; denotes the Hadamard product, which is the element-wise multiplication of matrices; <bold>
<italic>W</italic>
</bold>
<sub>r</sub>, <bold>
<italic>W</italic>
</bold>
<sub>z</sub>, <bold>
<italic>W</italic>
</bold>
<sub>h</sub>, <bold>
<italic>U</italic>
</bold>
<sub>r</sub>, <bold>
<italic>U</italic>
</bold>
<sub>z</sub>, <bold>
<italic>U</italic>
</bold>
<sub>h</sub> are the weight parameter matrices corresponding to the reset gate, update gate, and candidate state; <bold>
<italic>b</italic>
</bold>
<sub>r</sub>, <bold>
<italic>b</italic>
</bold>
<sub>z</sub>, <bold>
<italic>b</italic>
</bold>
<sub>h</sub> are the bias parameter matrices for the reset gate, update gate, and candidate state, respectively.</p>
<p>Thus, the multiplicative relationship expression for gradient calculation in the GRU structure becomes:<disp-formula id="e10">
<mml:math id="m11">
<mml:mrow>
<mml:mfenced open="{" close="" separators="&#x7c;">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x220f;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x220f;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>U</mml:mi>
<mml:mi>z</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi>U</mml:mi>
<mml:mi>h</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>U</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mspace width="0.17em"/>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>U</mml:mi>
<mml:mi>h</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mspace width="0.17em"/>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>r</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3b4;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x2202;</mml:mi>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mspace width="0.17em"/>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>
</p>
<p>In this expression: &#x3b4;<sub>
<italic>zj</italic>
</sub> is the partial derivative of the loss function concerning the update gate; &#x3b4;<sub>
<italic>rj</italic>
</sub> is the partial derivative of the loss function concerning the reset gate; &#x3b4;<sub>
<italic>hj</italic>
</sub> is the partial derivative of the loss function concerning the hidden state information.</p>
<p>By substituting the multiplicative terms in <xref ref-type="disp-formula" rid="e10">Equation 10</xref> into <xref ref-type="disp-formula" rid="e8">Equation 8</xref>, it can be observed that the original linear multiplicative relationship in <xref ref-type="disp-formula" rid="e8">Equation 8</xref> has been replaced by a nonlinear relationship. Consequently, the gradient vanishing and exploding issues in data prediction are resolved. At this point, the processed expanded sample data matrix can be used as input to the network, and data prediction is performed through the GRU network.</p>
<p>After obtaining the prediction values from the GRU network, the most commonly used metrics in time series prediction&#x2014;Root Mean Square Error (RMSE) in <xref ref-type="disp-formula" rid="e11">Equation 11</xref>, Mean Absolute Error (MAE) in <xref ref-type="disp-formula" rid="e12">Equation 12</xref>, and <italic>R</italic>
<sup>2</sup> (R-Square) in <xref ref-type="disp-formula" rid="e13">Equation 13</xref>&#x2014;are selected for error validation:<disp-formula id="e11">
<mml:math id="m12">
<mml:mrow>
<mml:mtext>RMSE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:msqrt>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>
<disp-formula id="e12">
<mml:math id="m13">
<mml:mrow>
<mml:mtext>MAPE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mfenced open="|" close="|" separators="&#x7c;">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>100</mml:mn>
<mml:mo>%</mml:mo>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>
<disp-formula id="e13">
<mml:math id="m14">
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="normal">R</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:msubsup>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:msubsup>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="&#x7c;">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(13)</label>
</disp-formula>
</p>
<p>In these expressions: <italic>y</italic>
<sub>
<italic>i</italic>
</sub> represents the true value of the sample; <inline-formula id="inf2">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the predicted value;<italic>n</italic> is the number of true values in the sample, <inline-formula id="inf3">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the mean of the predicted values.</p>
</sec>
<sec id="s3-3">
<label>3.3</label>
<title>Data prediction process</title>
<p>Combining the above analyses of the self-attention mechanism and GRU, the process begins by applying weight processing to similar data based on the self-attention mechanism to obtain expanded sample data. Then, the expanded sample data is used to train model parameters with the GRU model. Finally, the LCC data prediction results are output, and the accuracy of the prediction results is validated. The complete LCC data prediction workflow in <xref ref-type="fig" rid="F3">Figure 3</xref> is as follows.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>LCC data prediction flowchart.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g003.tif">
<alt-text content-type="machine-generated">Flowchart illustrating a process beginning with substation GIS equipment data collection and sorting. It involves sample size expansion using self-attention and weighted methods. The extended sample data matrix is divided into training and test sets. The training set adjusts GRU model parameters. Iterations continue until optimal or maximum numbers are reached. Once complete, the model parameters are determined, and test sets evaluate performance. The process ends with a finish stage.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Example verification and analysis</title>
<sec id="s4-1">
<label>4.1</label>
<title>Actual conditions of the substation</title>
<p>Taking the 110 kV GIS equipment of a substation in Henan Province as an example, this substation is selected as the sample substation, and the SA-GRU model proposed in this paper is applied to predict the LCC data of GIS equipment in this sample substation. The substation currently has two main transformers, and the 110 kV side distribution device uses outdoor GIS equipment. It is located in an E-grade pollution area. The basic information on the voltage level, environment, and scale of the sample substation is shown in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Basic information of the sample substation.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Item</th>
<th align="center">Numerical value</th>
<th align="center">Unit</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Voltage level</td>
<td align="center">220</td>
<td align="center">kV</td>
</tr>
<tr>
<td align="center">Number of main converters</td>
<td align="center">2</td>
<td align="center">Set number</td>
</tr>
<tr>
<td align="center">Principal variable capacity</td>
<td align="center">180</td>
<td align="center">MVA</td>
</tr>
<tr>
<td align="center">110 kV incoming and outgoing lines</td>
<td align="center">12</td>
<td align="center">Circuit number</td>
</tr>
<tr>
<td align="center">GIS equipment layout</td>
<td align="center">Outdoors</td>
<td align="center">&#x2014;</td>
</tr>
<tr>
<td align="center">Altitude</td>
<td align="center">&#x3c;1,000</td>
<td align="center">m</td>
</tr>
<tr>
<td align="center">Pollution grade</td>
<td align="center">E</td>
<td align="center">Level</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To demonstrate that the SA-GRU model proposed in this paper is suitable for predicting the LCC data of GIS equipment in substations, three similar substations are introduced to expand the LCC data samples based on the basic conditions, such as the voltage level, electrical scale, and surrounding environment of the example substation. The basic information of the introduced similar substations is shown in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Basic information of similar substations.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Item</th>
<th align="center">Substation 1</th>
<th align="center">Substation 2</th>
<th align="center">Substation 3</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Voltage level</td>
<td align="center">220 kV</td>
<td align="center">220 kV</td>
<td align="center">220 kV</td>
</tr>
<tr>
<td align="center">Number of main converters/unit</td>
<td align="center">3</td>
<td align="center">3</td>
<td align="center">2</td>
</tr>
<tr>
<td align="center">Main variable capacity/MVA</td>
<td align="center">240</td>
<td align="center">180</td>
<td align="center">180</td>
</tr>
<tr>
<td align="center">110 kV incoming/outgoing line/return</td>
<td align="center">12</td>
<td align="center">16</td>
<td align="center">10</td>
</tr>
<tr>
<td align="center">GIS equipment layout</td>
<td align="center">Outdoors</td>
<td align="center">Outdoors</td>
<td align="center">Outdoors</td>
</tr>
<tr>
<td align="center">Altitude m</td>
<td align="center">&#x3c;1,000</td>
<td align="center">&#x3c;1,000</td>
<td align="center">&#x3c;1,000</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Comparing the basic information of substations in <xref ref-type="table" rid="T1">Tables 1</xref> and <xref ref-type="table" rid="T2">2</xref> reveals that the three introduced substations have the same voltage level as the sample substation, with altitudes below 1,000 m, are located in an E-grade pollution area, and all GIS equipment is arranged outdoors. Substations 1 and 2 have one more main transformer than the sample substation, while Substation 3 has the same number of main transformers as the sample substation. In terms of incoming and outgoing line circuits, Substation 1 has the same number as the sample substation, Substation 2 has 4 more, and Substation 3 has 2 fewer circuits. Among these, Substation 3 has the highest similarity to the sample substation in terms of basic information, while Substations 2 and 1 have lower similarity.</p>
<p>After identifying the sample substation and similar substations, raw LCC data were obtained through the PMS and ERP system platforms and multidimensional lean reports. This data collection primarily focused on cost-related data, including voltage level, service life, asset original value, routine maintenance and repair material costs, major repair material costs, maintenance fees, labor costs, and other expenses. As of 31 December 2023, a total of 886 LCC sample data entries for GIS equipment were collected for the sample substation. Additionally, 4,101 LCC sample data entries were gathered for GIS equipment from the three similar substations, facilitating the expansion of LCC data samples.</p>
<p>A portion of the collected sample data on cost-related information for 110 kV GIS equipment is shown in <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Sample of cost-related data for 110 kV GIS equipment (partial).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Serial number</th>
<th align="center">Operation age limit</th>
<th align="center">Original value of assets</th>
<th align="center">Maintenance and repair materials cost</th>
<th align="center">Cost of major repair materials</th>
<th align="center">Repair charge</th>
<th align="center">Labor cost</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td align="center">38</td>
<td align="center">143435.27</td>
<td align="center">414.30</td>
<td align="center">1</td>
<td align="center">2959.91</td>
<td align="center">2631.77</td>
</tr>
<tr>
<td align="center">2</td>
<td align="center">32</td>
<td align="center">359302.77</td>
<td align="center">2100.87</td>
<td align="center">10676.74</td>
<td align="center">588.70</td>
<td align="center">35819.53</td>
</tr>
<tr>
<td align="center">3</td>
<td align="center">18</td>
<td align="center">1557504.77</td>
<td align="center">1515.03</td>
<td align="center">1515.03</td>
<td align="center">3680.27</td>
<td align="center">15061.98</td>
</tr>
<tr>
<td align="center">4</td>
<td align="center">23</td>
<td align="center">363904.65</td>
<td align="center">222.28</td>
<td align="center">0</td>
<td align="center">4864.15</td>
<td align="center">5059.33</td>
</tr>
<tr>
<td align="center">5</td>
<td align="center">4</td>
<td align="center">171930.08</td>
<td align="center">708.70</td>
<td align="center">0</td>
<td align="center">0</td>
<td align="center">4484.22</td>
</tr>
<tr>
<td align="center">6</td>
<td align="center">12</td>
<td align="center">873264.77</td>
<td align="center">134.56</td>
<td align="center">0</td>
<td align="center">1139.65</td>
<td align="center">2584.88</td>
</tr>
<tr>
<td align="center">7</td>
<td align="center">2</td>
<td align="center">124368.33</td>
<td align="center">7468.46</td>
<td align="center">5193.79</td>
<td align="center">7509.21</td>
<td align="center">1856.92</td>
</tr>
<tr>
<td align="center">8</td>
<td align="center">9</td>
<td align="center">645125.57</td>
<td align="center">8415.71</td>
<td align="center">9456.00</td>
<td align="center">1180.20</td>
<td align="center">53565.84</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To clearly present the data foundation of this study, <xref ref-type="table" rid="T4">Table 4</xref> lists representative samples of the LCC time-series data collected from a 110 kV GIS unit at a substation in Henan Province. The dataset contains cost records spanning the equipment&#x2019;s entire life cycle, reflecting the continuity and complexity characteristic of time-series data.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>LCC time-series data collected from a 110 kV GIS unit at a substation in Henan Province.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Serial number</th>
<th align="center">LCC cost (Yuan)</th>
<th align="center">Serial number</th>
<th align="center">LCC cost (Yuan)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td align="center">10253.92</td>
<td align="center">11</td>
<td align="center">21117.94</td>
</tr>
<tr>
<td align="center">2</td>
<td align="center">18594.72</td>
<td align="center">12</td>
<td align="center">22019.09</td>
</tr>
<tr>
<td align="center">3</td>
<td align="center">17078.66</td>
<td align="center">13</td>
<td align="center">18771.44</td>
</tr>
<tr>
<td align="center">4</td>
<td align="center">20873.63</td>
<td align="center">14</td>
<td align="center">24503.27</td>
</tr>
<tr>
<td align="center">5</td>
<td align="center">21922.14</td>
<td align="center">15</td>
<td align="center">25276.93</td>
</tr>
<tr>
<td align="center">6</td>
<td align="center">25626.14</td>
<td align="center">16</td>
<td align="center">28541.20</td>
</tr>
<tr>
<td align="center">7</td>
<td align="center">18292.75</td>
<td align="center">17</td>
<td align="center">28401.98</td>
</tr>
<tr>
<td align="center">8</td>
<td align="center">23898.47</td>
<td align="center">18</td>
<td align="center">30419.14</td>
</tr>
<tr>
<td align="center">9</td>
<td align="center">24465.10</td>
<td align="center">19</td>
<td align="center">30956.61</td>
</tr>
<tr>
<td align="center">10</td>
<td align="center">27341.49</td>
<td align="center">20</td>
<td align="center">25864.47</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The GRU model was configured with the following hyperparameters: 64 hidden units, a look-back window (time steps) of 10, a dropout rate of 0.3, a batch size of 16, the Adam optimizer with an initial learning rate of 0.001, and 200 training epochs. These hyperparameters were selected through preliminary tuning based on validation performance, and early stopping was employed during training to prevent overfitting.</p>
</sec>
<sec id="s4-2">
<label>4.2</label>
<title>Weight calculation results based on self-attention mechanism</title>
<p>Based on the LCC sample data expansion model using the self-attention mechanism, the influence weights of LCC data for GIS equipment in different substations can be obtained. A comparison of these weights with those obtained using the direct attention mechanism is shown in <xref ref-type="fig" rid="F4">Figure 4</xref>.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Influence weights obtained by different models.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g004.tif">
<alt-text content-type="machine-generated">Bar chart comparing impact weights obtained by different methods for various substations. Sample Substation shows weights of 0.568 for self-attention and 0.526 for attention. Substation 1 has weights of 0.115 and 0.120, Substation 2 has 0.102 and 0.103, and Substation 3 shows 0.195 and 0.151. Dark blue represents self-attention, light blue represents attention.</alt-text>
</graphic>
</fig>
<p>As shown in <xref ref-type="fig" rid="F4">Figure 4</xref>, the influence weights based on the self-attention mechanism for the sample substation, Substation 1, Substation 2, and Substation 3 are 0.588, 0.115, 0.102, and 0.195, respectively. The weight results indicate that, apart from the sample substation, Substation 3 has the highest weight, likely due to the similarity of its main transformer count, main transformer capacity, and the number of incoming and outgoing circuits with those of the sample substation, differing only by 2 fewer circuits. The weights of Substation 1 and Substation 2 are nearly 50% lower than that of Substation 3, likely because Substation 1 has a higher main transformer count and capacity, and Substation 2 has a higher main transformer count and a larger number of incoming and outgoing circuits than the sample substation. These quantitative differences lead to higher LCC data values for Substations 1 and 2. Substation 2 has the lowest weight, possibly because the number of incoming and outgoing circuits on the 110 kV side has a greater impact on the LCC data values.</p>
<p>Comparing the influence weights of the attention mechanism in <xref ref-type="fig" rid="F5">Figure 5</xref>, it can be seen that the weight of the sample substation based on the self-attention mechanism is reduced by 0.038 compared to the attention mechanism. This may be because the attention mechanism directly uses the raw sample data as sample features, whereas the self-attention mechanism extracts sample features from the raw data, leading to some differences between the two types of sample features. Except for Substation 3, the influence weights of the other similar substations based on the self-attention mechanism are all reduced. This may be because the LCC data features of Substation 3 are closer to the features extracted by the self-attention mechanism. Additionally, Substation 2 still has the smallest weight result, while Substation 3 still has the largest.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Comparison of training results from different models.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g005.tif">
<alt-text content-type="machine-generated">Line chart titled &#x22;LCC forecast results&#x22; comparing four models: True value, GRU, GRU-ATT, and Textual. The x-axis represents sample points from one to eleven, and the y-axis shows LCC in yuan, ranging from 2.8 to 4.4 (&#xD7;10&#x2074;). Each model's results are plotted with distinct line styles and markers, showing increasing trends with varying deviations.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s4-3">
<label>4.3</label>
<title>Comparative analysis of training results from different models and methods</title>
<p>Based on the typical lifespan of GIS equipment in substations, 40 years is selected as the overall time frame for LCC, with predictions made annually. Considering that the LCC data for substation GIS equipment, even after sample expansion, still constitutes a small dataset, no validation set is created. Instead, the sample data processed in <xref ref-type="disp-formula" rid="e7">Equation 7</xref> is split into a training set and a test set in a traditional 7:3 ratio. The training set is used to train model parameters, and the test set is used to validate the model&#x2019;s training results. To maintain continuity in the time-series data during training, a segmented approach is adopted for dividing the dataset: the first 32 consecutive years of data are used for model parameter training, and the last 8 consecutive years of data are used for validation. The divided training set is fed as input to the GRU model, with only LCC values as the output. Considering the components of the model in this paper, the SA-GRU model, ATT-GRU model (Attention combined with GRU model), and GRU model are selected for a comparative analysis of the training results from different models. The training results of the different models on the LCC data of substation GIS equipment are shown in <xref ref-type="fig" rid="F5">Figure 5</xref>, and the evaluation metrics RMSE, MAE, and <italic>R</italic>
<sup>2</sup> are presented in <xref ref-type="table" rid="T4">Table 4</xref>.</p>
<p>As seen in <xref ref-type="fig" rid="F5">Figure 5</xref>, from an overall training perspective, while the GRU model&#x2019;s general trend aligns with the true values, the discrepancies between its predictions and the actual values are noticeably larger than those of the SA-GRU and ATT-GRU models. The ATT-GRU model, which is an improved version of the GRU model with the inclusion of similar data during modeling, shows certain advantages in small-sample data prediction. Although its performance is not as accurate as the SA-GRU model, the overall prediction results are still better than the GRU model, closely approximating the true values, with only slight deviations at certain sample points and towards the end of the training. The SA-GRU model used in this paper provides results that are even closer to the true values compared to the ATT-GRU model. This improvement is due to the SA-GRU model&#x2019;s use of a self-attention mechanism to enhance sample data features, which increases the weights assigned to similar substations that closely match the sample data features. This adjustment reduces the negative impact on prediction accuracy caused by inconsistencies between the features of similar data and sample data.</p>
<p>From <xref ref-type="table" rid="T5">Table 5</xref>, it can be observed that the RMSE of the SA-GRU model is 715.93 yuan, the MAPE is 1.54%, and the <italic>R</italic>
<sup>2</sup> value is 94.86%; the RMSE of the ATT-GRU model is 1039.18 yuan, the MAPE is 2.35%, and the <italic>R</italic>
<sup>2</sup> value is 91.78%; while the GRU model&#x2019;s RMSE is 1677.33 yuan, the MAPE is 3.83%, and the <italic>R</italic>
<sup>2</sup> value is 83.32%. Compared to the ATT-GRU model, the SA-GRU model improves the RMSE accuracy by 31.11% and the MAPE accuracy by 34.47%, with the <italic>R</italic>
<sup>2</sup> value increasing from 91.78% to 94.86%.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>RMSE, MAE, and <italic>R</italic>
<sup>2</sup> metrics results for different models and different methods.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Prediction model</th>
<th align="center">RMSE/&#x5143;</th>
<th align="center">MAPE</th>
<th align="center">
<italic>R</italic>
<sup>2</sup>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">GRU model</td>
<td align="center">1677.33</td>
<td align="center">3.83%</td>
<td align="center">83.32%</td>
</tr>
<tr>
<td align="center">The ATT-GRU model</td>
<td align="center">1039.18</td>
<td align="center">2.35%</td>
<td align="center">91.78%</td>
</tr>
<tr>
<td align="center">Optimize partial least squares</td>
<td align="center">1388.11</td>
<td align="center">3.37%</td>
<td align="center">87.31%</td>
</tr>
<tr>
<td align="center">The fuzzy neural network method</td>
<td align="center">1160.57</td>
<td align="center">2.73%</td>
<td align="center">85.68%</td>
</tr>
<tr>
<td align="center">Textual method</td>
<td align="center">715.93</td>
<td align="center">1.54%</td>
<td align="center">94.86%</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To validate the accuracy of the proposed method for LCC data prediction, the proposed method was compared with the optimized partial least squares method and the fuzzy neural network method mentioned in the introduction. The training results are shown in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Comparison of training results from different methods.</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g006.tif">
<alt-text content-type="machine-generated">Line graph titled &#x22;LCC forecast results&#x22; comparing four models. The y-axis represents LCC in yuan and the x-axis represents sample points. The graph includes lines for &#x22;True value&#x22; (purple, solid), &#x22;Fuzzy neural network&#x22; (blue, dashed), &#x22;Optimize the least squares model&#x22; (orange, dashed), and &#x22;Textual model&#x22; (light blue, dotted). Lines show similar upward trends with varying fluctuations, indicating close correspondence among the models' predictions.</alt-text>
</graphic>
</fig>
<p>According to the evaluation index calculation formulas, the RMSE, MAE, and <italic>R</italic>
<sup>2</sup> results of different prediction methods are presented in <xref ref-type="table" rid="T4">Table 4</xref>. Combined with <xref ref-type="fig" rid="F6">Figure 6</xref>, it is evident that among time-series prediction methods, the SA-GRU model demonstrates superior training performance and evaluation results compared to other methods. The optimized partial least squares (PLS) method is significantly influenced by numerical fluctuations during training, resulting in an inability to accurately capture data trends. Although the fuzzy neural network method mitigates the impact of numerical fluctuations on training, it shows weaker capability in capturing data features. These limitations cause both methods to yield less accurate prediction results than the proposed method in this paper. As shown in <xref ref-type="table" rid="T4">Table 4</xref>, compared to the optimized PLS method, the proposed method reduces the RMSE by 672.18 yuan, decreases the MAPE by 1.83%, and improves the <italic>R</italic>
<sup>2</sup> by 7.55%. Compared to the fuzzy neural network method, the proposed method reduces the RMSE by 444.64 yuan, decreases the MAPE by 1.19%, and improves the <italic>R</italic>
<sup>2</sup> by 9.18%.</p>
<p>To visualize the predictive performance of the best model, <xref ref-type="fig" rid="F7">Figure 7</xref> plots observed versus predicted LCC on the held-out test set. The 45&#xb0; bisector (y &#x3d; x) is shown as the ideal reference, together with an ordinary-least-squares (OLS) fit of predicted on observed values. The legend reports <italic>R</italic>
<sup>2</sup>, RMSE, MAE, and MAPE to complement the visual assessment.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Observed vs. predicted LCC (Test Set, SA-GRU).</p>
</caption>
<graphic xlink:href="fenrg-13-1533878-g007.tif">
<alt-text content-type="machine-generated">Scatter plot depicting observed versus predicted LCC values in yuan, using the SA-GRU model. It includes a bisector line and an OLS trend line with equation \( \hat{y} &#x3d; -2403 + 1.048 \cdot x \). Key metrics: R-squared is 0.764, RMSE is 1549 yuan, MAE is 1157 yuan, and MAPE is 3.30%.</alt-text>
</graphic>
</fig>
<p>In <xref ref-type="fig" rid="F7">Figure 7</xref>, points cluster around the bisector. The OLS fit <inline-formula id="inf4">
<mml:math id="m17">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2403</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1.048</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> has a slope close to 1 and a small intercept, indicating limited systematic bias. Quantitatively, on the test set we obtain <inline-formula id="inf5">
<mml:math id="m18">
<mml:mrow>
<mml:msup>
<mml:mi>R</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.764</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, RMSE &#x3d; 1,549 yuan, MAE &#x3d; 1,157 yuan, and MAPE &#x3d; 3.30%, which is consistent with Table Y. The slope slightly above 1 suggests a mild overestimation at higher LCC levels (and slight underestimation for low-cost years), but the deviation is small relative to the overall range. Overall, SA-GRU provides accurate and decision-grade LCC forecasts.</p>
<p>In summary, the SA-GRU model proposed in this paper can effectively improve the accuracy of LCC data prediction for GIS equipment in substations.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<label>5</label>
<title>Conclusion</title>
<p>Considering the limited sample size and presence of outlier data in the GIS equipment LCC data of substations, this study proposes an SA-GRU-based prediction method for LCC data. Taking the 110 kV GIS equipment of a substation in Henan Province as an example, the conclusions are as follows:<list list-type="order">
<list-item>
<p>The SA-GRU model outperforms the Attention-GRU model and the GRU model in terms of RMSE, MAPE, and <italic>R</italic>
<sup>2</sup> evaluation metrics. Models incorporating self-attention mechanisms also demonstrate superior performance compared to those using only attention mechanisms, confirming the necessity of using self-attention to process data from similar substations.</p>
</list-item>
<list-item>
<p>Compared with traditional optimized partial least squares (PLS) and fuzzy neural network methods, the proposed method reduces the RMSE by 672.18 yuan and 444.64 yuan, decreases the MAPE by 1.83% and 1.19%, and increases the <italic>R</italic>
<sup>2</sup> by 7.55% and 9.18%, respectively. These results validate the effectiveness of the proposed method in improving the accuracy of GIS equipment LCC data prediction and addressing the challenges in accurately forecasting LCC data for substations.</p>
</list-item>
<list-item>
<p>The proposed method quantifies the influence weights of data from the sample substation and similar substations to obtain weighted data from similar substations, thereby expanding the sample size. However, the current approach to weight calculation does not consider the qualitative characteristics of individual substations. Future research could focus on quantifying the scale and environmental characteristics of substations and developing appropriate methods to further enhance the reliability of weight calculations for substation interactions.</p>
</list-item>
</list>
</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>WX: Writing &#x2013; original draft, Writing &#x2013; review and editing. CA: Conceptualization, Data curation, Resources, Writing &#x2013; review and editing. TS: Conceptualization, Funding acquisition, Project administration, Supervision, Writing &#x2013; review and editing. PD: Data curation, Formal Analysis, Investigation, Software, Writing &#x2013; review and editing. YZ: Writing &#x2013; review and editing, Methodology, Resources, Visualization. JZ: Formal Analysis, Project administration, Supervision, Writing &#x2013; review and editing.</p>
</sec>
<ack>
<title>Acknowledgements</title>
<p>This paper thanks all the authors for their dedicated work and the financial support of various institutions. And thanks for the comments from all reviewers.</p>
</ack>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of interest</title>
<p>Authors WX, CA, TS, PD, YZ, and JZ were employed by State Grid Henan Electric Power Company.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn fn-type="custom" custom-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1031506/overview">Manoj Kumar Nallapaneni</ext-link>, Maastricht University, Netherlands</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/615262/overview">Kenneth E. Okedu</ext-link>, Melbourne Institute of Technology, Australia</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3095100/overview">Javier Tarr&#xed;o-Saavedra</ext-link>, Universidade da Coru&#xf1;a, Spain</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>A review of life-cycle cost (LCC) technology applications in power systems</article-title>. <source>Power Syst. Prot. Control</source> <volume>39</volume> (<issue>17</issue>), <fpage>149</fpage>&#x2013;<lpage>154</lpage>. <pub-id pub-id-type="doi">10.7667/j.issn.1674-3415.2011.17.029</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheng</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Short-term wind-farm forecasting feature selection and multi-level deep transfer learning</article-title>. <source>High. Volt. Eng.</source> <volume>48</volume> (<issue>2</issue>), <fpage>497</fpage>&#x2013;<lpage>503</lpage>. <pub-id pub-id-type="doi">10.13336/j.1003-6520.hve.20210032</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Du</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Life-cycle cost analysis and management of underground substations based on fuzzy theory</article-title>,&#x201d; in <source>Master&#x2019;s thesis</source>. <publisher-name>Shanghai Jiao Tong University</publisher-name>.</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Quality-related fault detection using orthogonal signal correction and efficient partial least squares</article-title>. <source>Control Decis.</source> <volume>35</volume> (<issue>5</issue>), <fpage>1167</fpage>&#x2013;<lpage>1174</lpage>. <pub-id pub-id-type="doi">10.13195/j.kzyjc.2018.0708</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Time-series forecasting with multi-time-scale RNN</article-title>. <source>Comput. Appl. Softw.</source> <volume>35</volume> (<issue>7</issue>), <fpage>33</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.3969/j.issn.1000-386x.2018.07.006</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Tian</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Key technologies of health management and intelligent O&#x26;M for power equipment in the new power system</article-title>. <source>Power Syst. Technol.</source>, <fpage>1</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.13335/j.1000-3673.pst.2022.2451</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020a</year>). <article-title>LCC prediction for substations in frigid high-latitude regions using SSA-LS-SVM</article-title>. <source>Smart Power</source> <volume>48</volume> (<issue>6</issue>), <fpage>54</fpage>&#x2013;<lpage>60</lpage>.</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2020b</year>). <article-title>Fuzzy estimation model of LCC for 110 kV GIS substations based on fuzzy smoothing</article-title>. <source>J. Guangxi Normal Univ. (Natural Science Edition)</source> <volume>38</volume> (<issue>5</issue>), <fpage>24</fpage>&#x2013;<lpage>33</lpage>. <pub-id pub-id-type="doi">10.16088/j.issn.1001-6600.2020.05.003</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Peng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Methods to improve short-term and ultra-short-term wind-power forecasting accuracy: a review</article-title>. <source>Proc. CSEE</source> <volume>36</volume> (<issue>23</issue>), <fpage>6315</fpage>&#x2013;<lpage>6326&#x2b;6596</lpage>. <pub-id pub-id-type="doi">10.13334/j.0258-8013.pcsee.161167</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qiao</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Substation LCC prediction model via GA-tuned LS-SVM</article-title>. <source>China Electr. Power</source> <volume>48</volume> (<issue>11</issue>), <fpage>142</fpage>&#x2013;<lpage>148</lpage>.</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Intelligent statistical algorithm for equipment cost of power-grid projects based on physical ID</article-title>. <source>Power Grid Clean Energy</source> <volume>36</volume> (<issue>2</issue>), <fpage>68</fpage>&#x2013;<lpage>74</lpage>.</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Su</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Substation life-cycle cost planning based on GIS spatial analysis and improved particle swarm optimization</article-title>. <source>Proc. CSEE</source> <volume>32</volume> (<issue>16</issue>), <fpage>92</fpage>&#x2013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.13334/j.0258-8013.pcsee.2012.16.013</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Concepts, models and methods for short-term wind-power forecasting</article-title>. <source>Electr. Power Eng. Technol.</source> <volume>37</volume> (<issue>5</issue>), <fpage>7</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.19464/j.cnki.cn32-1541/tm.2018.05.002</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Geng</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Life-cycle cost prediction of substations based on optimized orthogonal partial least squares</article-title>. <source>Smart Power</source> <volume>48</volume> (<issue>5</issue>), <fpage>119</fpage>&#x2013;<lpage>124</lpage>. <pub-id pub-id-type="doi">10.3969/j.issn.1673-7598.2020.05.019</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Life-cycle cost modeling and sensitivity analysis for an unattended 110 kV substation in a coal mine</article-title>. <source>China Min. Mag.</source> <volume>30</volume> (<issue>S1</issue>), <fpage>68</fpage>&#x2013;<lpage>73</lpage>.</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Yin</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Few-data wind-power forecasting considering spatial coupling</article-title>. <source>South. Power Syst. Technol.</source> <volume>16</volume> (<issue>6</issue>), <fpage>75</fpage>&#x2013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.13648/j.cnki.issn1674-0629.2022.06.008</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021a</year>). <article-title>Substation LCC prediction using quantum-behaved particle swarm optimization and LS-SVM</article-title>. <source>Electr. Meas. Instrum.</source> <volume>58</volume> (<issue>6</issue>), <fpage>76</fpage>&#x2013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.19753/j.issn1001-1390.2021.06.011</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ke</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2021b</year>). <article-title>Prediction of substation maintenance and O&#x26;M cost using an improved BP neural network</article-title>. <source>J. Electr. Power Sci. Technol.</source> <volume>36</volume> (<issue>4</issue>), <fpage>44</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.19781/j.issn.1673-9140.2021.04.006</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xue</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Ultra-short-term wind-power prediction with a CNN-GRU hybrid model</article-title>. <source>Renew. Energy</source> <volume>37</volume> (<issue>3</issue>), <fpage>456</fpage>&#x2013;<lpage>462</lpage>. <pub-id pub-id-type="doi">10.13941/j.cnki.21-1469/tk.2019.03.023</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Ying</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Estimation of life-cycle cost for substation construction</article-title>. <source>Comput. Simul.</source> <volume>34</volume> (<issue>1</issue>), <fpage>123</fpage>&#x2013;<lpage>128</lpage>.</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Short-term wind-power forecasting with neighboring wind-farm information and CNN-BiLSTM</article-title>. <source>South. Power Syst. Technol.</source> <volume>17</volume> (<issue>2</issue>), <fpage>47</fpage>&#x2013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.13648/j.cnki.issn1674-0629.2023.02.006</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ye</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Pei</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Combination forecasting for wind power under cold-wave and small-sample conditions</article-title>. <source>Proc. CSEE</source> <volume>43</volume> (<issue>2</issue>), <fpage>543</fpage>&#x2013;<lpage>555</lpage>. <pub-id pub-id-type="doi">10.13334/j.0258-8013.pcsee.221814</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yin</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ou</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A transfer-learning approach for wind-power prediction based on a serio-parallel deep-learning architecture</article-title>. <source>Energy</source> <volume>234</volume>, <fpage>121171</fpage>. <pub-id pub-id-type="doi">10.1016/j.energy.2021.121271</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2021</year>). <source>Research on attention mechanism in neural networks</source>. <publisher-name>University of Science and Technology of China</publisher-name>. <pub-id pub-id-type="doi">10.27517/d.cnki.gzkju.2021.000623</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Scenario data transfer for new-energy generation via conditional deep convolutional GAN</article-title>. <source>Power Syst. Technol.</source> <volume>46</volume> (<issue>6</issue>), <fpage>2182</fpage>&#x2013;<lpage>2190</lpage>. <pub-id pub-id-type="doi">10.13335/j.1000-3673.pst.2021.1008</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Upper-bound assessment of GIS equipment failures using time-series models and deep learning</article-title>. <source>J. Jilin Univ. (Engineering Technol. Edition)</source>, <fpage>1</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.13229/j.cnki.jdxbgxb.20230307</pub-id>
</mixed-citation>
</ref>
</ref-list>
</back>
</article>