<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Energy Res.</journal-id>
<journal-title>Frontiers in Energy Research</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Energy Res.</abbrev-journal-title>
<issn pub-type="epub">2296-598X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">856635</article-id>
<article-id pub-id-type="doi">10.3389/fenrg.2022.856635</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Energy Research</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Anomaly Detection of Hydropower Units Based on Recurrent Neural Network</article-title>
<alt-title alt-title-type="left-running-head">Xiong et al.</alt-title>
<alt-title alt-title-type="right-running-head">Anomaly Detection of Hydropower Units</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Xiong</surname>
<given-names>Lei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Liu</surname>
<given-names>Jiajun</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1639198/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yang</surname>
<given-names>Feng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Gang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Dang</surname>
<given-names>Jian</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1620624/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Electrical Engineering</institution>, <institution>Xi&#x2019;an University of Technology</institution>, <addr-line>Xi&#x2019;an</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Xi&#x2019;an Key Laboratory of Intelligent Energy</institution>, <institution>Xi&#x2019;an University of Technology</institution>, <addr-line>Xi&#x2019;an</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1259467/overview">Liansong Xiong</ext-link>, Nanjing Institute of Technology (NJIT), China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1646149/overview">Yu Zhou</ext-link>, Taiyuan University of Technology, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1646783/overview">Rize Jin</ext-link>, Tianjin Polytechnic University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1615134/overview">Yingjie Wang</ext-link>, University of Liverpool, United Kingdom</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Jiajun Liu, <email>xautljj@163.com</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Process and Energy Systems Engineering, a section of the journal Frontiers in Energy Research</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>07</day>
<month>04</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>10</volume>
<elocation-id>856635</elocation-id>
<history>
<date date-type="received">
<day>17</day>
<month>01</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>24</day>
<month>02</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 Xiong, Liu, Yang, Zhang and Dang.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Xiong, Liu, Yang, Zhang and Dang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Anomaly detection for hydraulic turbine unit has an important role in hydropower system. In hydropower systems, different components will produce n-dimensional heterogeneous time series with different characteristics at all times. Due to the characteristic evolution and time dependence, vibration-based anomaly detection for hydraulic turbine unit is extremely challenging. In this paper, we propose a conditional quantile regression based recurrent neural network (QRNN), which models the time dependence and probability distribution between random variables. The proposed method aims to extract the actual representation patterns from the fitted models and it can effectively detect anomalies in the non-uniform time series of feature evolution. The experimental results show that the proposed method has better accuracy in anomaly detection (error reduction by 34%) than the traditional method, and saves at least 25.6% of execution time.</p>
</abstract>
<kwd-group>
<kwd>anomaly detection</kwd>
<kwd>hydropower system</kwd>
<kwd>n-dimensional heterogeneous time series</kwd>
<kwd>quantile regression based recurrent neural network (QRNN)</kwd>
<kwd>characteristic evolution</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>In order to control the stability and security, data analysis and detection methods play an important role in power systems (<xref ref-type="bibr" rid="B22">Liu et al., 2017</xref>; <xref ref-type="bibr" rid="B37">Xiong et al., 2019</xref>; <xref ref-type="bibr" rid="B36">Xiong et al., 2021</xref>). Different from the wind power system (<xref ref-type="bibr" rid="B23">Liu et al., 2020</xref>), the operation process of a hydropower unit is complex. According to the statistics, the vibration signal of the unit can reflect more than 80% of the fault characteristics. Therefore, we propose to detect the fault of hydropower turbine unit based on mining the vibration signal data in this paper. Due to highly nonlinear, instability and time-varying characteristics of the vibration signal data of the hydraulic turbine unit, it is difficult to be modeled precisely. At present, the non-stationary vibration signal processing methods include short-time Fourier transform, Mode Decomposition and so on. In order to solve the problems of empirical mode decomposition on mathematical theory, Konstantin et al. (<xref ref-type="bibr" rid="B14">Iyer et al., 2016</xref>) proposed variational mode decomposition method. Assuming that the signal is composed of modal functions with different center frequencies, it is decomposed by self-adaptive and quasi-orthogonal methods in the variational framework. Breiman et al. (<xref ref-type="bibr" rid="B3">Breiman, 2001</xref>) proposed the random forest theory. Based on decision tree, this method generates several decision trees by randomly feature selection and samples. After each decision tree is trained separately, the final sample category is voted on. However, they are not suitable for processing the multivariate vibration time series data of hydraulic turbine unit.</p>
<p>Extreme value theory (EVT) (<xref ref-type="bibr" rid="B38">Xu et al., 2018</xref>), Peak over-threshold (POT) (<xref ref-type="bibr" rid="B13">Hundman et al., 2018</xref>), and distance-based (<xref ref-type="bibr" rid="B6">Dragomiretskiy and Zosso, 2013</xref>) can be used for time series anomaly detection (<xref ref-type="bibr" rid="B35">Wu et al., 2014</xref>). All these methods may involve a large number of distribution assumptions, parameter adjustments, and heavy calculations in the conversion process. This may increase the calculation cost. Although there are many studies on anomaly detection (<xref ref-type="bibr" rid="B25">Mehrotra et al., 2017</xref>), they either focus on deterministic methods (<xref ref-type="bibr" rid="B7">Filonov et al., 2016</xref>) or random methods (<xref ref-type="bibr" rid="B5">Dasgupta and Osogami, 2017</xref>; <xref ref-type="bibr" rid="B20">Lai et al., 2018</xref>), and ignore time dependence of the time series (<xref ref-type="bibr" rid="B11">Guha et al., 2016</xref>). In (<xref ref-type="bibr" rid="B24">Manzoor et al., 2018</xref>), the goal is to identify independent anomalous objects rather than identifying anomalous time series patterns based on their time dependence. In time series modeling, historical observations are very important for understanding current data. Therefore, it is better to use a sequence of observations: X_(t-T), X_(t-T&#x2b;1), &#x2026;, X_t rather than X_t to calculate the anomaly score in time series.</p>
<p>Vector auto regression (VAR) and support vector regression (SVR) can be applied to multivariate time series (<xref ref-type="bibr" rid="B20">Lai et al., 2018</xref>). However, many of these models are difficult to scale up and include exogenous variables (<xref ref-type="bibr" rid="B23">Liu et al., 2020</xref>). Supervised methods (<xref ref-type="bibr" rid="B21">Laptev et al., 2015</xref>; <xref ref-type="bibr" rid="B32">Shipmon et al., 2017</xref>) require labeled data for model training, and can only identify known types of anomalies. The unsupervised method does not need to label the data, which can be divided into two categories: the deterministic model and random model. For deterministic model. In order to capture both the long-term dynamics and short-term effects simultaneously, the non-linear dynamics modeling, long short term memory (LSTM) methods (<xref ref-type="bibr" rid="B12">Hochreiter and Schmidhuber, 1997</xref>; <xref ref-type="bibr" rid="B9">Gers et al., 2000</xref>; <xref ref-type="bibr" rid="B1">Borovykh et al., 2017</xref>; <xref ref-type="bibr" rid="B10">Grob et al., 2018</xref>) have been proposed for forecasting events based on sequence models. For deterministic model, (<xref ref-type="bibr" rid="B7">Filonov et al., 2016</xref>) proposed a LSTM-based predictive models to detect spacecraft anomalies. Although LSTMs are deterministic and have no random variables, they can handle the time dependence of heterogeneous time series. For random model, recurrent neural network (<xref ref-type="bibr" rid="B5">Dasgupta and Osogami, 2017</xref>; <xref ref-type="bibr" rid="B20">Lai et al., 2018</xref>) is used for time series anomaly detection. Auto-encoder-based methods are used for time series anomaly detection in (<xref ref-type="bibr" rid="B2">Boudiaf et al., 2016</xref>; <xref ref-type="bibr" rid="B38">Xu et al., 2018</xref>). Variants of convolution and recurrent neural networks are used for modeling temporal patterns (<xref ref-type="bibr" rid="B4">Calvo-Bascones et al., 2021</xref>). Deep convolutional neural networks (<xref ref-type="bibr" rid="B15">Kim, 2014</xref>; <xref ref-type="bibr" rid="B39">Yang et al., 2015</xref>) have been used for time series human activity recognition. Memory guided normality for anomaly detection is proposed in (<xref ref-type="bibr" rid="B28">Park et al., 2020</xref>). Deep learning based anomaly detection methods for video and industrial control system are proposed in <xref ref-type="bibr" rid="B34">Wang et al. (2020)</xref> and <xref ref-type="bibr" rid="B26">Nayak et al. (2021)</xref>, respectively. The deep neural network (<xref ref-type="bibr" rid="B31">Sen et al., 2019</xref>), (<xref ref-type="bibr" rid="B30">Salinas et al., 2017</xref>) has been proposed to high dimensional time series forecasting, however, it is limited to training on entire time series and then perform multi-step ahead forecast which is practically computational resource demanding in feature evolving heterogeneous time series. Recently, (<xref ref-type="bibr" rid="B27">Pang et al., 2021</xref>) give a survey of deep learning methods for anomaly detection. However, in the hydraulic turbine system, different subsystems and parts will produce <italic>n</italic>-dimensional heterogeneous time series with different characteristics at all times. This will form a heterogeneous time series of feature evolution.</p>
<p>In this paper, we propose a quantile regression (<xref ref-type="bibr" rid="B8">Geraci and Bottai, 2007</xref>) based recurrent neural network (QRNN) for anomaly detection of hydropower units. The proposed method can explicitly model the temporal dependence among stochastic variables with properly estimated probability distributions. Considering the feature evolution of heterogeneous time series, the most difficult part in this method is still how to detect anomalies. The most intuitive way to verify the reliability of the detection model is to find the correct estimated distribution probability. The higher the probability, the higher the confidence. But this does not mean higher accuracy. The proposed method aims to extract the actual representational patterns from the fitted models, thereby studying their reconstruction error probability distribution by using a conditional quantile regression method to limit mistakes when it is forced to make decision on normality by guidance of principled uncertainty estimates. Inspired by the distribution assumptions-free QR (<xref ref-type="bibr" rid="B16">Koenker and Bassett Jr, 1978</xref>; <xref ref-type="bibr" rid="B18">Koenker and Machado, 1999</xref>; <xref ref-type="bibr" rid="B17">Koenker and Hallock, 2001</xref>) build on asymmetric laplace distribution (<xref ref-type="bibr" rid="B19">Kotz et al., 2001</xref>), we proposed QRNN to model the conditional distribution <inline-formula id="inf1">
<mml:math id="m1">
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> is the predicted value. The contributions of this paper can be summarized as follows:<list list-type="simple">
<list-item>
<p>&#x2022; According to the characteristics of hydraulic turbine data, we propose a quantile regression based recurrent neural network for anomaly detection, which can model temporal dependencies and stochasticity explicitly in vibration time series data. And it can study the reconstruction error probability distribution by using a conditional quantile regression method.</p>
</list-item>
<list-item>
<p>&#x2022; We propose an incremental online method to update the co-evolving heterogeneous for multiple streaming data.</p>
</list-item>
<list-item>
<p>&#x2022; We propose to apply stochastic methods to detect anomalies in heterogeneous time series with feature evolution, which is proved robust and powerful in the experiment.</p>
</list-item>
</list>
</p>
<p>The rest of this paper is organized as follows. In <xref ref-type="sec" rid="s2">Section 2</xref>, we introduce the problem and the related model used in this paper. In <xref ref-type="sec" rid="s3">Section 3</xref>, we describe the proposed method in detail. In <xref ref-type="sec" rid="s4">Section 4</xref>, we conduct several experiments to evaluate the advantage of the proposed method. At the end, we conclude the paper in <xref ref-type="sec" rid="s5">Section 5</xref>.</p>
</sec>
<sec id="s2">
<title>2 Problem and Model</title>
<sec id="s2-1">
<title>2.1 Formal Description of the Problem</title>
<p>Formalize the anomaly detection problem in the feature evolution time series, [<italic>T</italic>
<sub>0</sub>, <italic>T</italic>) represents the observation window of the observed event. In general, it can be assumed that <italic>T</italic>
<sub>0</sub> &#x3d; 0. Each feature F uses a continuous time stamp sequence <italic>T</italic>
<sub>
<italic>u</italic>
</sub> &#x3d; <italic>t</italic>
<sub>1</sub>, <italic>&#x2026;</italic>, <italic>t</italic>
<sub>
<italic>n</italic>
</sub>. <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>:<italic>t</italic>
</sub> &#x2208; <italic>R</italic>
<sup>
<italic>M</italic>&#xd7;(<italic>T</italic>&#x2b;1)</sup> represents the observation sequence <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>
</sub>, <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>&#x2b;1</sub>, &#x2026;, <italic>X</italic>
<sub>
<italic>t</italic>
</sub>. <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> and <italic>y</italic> represents the predicted value and the true value. L represents the loss function. In the problem of time series anomaly detection, past historical observations are very important for understanding and capturing the dynamic patterns in the current data stream. We use the observation sequence <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>:<italic>t</italic>
</sub> to calculate the anomaly score of <italic>X</italic>
<sub>
<italic>t</italic>
</sub>. Given <italic>X</italic> &#x3d; <italic>x</italic>
<sub>1</sub>, <italic>&#x2026;</italic>, <italic>x</italic>
<sub>
<italic>N</italic>
</sub>, where <italic>N</italic> is the length of the vector <italic>x</italic>. The observation <italic>x</italic>
<sub>
<italic>t</italic>
</sub> is the <italic>N</italic>-dimensional vector <inline-formula id="inf4">
<mml:math id="m4">
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and <italic>X</italic> &#x2208; <italic>R</italic>
<sup>
<italic>M</italic>&#xd7;<italic>N</italic>
</sup> at time <italic>t</italic>(<italic>t</italic> &#x2264; <italic>N</italic>). According to the past historical observation data <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>:<italic>t</italic>
</sub>, the problem of anomaly detection in vibration time series data of hydraulic turbine unit is formally defined as follows.</p>
<p>Problem: Given the historical observation sequence represented by the time series in the order given as <italic>X</italic>
<sub>
<italic>t</italic>&#x2212;<italic>T</italic>:<italic>t</italic>
</sub>, where t is the current time point, report an anomaly score for a given current data point <italic>X</italic>
<sub>
<italic>t</italic>
</sub> instantly at any time.</p>
</sec>
<sec id="s2-2">
<title>2.2 Long Short-Term Memory</title>
<p>LSTM is a kind of deep recurrent neural network. It is widely used in time series. LSTM realizes the memory function in time through the opening and closing of the door. It can effectively solve the problem of gradient disappearance and gradient explosion in general situations. The key is to introduce a gating unit system. The system stores historical information through the internal memory unit-cell state unit. Different gates can make the network know when to forget historical information, when to update cells status dynamically. Cells are cyclically connected to each other, instead of hidden units in the general cyclic network. If the input gate sigmoid allows new information input, its value can be added to the state. The state unit has linear self-circulation. Its weight is controlled by the forget gate. The output of the cell can be closed by the output gate. The status unit can also be used as an additional input for the gating unit.<disp-formula id="e1">
<mml:math id="m5">
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(1)</label>
</disp-formula>
<disp-formula id="e2">
<mml:math id="m6">
<mml:msub>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(2)</label>
</disp-formula>
<disp-formula id="e3">
<mml:math id="m7">
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>E</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(3)</label>
</disp-formula>
<disp-formula id="e4">
<mml:math id="m8">
<mml:msub>
<mml:mrow>
<mml:mi>o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>o</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>o</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="e5">
<mml:math id="m9">
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>o</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mi>E</mml:mi>
<mml:mi>L</mml:mi>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>Here f, i, o, h, C, c, w, Z, b, <italic>&#x3c3;</italic>() respectively represents forget gate, input gate, output gate, hidden state vector, unit state vector, weight matrix, feature vector, bias vector and sigmoid function. In this paper, we select to use LSTM due to the following features: <italic>1</italic>) it is able to support end to end modeling; <italic>2</italic>) it is easy to incorporate exogenous; <italic>3</italic>) it is powerful in feature extraction for vibration time series data of hydraulic turbine unit.</p>
</sec>
</sec>
<sec id="s3">
<title>3 The Proposed Method</title>
<sec id="s3-1">
<title>3.1 The Architecture of QRNN</title>
<p>This paper propose to combine LSTM and conditional quantile regression to model explicitly temporal dependencies and stochasticity in vibration time series data of hydraulic turbine unit. The overall structure of QRNN consists of two parts: <italic>1</italic>) offline training <italic>2</italic>) online anomaly detection. As shown in <xref ref-type="fig" rid="F2">Figure 2</xref>, the proposed neural network is composed of two recurrent neural networks, such as the LSTM encoder and predictor-detector network. It can be considered as <italic>f</italic> &#x3d; <italic>pd</italic>(<italic>e</italic>(<italic>h</italic>
<sub>
<italic>t</italic>
</sub>, (<italic>x</italic>,<italic>y</italic>)<sub>
<italic>t</italic>
</sub>)) where f is the predicted scores, <italic>x</italic>
<sub>
<italic>t</italic>
</sub> and <italic>y</italic>
<sub>
<italic>t</italic>
</sub> are the input variables, <italic>h</italic>
<sub>
<italic>t</italic>
</sub> is the encoder state, <italic>e</italic>(.) is an encoder and <italic>pd</italic>(.) is the predictor-detector network. The LSTM encoder is trained to extract useful temporal and non-linear patterns contained in vibration time series data, which can be used to guide the predictor-detector network of anomaly detection. If exogenous variables are available, they can get concatenated with extracted feature vectors from the LSTM encoder and used as input to the predictor-detector network. The predictor-detector network consists of an LSTM and dense layers for output prediction. The training process can be carried out daily training according to business needs, such as once a week or once a month. The offline part is composed of preprocessing sub-module (shared by online and offline modules) and start-up sub-module. The online part consists of real-time detection and update sub-modules. The data is preprocessed in the preprocessing module. The data is divided into sequences through a sliding window of length <italic>T</italic>. The startup module builds the model and deploys it to memory after training, testing, and verification. The fitted model can now perform real-time anomaly detection. In streaming or online settings, new observations after preprocessing (such as <italic>X</italic>
<sub>
<italic>t</italic>
</sub> at time <italic>t</italic>) can be sent to the detection module and provide anomaly scores. If the abnormal score of <italic>X</italic>
<sub>
<italic>t</italic>
</sub> is lower or higher than the abnormal threshold, <italic>X</italic>
<sub>
<italic>t</italic>
</sub> is marked as abnormal. The update of sub-module will update the parameters of each operation in QRNN db. The overall structure is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The proposed neural network model with LSTM encoder and predictor-detector network.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g001.tif"/>
</fig>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The overall structure of QRNN.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g002.tif"/>
</fig>
<p>QRNN is composed of LSTM encoder and prediction-detector network. The network can essentially be regarded as a large neural network, expressed as <italic>f</italic> &#x3d; <italic>pd</italic>(<italic>e</italic>(<italic>h</italic>
<sub>
<italic>t</italic>
</sub>, (<italic>x</italic>,<italic>y</italic>)<sub>
<italic>t</italic>
</sub>)), where f is the prediction score, <italic>x</italic>
<sub>
<italic>t</italic>
</sub> and <italic>y</italic>
<sub>
<italic>t</italic>
</sub> are input variables, <italic>h</italic>
<sub>
<italic>t</italic>
</sub> is the state of the encoder, <italic>e</italic>(.) is the ana encoder, and <italic>pd</italic>(.) is the prediction-detection network. The trained LSTM encoder can extract useful time and nonlinear dynamic patterns contained in the heterogeneous time series of feature evolution. These patterns can be used to guide the prediction-detection network to perform anomaly detection.</p>
<p>If exogenous variables are available, they can be connected with the feature vector extracted from the LSTM encoder and used as input to the predictor-detector network. This is because in reality, it is not necessary to observe all the content needed to predict the output through the input. After the LSTM encoder network is trained, the output from the last cell state is sent to the prediction-detection network. This is also the training prediction score evolution data stream <inline-formula id="inf5">
<mml:math id="m10">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, where <italic>h</italic>(.) is the hidden coding state, as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. The proposed QRNN prediction-detection network consists of an LSTM layer and a fully connected output prediction layer. The last dense layer is modeled as three outputs and three loss functions. One for the desired fitting model. The other two are used to guide the upper and lower bounds of forecast uncertainty estimates. It is worth noting that the total loss is calculated as the sum of individual quantile losses. And the output is all quantile predictions of different quantile values defined by <italic>&#x3c4;</italic> &#x2208; [0, 1].</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>QRNN architecture diagram.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g003.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>3.2 Fault Detection and Loss Function</title>
<p>QRNN is an anomaly detector based on prediction. The anomaly detection depends on the quality of the predicted value. When predicting normal data, the probability distribution of the error can be calculated, and then used to find the maximum posterior probability estimate of the normal behavior of the test data. In order to predict more accurately, it is necessary to set thresholds for the upper and lower bounds. Beyond this threshold, data points can be marked as abnormal, as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. In the deep learning regression task setting, the mean square error function is the most commonly used loss function.<disp-formula id="e6">
<mml:math id="m11">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Schematic diagram of anomaly detection.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g004.tif"/>
</fig>
<p>Intuitively speaking, taking the negative power of the mean square error function approximates the Gaussian process whose mode corresponds to the mean parameter. Non-parametric distributions like conditional quantiles are very useful in quantifying uncertainty estimation in decision-making and minimizing risk.</p>
<p>The goal of Quantile regression (QR) is to estimate the conditional median or any other quantile of the distribution. It can be done by solving the following formula.<disp-formula id="e7">
<mml:math id="m12">
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(7)</label>
</disp-formula>Where the function <italic>&#x3c1;</italic>
<sub>
<italic>r</italic>
</sub>(.) is the tilted absolute value function that yields the <italic>&#x3c4;</italic>th sample quantiles. To obtain an estimate of the conditional median function, we replace the scalar <italic>&#x3be;</italic> by the parametric function <italic>&#x3be;</italic>(<italic>x</italic>
<sub>
<italic>i</italic>
</sub>, <italic>&#x3b2;</italic>), which can be formulated as a linear function of parameters.<disp-formula id="e8">
<mml:math id="m13">
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<p>Quantile regression learns to predict the conditional quantiles <inline-formula id="inf6">
<mml:math id="m14">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for the given target distribution <italic>via</italic> formula <xref ref-type="disp-formula" rid="e9">(9)</xref>, in which <inline-formula id="inf7">
<mml:math id="m15">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is the predicted value at the given quantile.<disp-formula id="e9">
<mml:math id="m16">
<mml:mi>P</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2264;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
</mml:math>
<label>(9)</label>
</disp-formula>
</p>
<p>We focus on putting weights to the distances between points on the distribution function and the fitted regression line based on the selected quantile. We select to use QR due to it has the following key features: <italic>1</italic>) It does not make any distribution assumption on the error; <italic>2</italic>) QR can describe the outcome variable of the entire conditional distribution; <italic>3</italic>) QR is more robust to outliers and setting errors of the error distribution; <italic>4</italic>) QR can expand the flexibility of parametric and non-parametric regression methods.</p>
<p>We set <italic>&#x3c4;</italic> &#x2208; [0, 1] to generate the predicted value <italic>y</italic>
<sub>
<italic>t</italic>&#x2b;1:<italic>M</italic>
</sub> with the smallest reconstruction error <italic>&#x3be;</italic>, as shown in the formula <xref ref-type="disp-formula" rid="e6">(6)</xref> in the predictive detector network. And calculate the quantile loss of a single data point by formula <xref ref-type="disp-formula" rid="e10">(10)</xref>. Because we needed a complete conditional distribution rather than a point estimate of uncertainty estimation, the average loss function <italic>L</italic>(.) on a given data distribution can be defined by formula <xref ref-type="disp-formula" rid="e11">(11)</xref>, where <italic>f</italic>
<sup>
<italic>W</italic>
</sup>(<italic>x</italic>
<sub>
<italic>t</italic>
</sub>) is the fitting model under the given quantile <italic>&#x3c4;</italic>, <italic>y</italic>
<sub>
<italic>t</italic>
</sub> and <italic>x</italic>
<sub>
<italic>t</italic>
</sub> are true values.<disp-formula id="e10">
<mml:math id="m17">
<mml:mi>L</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="array">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mi>&#x3c4;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd columnalign="center"/>
<mml:mtd columnalign="left">
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd columnalign="center"/>
<mml:mtd columnalign="left">
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(10)</label>
</disp-formula>
<disp-formula id="e11">
<mml:math id="m18">
<mml:mi>L</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi>L</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(11)</label>
</disp-formula>
</p>
<p>In the case of a loss function, the prediction-detector network is modeled as three outputs and three loss functions, with the lower quantile, the median (fitting model) and the upper quantile. At <italic>&#x3c4;</italic> &#x3d; 0.5, the loss function will estimate the median value instead of the average value. With the upper quantile and the lower quantile, a reliable uncertainty estimate can be provided for forecasting.</p>
</sec>
<sec id="s3-3">
<title>3.3 Time Series Evolution</title>
<p>It needs to be considered that during the operation of the turbine, the data update of each subsystem arrives in the form of a stream over time. This may include updates to existing features, or new features, as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. Generally, a time series may have endogenous variables (for example, the output is a function of the previous output), or may not have exogenous variables that are not affected by other variables in the system. But the output depends on it. Most work ignores exogenous variables. But in order to improve quality and improve anomaly detection, this article introduces exogenous variables. When performing anomaly detection on the time series of various hydraulic turbines, sometimes there may be patterns that were not previously available. It deviates greatly from the training model. This may be a training data set observed in a specific mode. In general, it requires that each time a new data set is reached, the entire process of training the model must start from the beginning. For hydraulic turbine time series fault detection models, training the entire model requires a large amount of data sets and a large amount of time. In actual situations, it is impractical to keep training the model with the arrival of new data. In order to solve this problem, an online incremental update scheme is needed to provide threshold and model updates. QRNN can receive previously unavailable characteristics or evolved data points for learning, without starting from scratch, so as to mark abnormal data points and update model parameters in real time, as shown in formula <xref ref-type="disp-formula" rid="e12">(12)</xref> and <xref ref-type="disp-formula" rid="e13">(13)</xref>, where <italic>th</italic>
<sub>
<italic>up</italic>
</sub> and <italic>th</italic>
<sub>
<italic>lo</italic>
</sub> are the upper and lower thresholds of the abnormal score calculated by the selected quantile, <italic>&#x3b1;</italic>1 and <italic>&#x3b1;</italic>2 are the higher and lower bounds, respectively, and <inline-formula id="inf8">
<mml:math id="m19">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is the fitted model at the corresponding quantile (<italic>&#x3c4;</italic> &#x3d; 0.5 corresponds to the median).<disp-formula id="e12">
<mml:math id="m20">
<mml:mi>t</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>u</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:math>
<label>(12)</label>
</disp-formula>
<disp-formula id="e13">
<mml:math id="m21">
<mml:mi>t</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:math>
<label>(13)</label>
</disp-formula>
</p>
<p>Based on the conditional quantile regression, with the emergence of a new data set, the confidence of the model parameter distribution will be automatically updated. The threshold value will be incrementally updated, as described in the loss function above. With this setting, QRNN can perform anomaly detection in vibration time series data of hydraulic turbine unit.</p>
<p>
<xref ref-type="fig" rid="F5">Figure 5</xref> shows the overall algorithm of QRNN. For preprocessing, offline training, evaluation, testing and detection is performed. The fitted model <inline-formula id="inf9">
<mml:math id="m22">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is loaded into memory ready for anomaly detection which is performed in formula <xref ref-type="disp-formula" rid="e12">(12)</xref> and <xref ref-type="disp-formula" rid="e13">(13)</xref>. After the traversal is completed, the fitted model is saved to QRNNdb. QRNN is design to operate with multiple settings depending on whether exogenous features are present. During fitting, different processing needs to be performed according to whether the exogenous feature (XF) is existed. During this procedure, LSTM encoder can be used as feature extractor. Upon the arrival of vibration time series data of hydraulic turbine unit, the proposed QRNN will predict its value and seamlessly flag its anomaly score. At the same time, the model parameters will be updated. The prediction, model updates and abnormal scoring can be completed in a single pass at O(1) time complexity.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>The overall algorithm of QRNN.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g005.tif"/>
</fig>
</sec>
</sec>
<sec id="s4">
<title>4 Evaluation</title>
<p>In order to verify the efficiency of the proposed method, we adopted the hydraulic turbine working simulation data set provided by Case Western Reserve University (CWRU) to perform the experiment and evaluation. The test bench consists of a two-horsepower motor, a torque sensor/encoder, and an electronic device for power measurement and control. The vibration data is collected by an accelerometer. The accelerometer is placed at the 12 o&#x2019;clock position of the drive end and the fan end of the motor housing, which using a 16-channel DAT recorder (<xref ref-type="bibr" rid="B33">Smith and Randall, 2015</xref>). Experimental data of the drive end is collected at 12,000 sampling points per second and 48,000 sampling points per second. The fan data is collected at 12,000 sample per second (<xref ref-type="bibr" rid="B2">Boudiaf et al., 2016</xref>). The detailed attributes of the data set used in the experiment are shown in <xref ref-type="table" rid="T1">Table 1</xref>. We use python and Keras framework to build neural network. The experimental setting is CPU 2.30&#xa0;GHz, 8G DDR4 3200&#xa0;MHZ, and GeForce GTX 1050Ti GPU.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Data set attributes.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Loads</th>
<th align="center">Rotating speed (rpm)</th>
<th align="center">Sequence length</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0HP</td>
<td align="center">1,797</td>
<td align="center">2,43,938</td>
</tr>
<tr>
<td align="left">1HP</td>
<td align="center">1,772</td>
<td align="center">4,83,903</td>
</tr>
<tr>
<td align="left">2HP</td>
<td align="center">1,750</td>
<td align="center">4,83,903</td>
</tr>
<tr>
<td align="left">3HP</td>
<td align="center">1,730</td>
<td align="center">4,83,903</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Adam optimizer is used in this method, which is an extension of SGD. In this algorithm, an adaptive gradient algorithm is introduced to adjust the learning rate. The errors of different methods on different load data sets are shown in <xref ref-type="table" rid="T2">Table 2</xref>. &#x201c;1HP&#x201d; in the table represents that there is a load in the device. In the QRNN method, the ability of anomaly detection largely depends on trend prediction, so the accuracy of trend prediction is very important. It can be seen from the table that the QRNN method has a greater improvement in the accuracy of trend prediction compared to other methods. Compared with the errors of CNN, GRU and BiLSTM, the QRNN method has reduced 41, 44, and 65% respectively.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Mean square error of trend prediction of QRNN, CNN, GRU, and BiLSTM methods on data sets under different loads.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">CNN</th>
<th align="center">GRU</th>
<th align="center">BiLSTM</th>
<th align="center">QRNN</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0HP</td>
<td align="center">1.261 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.326 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">2.082 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.122 &#xd7; 10<sup>&#x2013;4</sup>
</td>
</tr>
<tr>
<td align="left">1HP</td>
<td align="center">1.217 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.345 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">2.139, &#xd7;, 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.056 &#xd7; 10<sup>&#x2013;4</sup>
</td>
</tr>
<tr>
<td align="left">2HP</td>
<td align="center">1.155 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.272 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.947 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.021 &#xd7; 10<sup>&#x2013;4</sup>
</td>
</tr>
<tr>
<td align="left">3HP</td>
<td align="center">1.268 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.333 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">1.445 &#xd7; 10<sup>&#x2013;4</sup>
</td>
<td align="center">0.909 &#xd7; 10<sup>&#x2013;4</sup>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>One-dimensional CNN can be well applied to the time series analysis of sensor data (<xref ref-type="bibr" rid="B29">Roska and Chua, 1993</xref>) (such as gyroscope or accelerometer data). It can also be used to analyze signal data with a fixed length period (such as audio signals). It can well identify simple patterns in the data, and then use these patterns to generate more complex patterns in higher-level layers. It is generally feasible to use one-dimensional convolution to process time series. The results of CNN are shown in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>CNN for anomaly detection.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g006.tif"/>
</fig>
<p>Long short-term memory network (LSTM) was born to solve the problem of long-term dependence. As a variant of LSTM, GRU combines the forget gate and input gate into a single update gate. It mixes the cell state and the hidden state, and adds some other changes. The final model is simpler than the standard LSTM model. BiLSTM is also a variant of LSTM, which takes context into account. The results of GRU and BiLSTM are shown in <xref ref-type="fig" rid="F7">Figures 7</xref>, <xref ref-type="fig" rid="F8">8</xref>, respectively.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>GRU for anomaly detection.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g007.tif"/>
</fig>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>BiLSTM for anomaly detection.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g008.tif"/>
</fig>
<p>The main body of the QRNN framework is composed of LSTM encoders. LSTM network is used to solve the vanishing gradient problem. It supports end-to-end modeling and nonlinear feature extraction. In the process of training the neural network, the error margin has a direct impact on the accuracy of the model, especially the time. This is a hyper parameter and needs to be specified before training. The basis for the formulation is the convergence speed of the model and the learning accuracy of the sample. When it is too large, neither the training accuracy nor the test accuracy is high enough to meet actual needs. If it is too small, although the learning effect is more ideal, it takes a lot of time. The parameter is set between 10<sup>&#x2013;4</sup> 10<sup>&#x2013;2</sup> generally. When the training error is lower than this value, the training is considered sufficient, and the training should be stopped. The results of the QRNN are shown in <xref ref-type="fig" rid="F9">Figure 9</xref>.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>QRNN for anomaly detection.</p>
</caption>
<graphic xlink:href="fenrg-10-856635-g009.tif"/>
</fig>
<p>We count TP (True Positive), FP (False Positive), FN (False Negative), and TN (True Negative) based on the experimental data. TP represents the number of abnormal points marked by the neural network that are correctly marked. FP means that it is actually a normal point but is marked as an abnormal point. FN represents the number of normal points marked as abnormal points. TN represents the number of normal points that are marked correctly. We calculate the precision and recall rate of each method based on TP, FP, FN, and TN. The precision rate <inline-formula id="inf10">
<mml:math id="m23">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is used to measure the degree to which the method finds the right way, and the recall rate <inline-formula id="inf11">
<mml:math id="m24">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> measures the degree to which the method is found to be complete. As shown in <xref ref-type="table" rid="T3">Table 3</xref>, it can be seen that the QRNN has the highest accuracy and recall rates (95.9%, 94.3%). In addition, although the structure of QRNN is relatively complex and the training time is longer than CNN and GRU, the overall running time (125.3 MS) is significantly shorter than them.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>TP, FP, FN, TN value, precision, recall and running time of each method.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">TP</th>
<th align="center">FP</th>
<th align="center">FN</th>
<th align="center">TN</th>
<th align="center">Precision (%)</th>
<th align="center">Recall (%)</th>
<th align="center">Time (ms)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">CNN</td>
<td align="center">323</td>
<td align="center">53</td>
<td align="center">102</td>
<td align="center">9,522</td>
<td align="char" char=".">85.9</td>
<td align="char" char=".">76.0</td>
<td align="char" char=".">154.7</td>
</tr>
<tr>
<td align="left">GRU</td>
<td align="center">367</td>
<td align="center">37</td>
<td align="center">58</td>
<td align="center">9,538</td>
<td align="char" char=".">90.8</td>
<td align="char" char=".">86.4</td>
<td align="char" char=".">164.3</td>
</tr>
<tr>
<td align="left">BiLSTM</td>
<td align="center">379</td>
<td align="center">27</td>
<td align="center">46</td>
<td align="center">9,548</td>
<td align="char" char=".">93.3</td>
<td align="char" char=".">89.2</td>
<td align="char" char=".">159.4</td>
</tr>
<tr>
<td align="left">QRNN</td>
<td align="center">398</td>
<td align="center">17</td>
<td align="center">27</td>
<td align="center">9,558</td>
<td align="char" char=".">95.9</td>
<td align="char" char=".">93.6</td>
<td align="char" char=".">125.3</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s5">
<title>5 Conclusion</title>
<p>Anomaly detection of evolution time series is important for hydraulic turbine systems. This paper introduces a vibration-based anomaly detection method. It can deal with time-dependent and non-linear complex dynamic sequences. As a window-based anomaly detection method, it has scalability, high-efficiency stream processing efficiency and can effectively deal with the heterogeneity and randomness in the constantly changing data stream. After experimental evaluation based on real data sets, it indicates that the proposed method is fast, robust and accurate compared to the traditional anomaly detection methods.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The data underlying this article are available in the article and in its online Supplementary Material.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>LX wrote the first draft of the paper, JL directed the paper writing and provided project support for the paper, and FY, GZ and the JD put forward many valuable amendments in the paper writing.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This research was funded by National Natural Science Foundation of China (52009106); Natural Science Basic Research Plan of Shaanxi Province (2019JQ-130).</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Borovykh</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bohte</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Oosterlee</surname>
<given-names>C. W.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Conditional Time Series Forecasting with Convolutional Neural Networks</article-title>,&#x201d; in <source>Proceedings of the International Conference on Artificial Neural Networks</source>. <publisher-loc>United States</publisher-loc>: <publisher-name>arXiv</publisher-name>. </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boudiaf</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Moussaoui</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dahane</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Atoui</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A Comparative Study of Various Methods of Bearing Faults Diagnosis Using the Case Western reserve university Data</article-title>. <source>J. Fail. Anal. Preven.</source> <volume>16</volume>, <fpage>271</fpage>&#x2013;<lpage>284</lpage>. <pub-id pub-id-type="doi">10.1007/s11668-016-0080-7</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Breiman</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2001</year>). <article-title>Random Forests</article-title>. <source>Machine Learn.</source> <volume>45</volume>, <fpage>5</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1023/a:1010933404324</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Calvo-Bascones</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Sanz-Bobi</surname>
<given-names>M. A.</given-names>
</name>
<name>
<surname>Welte</surname>
<given-names>T. M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Anomaly Detection Method Based on the Deep Knowledge behind Behavior Patterns in Industrial Components. Application to a Hydropower Plant</article-title>. <source>Comput. Industry</source> <volume>125</volume>, <fpage>103376</fpage>. <pub-id pub-id-type="doi">10.1016/j.compind.2020.103376</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dasgupta</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Osogami</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Nonlinear Dynamic Boltzmann Machines for Time-Series Prediction</article-title>,&#x201d; in <source>The AAAI Conference on Artificial Intelligence</source>. <publisher-loc>San Francisco, United States</publisher-loc>: <publisher-name>AAAI Press</publisher-name>, <volume>31</volume>. </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dragomiretskiy</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zosso</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Variational Mode Decomposition</article-title>. <source>IEEE Transactions Signal. Processing</source> <volume>62</volume>, <fpage>531</fpage>&#x2013;<lpage>544</lpage>. <pub-id pub-id-type="doi">10.1109/TSP.2013.2288675</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Filonov</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Lavrentyev</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Vorontsov</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Multivariate Industrial Time Series with Cyber-Attack Simulation: Fault Detection Using an Lstm-Based Predictive Data Model</article-title>. <comment>NIPS.</comment> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Geraci</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bottai</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Quantile Regression for Longitudinal Data Using the Asymmetric Laplace Distribution</article-title>. <source>Biostatistics</source> <volume>8</volume>, <fpage>140</fpage>&#x2013;<lpage>154</lpage>. <pub-id pub-id-type="doi">10.1093/biostatistics/kxj039</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gers</surname>
<given-names>F. A.</given-names>
</name>
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Cummins</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Learning to Forget: Continual Prediction with Lstm</article-title>. <source>Neural Comput.</source> <volume>12</volume>, <fpage>2451</fpage>&#x2013;<lpage>2471</lpage>. <pub-id pub-id-type="doi">10.1162/089976600300015015</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Grob</surname>
<given-names>G. L.</given-names>
</name>
<name>
<surname>Cardoso</surname>
<given-names>&#xc2;.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Little</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Chamberlain</surname>
<given-names>B. P.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>A Recurrent Neural Network Survival Model: Predicting Web User Return Time</article-title>,&#x201d; in <source>Joint European Conference on Machine Learning and Knowledge Discovery in Databases</source> (<publisher-loc>United States</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>152</fpage>&#x2013;<lpage>168</lpage>. </citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Guha</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mishra</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Roy</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Schrijvers</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Robust Random Cut forest Based Anomaly Detection on Streams</article-title>,&#x201d; in <source>International Conference on Machine Learning</source> (<publisher-loc>Canada</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>2712</fpage>&#x2013;<lpage>2721</lpage>. </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hochreiter</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>Long Short-Term Memory</article-title>. <source>Neural Comput.</source> <volume>9</volume>, <fpage>1735</fpage>&#x2013;<lpage>1780</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hundman</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Constantinou</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Laporte</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Colwell</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Soderstrom</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Detecting Spacecraft Anomalies Using Lstms and Nonparametric Dynamic Thresholding</article-title>,&#x201d; in <source>The 24th ACM SIGKDD International Conference</source>. <publisher-loc>New York, United States</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>387</fpage>&#x2013;<lpage>395</lpage>. <pub-id pub-id-type="doi">10.1145/3219819.3219845</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Iyer</surname>
<given-names>A. P.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L. E.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Stoica</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Time-evolving Graph Processing at Scale</article-title>,&#x201d; in <source>The Fourth International Workshop on Graph Data Management Experiences and Systems</source>. <publisher-loc>New York, United States</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1145/2960414.2960419</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kim</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Convolutional Neural Networks for Sentence Classification</article-title>,&#x201d; in <source>Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing</source>, <fpage>1746</fpage>&#x2013;<lpage>1751</lpage>. <pub-id pub-id-type="doi">10.3115/v1/d14-1181</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Koenker</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Bassett</surname>
<given-names>G.</given-names>
<suffix>Jr</suffix>
</name>
</person-group> (<year>1978</year>). <article-title>Regression Quantiles</article-title>. <source>Econometrica</source>, <fpage>33</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.2307/1913643</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Koenker</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Hallock</surname>
<given-names>K. F.</given-names>
</name>
</person-group> (<year>2001</year>). <article-title>Quantile Regression</article-title>. <source>J. Econ. Perspect.</source> <volume>15</volume>, <fpage>143</fpage>&#x2013;<lpage>156</lpage>. <pub-id pub-id-type="doi">10.1257/jep.15.4.143</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Koenker</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Machado</surname>
<given-names>J. A. F.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Goodness of Fit and Related Inference Processes for Quantile Regression</article-title>. <source>J. Am. Stat. Assoc.</source> <volume>94</volume>, <fpage>1296</fpage>&#x2013;<lpage>1310</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.1999.10473882</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kotz</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kozubowski</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Podg&#xf3;rski</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2001</year>). <source>The Laplace Distribution and Generalizations: A Revisit with Applications to Communications, Economics, Engineering, and Finance</source>, <volume>183</volume>. <publisher-loc>Basel</publisher-loc>: <publisher-name>Birkhauser</publisher-name>. </citation>
</ref>
<ref id="B20">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lai</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>W.-C.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Modeling Long-And Short-Term Temporal Patterns with Deep Neural Networks</article-title>,&#x201d; in <source>The 41st International ACM SIGIR Conference</source>. <publisher-loc>United States</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>95</fpage>&#x2013;<lpage>104</lpage>. <pub-id pub-id-type="doi">10.1145/3209978.3210006</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Laptev</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Amizadeh</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Flint</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Generic and Scalable Framework for Automated Time-Series Anomaly Detection</article-title>,&#x201d; in <source>The 21th ACM SIGKDD International Conference</source>. <publisher-loc>New York, United States</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>1939</fpage>&#x2013;<lpage>1947</lpage>. <pub-id pub-id-type="doi">10.1145/2783258.2788611</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Recognition and Vulnerability Analysis of Key Nodes in Power Grid Based on Complex Network Centrality</article-title>. <source>IEEE Trans. Circuits Syst. Express Briefs</source> <volume>65</volume>, <fpage>346</fpage>&#x2013;<lpage>350</lpage>. <pub-id pub-id-type="doi">10.1109/tcsii.2017.2705482</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Samson</surname>
<given-names>S. Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Oo</surname>
<given-names>A. M.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Impedance Modeling and Controllers Shaping Effect Analysis of Pmsg Wind Turbines</article-title>. <source>IEEE J. Emerging Selected Top. Power Electro.</source> <volume>9</volume>, <fpage>1465</fpage>&#x2013;<lpage>1478</lpage>. <pub-id pub-id-type="doi">10.1109/jestpe.2020.3014412</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Manzoor</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Lamba</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Akoglu</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Xstream: Outlier Detection in Feature-Evolving Data Streams</article-title>,&#x201d; in <source>The 24th ACM SIGKDD International Conference</source>. <publisher-loc>New York, United States</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>1963</fpage>&#x2013;<lpage>1972</lpage>. </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mehrotra</surname>
<given-names>K. G.</given-names>
</name>
<name>
<surname>Mohan</surname>
<given-names>C. K.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Anomaly Detection Principles and Algorithms</source>. <publisher-loc>USA</publisher-loc>: <publisher-name>Springer</publisher-name>. </citation>
</ref>
<ref id="B26">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nayak</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Pati</surname>
<given-names>U. C.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>S. K.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>A Comprehensive Review on Deep Learning-Based Methods for Video Anomaly Detection</article-title>,&#x201d; in <source>Image and Vision Computing</source>. <publisher-loc>Holland</publisher-loc>: <publisher-name>ELSEVIER</publisher-name>, <fpage>106</fpage>. <pub-id pub-id-type="doi">10.1016/j.imavis.2020.104078</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Hengel</surname>
<given-names>A. V. D.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Deep Learning for Anomaly Detection</article-title>. <source>ACM Comput. Surv.</source> <volume>54</volume>, <fpage>1</fpage>&#x2013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1145/3439950</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Park</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Noh</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ham</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Learning Memory-Guided Normality for Anomaly Detection</article-title>,&#x201d; in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>. <publisher-loc>United States</publisher-loc>: <publisher-name>IEEE</publisher-name>, <fpage>14372</fpage>&#x2013;<lpage>14381</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr42600.2020.01438</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roska</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Chua</surname>
<given-names>L. O.</given-names>
</name>
</person-group> (<year>1993</year>). <article-title>The Cnn Universal Machine: an Analogic Array Computer</article-title>. <source>IEEE Trans. Circuits Syst.</source> <volume>40</volume>, <fpage>163</fpage>&#x2013;<lpage>173</lpage>. <pub-id pub-id-type="doi">10.1109/82.222815</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Salinas</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Flunkert</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Gasthaus</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Deep</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Probabilistic Forecasting with Autoregressive Recurrent Networks (NIPS).</article-title> </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>H.-F.</given-names>
</name>
<name>
<surname>Dhillon</surname>
<given-names>I. S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Think Globally, Act Locally: A Deep Neural Network Approach to High-Dimensional Time Series Forecasting</article-title>. <source>Proc. Neural Inf. Process. Syst. Conf. (Nips)</source> <volume>32</volume>, <fpage>4837</fpage>&#x2013;<lpage>4846</lpage>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1905.03806">https://arxiv.org/abs/1905.03806</ext-link>. </citation>
</ref>
<ref id="B32">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Shipmon</surname>
<given-names>D. T.</given-names>
</name>
<name>
<surname>Gurevitch</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Piselli</surname>
<given-names>P. M.</given-names>
</name>
<name>
<surname>Edwards</surname>
<given-names>S. T.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Time Series Anomaly Detection; Detection of Anomalous Drops with Limited Features and Sparse Examples in Noisy Highly Periodic Data</source>. <comment>Preprint Repository Name [arXiv]. Available at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1708.03665">https://arxiv.org/abs/1708.03665</ext-link>
</comment>. </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Smith</surname>
<given-names>W. A.</given-names>
</name>
<name>
<surname>Randall</surname>
<given-names>R. B.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Rolling Element Bearing Diagnostics Using the Case Western reserve university Data: A Benchmark Study</article-title>. <source>Mech. Syst. signal Process.</source> <volume>64-65</volume>, <fpage>100</fpage>&#x2013;<lpage>131</lpage>. <pub-id pub-id-type="doi">10.1016/j.ymssp.2015.04.021</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Qu</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Anomaly Detection for Industrial Control System Based on Autoencoder Neural Network</article-title>,&#x201d; in <source>Wireless Communications and Mobile Computing</source>. <publisher-loc>United Kingdom</publisher-loc>: <publisher-name>Hindawi</publisher-name>. <pub-id pub-id-type="doi">10.1155/2020/8897926</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Edwards</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Philip</surname>
<given-names>S. Y.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Rs-forest: A Rapid Density Estimator for Streaming Anomaly Detection</article-title>,&#x201d; in <source>ICDM</source> (<publisher-loc>United States</publisher-loc>: <publisher-name>IEEE</publisher-name>) <fpage>600</fpage>&#x2013;<lpage>609</lpage>. <pub-id pub-id-type="doi">10.1109/icdm.2014.45</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Amplitude-phase Detection for Power Converters Tied to Unbalanced Grids with Large X/r Ratios</article-title>. <source>IEEE Trans. Power Electro.</source> <volume>37</volume>, <fpage>2100</fpage>&#x2013;<lpage>2112</lpage>. <pub-id pub-id-type="doi">10.1109/tpel.2021.3104591</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhuo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A Fast and Robust Real-Time Detection Algorithm of Decaying Dc Transient and Harmonic Components in Three-phase Systems</article-title>. <source>IEEE Trans. Power Electro.</source> <volume>35</volume>, <fpage>3332</fpage>&#x2013;<lpage>3336</lpage>. <pub-id pub-id-type="doi">10.1109/tpel.2019.2940891</pub-id> </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Xu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Bu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). &#x201c;<article-title>Unsupervised Anomaly Detection via Variational Auto-Encoder for Seasonal Kpis in Web Applications</article-title>,&#x201d; in <source>The 2018 World Wide Web Conference</source>. <publisher-loc>Switzerland</publisher-loc>: <publisher-name>ACM</publisher-name>, <fpage>187</fpage>&#x2013;<lpage>196</lpage>. <pub-id pub-id-type="doi">10.1145/3178876.3185996</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>M. N.</given-names>
</name>
<name>
<surname>San</surname>
<given-names>P. P.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X. L.</given-names>
</name>
<name>
<surname>Krishnaswamy</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Deep Convolutional Neural Networks on Multichannel Time Series for Human Activity Recognition</article-title>,&#x201d; in <source>Deep Convolutional Neural Networks on Multichannel Time Series for Human Activity Recognition</source> (<publisher-loc>Buenos Aires, Argentina</publisher-loc>: <publisher-name>IJCAI Press</publisher-name>). </citation>
</ref>
</ref-list>
</back>
</article>