<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2023.1135850</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Multi-channel EEG emotion recognition through residual graph attention neural network</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Chao</surname> <given-names>Hao</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Cao</surname> <given-names>Yiming</given-names></name>
<uri xlink:href="http://loop.frontiersin.org/people/2117289/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Yongli</given-names></name>
</contrib>
</contrib-group>
<aff><institution>College of Computer Science and Technology, Henan Polytechnic University</institution>, <addr-line>Jiaozuo</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Si Chen, Hong Kong Polytechnic University, Hong Kong SAR, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Yixin Zhang, Hong Kong Polytechnic University, Hong Kong SAR, China; Guijun Chen, Taiyuan University of Technology, China; Redha Taiar, Universit&#x000E9; de Reims Champagne-Ardenne, France</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Hao Chao <email>chaohao1981&#x00040;163.com</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>25</day>
<month>07</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>17</volume>
<elocation-id>1135850</elocation-id>
<history>
<date date-type="received">
<day>02</day>
<month>01</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>26</day>
<month>06</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Chao, Cao and Liu.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Chao, Cao and Liu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>In this paper, a novel EEG emotion recognition method based on residual graph attention neural network is proposed. The method constructs a three-dimensional sparse feature matrix according to the relative position of electrode channels, and inputs it into the residual network to extract high-level abstract features containing electrode spatial position information. At the same time, the adjacency matrix representing the connection relationship of electrode channels is constructed, and the time-domain features of multi-channel EEG are modeled using graph. Then, the graph attention neural network is utilized to learn the intrinsic connection relationship between EEG channels located in different brain regions from the adjacency matrix and the constructed graph structure data. Finally, the high-level abstract features extracted from the two networks are fused to judge the emotional state. The experiment is carried out on DEAP data set. The experimental results show that the spatial domain information of electrode channels and the intrinsic connection relationship between different channels contain salient information related to emotional state, and the proposed model can effectively fuse these information to improve the performance of multi-channel EEG emotion recognition.</p>
</abstract>
<kwd-group>
<kwd>EEG</kwd>
<kwd>emotion recognition</kwd>
<kwd>residual network</kwd>
<kwd>graph attention neural network</kwd>
<kwd>feature fusion</kwd>
</kwd-group>
<counts>
<fig-count count="14"/>
<table-count count="4"/>
<equation-count count="17"/>
<ref-count count="43"/>
<page-count count="15"/>
<word-count count="8250"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Auditory Cognitive Neuroscience</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Emotion is a physiological state of human beings accompanied by cognition and consciousness. People&#x00027;s daily cognitive and behavioral activities are almost driven by emotion, which also affects interpersonal interaction and group activities (Guozhen et al., <xref ref-type="bibr" rid="B13">2016</xref>). Affective computing is a representative field, which aims to give computer systems the ability to automatically recognize, understand and respond to human emotions, so as to realize intelligent human-computer interaction. As the core and important component of affective computing, emotion recognition has a wide range of applications in psychology, emotional computing, artificial intelligence, computer vision, medical, and other fields (Ramirez et al., <xref ref-type="bibr" rid="B31">2001</xref>; Hu et al., <xref ref-type="bibr" rid="B16">2019</xref>; F&#x000FC;rbass et al., <xref ref-type="bibr" rid="B9">2020</xref>).</p>
<p>Physiological signals mainly include electrocardiogram (ECG), electromyography (EMG), and electroencephalogram (EEG). Compared with facial expressions and voice signals, physiological signals are not easy to disguise, and are more objective and reliable in capturing the real emotional state of human beings. With the rapid development of wearable devices, long-term monitoring of physiological signals has become a reality, which makes it feasible and practical to judge emotional state based on EEG signals. In the medical field, EEG classification models play a role in automatic diagnosis of psychiatric disorders. Depression is one of the largest health problems in the world. It is a serious mental illness, and there is a problem of untimely treatment. Severe patients often have thoughts of suicide. Current diagnostic criteria for depression are still based on subjective clinical rating scales, such as The Hamilton Depression Rating Scale (Hamilton, <xref ref-type="bibr" rid="B14">1960</xref>), and require physician input (Sung et al., <xref ref-type="bibr" rid="B34">2005</xref>). Some research focuses on automatic diagnosis of depression based on EEG (Alhaj et al., <xref ref-type="bibr" rid="B2">2011</xref>; Mohammadi et al., <xref ref-type="bibr" rid="B27">2015</xref>), which can enable patients to quickly diagnose and understand their own condition, so as to carry out scientific treatment in advance.</p>
<p>The main components of the EEG signal are brain rhythm from different brain regions, reflecting the activity of that region (Niedermeyer and da Silva, <xref ref-type="bibr" rid="B28">2005</xref>). The electrical activity of the cerebral cortex is transmitted to the scalp through the anatomical structure. Therefore, the acquired EEG is a mixture of source signals from different brain regions, carrying a large amount of spatial location information (Xing et al., <xref ref-type="bibr" rid="B40">2019</xref>). In the research field of emotion recognition based on EEG, some studies have explored asymmetric features of brain regions, such as DASM (differential asymmetry), RASM (rational asymmetry), DCAU (differential causality) (Gogna et al., <xref ref-type="bibr" rid="B11">2016</xref>; Li et al., <xref ref-type="bibr" rid="B24">2018b</xref>). And other works studied the connectivity of EEG signals (Nolte et al., <xref ref-type="bibr" rid="B29">2004</xref>, <xref ref-type="bibr" rid="B30">2008</xref>; Supp et al., <xref ref-type="bibr" rid="B35">2007</xref>; Haufe et al., <xref ref-type="bibr" rid="B15">2013</xref>). Castelnovo et al. finds that the electrical activity of the brain is mainly concentrated in specific brain regions when people are in different sleep states, scalp EEG analysis of all night NREM (non-rapid eye movement) sleep revealed a localized decrease in slow wave activity (SWA) power (1&#x02013;4 Hz) over centro-parietal regions relative to the rest of the brain in SADs compared to good sleeping healthy controls (Castelnovo et al., <xref ref-type="bibr" rid="B3">2016</xref>). Nowadays, there are also some works that make better use of the spatial domain information of EEG channels in EEG classification tasks. In order to learn the spatiotemporal characteristics of EEG signals, Salama et al. divided the original EEG signals into multiple frames, and combined the original EEG signals of multiple channels into a two-dimensional matrix in each frame, where the first dimension represents the number of channels, and the second dimension Indicates the time length of a frame. Multiple frames are then superimposed to form a three-dimensional matrix, with the third dimension representing time. Finally, the 3D matrix is used as the input for 3D-CNN (3d convolutional neural networks) training. Since the left and right hemispheres of the human brain respond asymmetrically to emotion, a bi-hemisphere domain adversarial neural network (BiDANN) model is proposed to learn the discriminative emotional features of each hemisphere, BiDANN contains one global and two local domain discriminators, and learns discriminative sentiment features for each hemisphere by adversarial with local domain discriminators and classifiers (Li et al., <xref ref-type="bibr" rid="B26">2018c</xref>). Li et al. (<xref ref-type="bibr" rid="B25">2017</xref>) captures the spatial domain information contained in electrode positions by mapping into EEG multidimensional feature image following a 10/20 system. First, the spatial features, frequency domain and time features of the EEG signal are integrated, and mapped into a feature matrix according to the international 10/20 system, and then the EEG multidimensional feature image is generated using the interpolation method, using a combination of convolutional neural network (CNN) and long-term and short-term A hybrid deep network of memory (LSTM) recurrent neural network (RNN) recognizes emotional states. Li et al. (<xref ref-type="bibr" rid="B22">2018a</xref>) also used the distribution of electrodes on the scalp to extract the spatial domain information of electrode locations. First, the differential entropy features from 62 EEG signal channels are organized into a two-dimensional map of 8&#x000D7;9, and are mapped to a 20&#x000D7;20 input map through sparse operations to avoid information leakage in convolution and pooling operations. Finally, hierarchical convolutional neural network (HCNN) is used to classify positive, neutral and negative emotional states.</p>
<p>To a certain extent, the above research has applied the extraction of the spatial domain information of the EEG channel, and used the multi-dimensional feature matrix mapped according to the international 10/20 system and CNN to fuse the information of the neighbor nodes. However, there still exist several challenges in multi-channel EEG-based emotion recognition. First of all, the brain activity in emotional state is complex, and multiple brain regions are involved in the interaction. How to effectively characterize the interaction between brain regions is a problem to be considered. Furthermore, due to the local perception characteristics, CNN (Convolutional Neural Networks) tends to pay more attention to adjacent electrode channels and is good at learning local spatial patterns. Therefore, in the process of extracting electrode spatial position information, CNN can mine the significant information of correlation and interaction of different EEG signals in the same brain region. However, it cannot effectively capture the intrinsic connection relationship between EEG channels located in different brain regions and the global spatial position information of electrodes. Finally, the features extracted from the EEG signal and the distance between different electrodes are a kind of non-Euclidean data, only mapping the features extracted from each channel into a multi-dimensional sparse feature matrix according to the international 10/20 system ignores the distance information between electrodes, and ignores that all electrodes are not positioned in an absolute plane on the scalp.</p>
<p>To solve the above problems, this paper proposes a noval emotion recognition method based on residual graph attention neural network (ResGAT). In the proposed method, the residual network is utilized to achieve the spatial position information of the electrode channel and the correlation information of the adjacent EEG channels through the 3D feature matrix. Considering that the graph neural network (GAT) can update the state of vertices by periodically exchanging neighborhood information without being limited by vertex distance, it is employed to learn the neural functional connections between different brain regions, and the multi-head self-attention mechanism is used to adaptively adjust the adjacency matrix in the network. Therefore, the ResGAT model makes full use of the electrode spatial position information and the intrinsic connection relationship between EEG channels located in different brain regions. Moreover, when the EEG channel aggregates the characteristics of neighboring nodes, it pays more attention to the channel that is more relevant to itself. Finally, the high-level abstract features representing electrode space domain information and the high-level abstract features representing intrinsic connection relationship between EEG channels located in different brain regions are fused to judge the emotional state.</p>
</sec>
<sec id="s2">
<title>2. Datasets and feature extraction</title>
<sec>
<title>2.1. Data set</title>
<p>The DEAP data set used in the experiment is an open data set collected through experiments by Koelstra et al. from Queen Mary University of London, University of Twente, University of Geneva, Switzerland, and Swiss federal Institute of Technology in Lausanne to analyze human emotional states (Koelstra et al., <xref ref-type="bibr" rid="B20">2011</xref>). The dataset records multimodal physiological signals of 32 volunteers under the stimulation of selected music videos, including EEG and peripheral physiological signals, and 22 of the 32 volunteers also record facial expression videos. Each volunteer needs to watch 40 1-min long videos using 32 active AgCl electrodes (Fp1, AF3, F3, F7, FC5, FC1, C3, T7, CP5, CP1, P3, P7, PO3, O1, Oz, Pz, Fp2, AF4, Fz, F4, F8, FC6, FC2, Cz, C4, T8, CP6, CP2, P4, P8, PO4, and O2) recording EEG signals, these electrodes were placed on the scalp according to the international 10/20 system. At the end of each trial, the valence, arousal, dominance, and liking of the video were evaluated on a scale of 0&#x02013;9. Physiological signals were sampled at 512 Hz and resampled at 128 Hz. The physiological signal matrix of each subject is 40&#x000D7;40&#x000D7;8064 (40 trials, 40 channels, 8,064 sampling points). Eighty thousand and sixty-four is 63 s data at 128 Hz sampling rate, 3 s silent time.</p>
<p>In addition, the effectiveness of ResGAT was verified using the SEED-IV brain emotion dataset (Zheng et al., <xref ref-type="bibr" rid="B42">2018</xref>). This data set selected 72 movie clips containing 4 emotions (Happy, Sad, Neutral, and Fear) as EEG-induced materials. A total of 15 subjects recorded 62-channel EEG signals and eye movements when watching movie clips.</p>
</sec>
<sec>
<title>2.2. Feature extraction</title>
<p>In the DEAP set, each person watched 40 emotion-inducing videos, and the duration of EEG signals recorded in each video was 60 s. In the experiment, a sliding window divides the raw EEG signal of each channel into several segments, and the duration of each sliding window is set to 6 s. The segments do not over lap. Each segment is considered an independent sample, and the six new samples inherited the labels of the original. Thus, 12,800 samples were be obtained. A set of time domain features can be extracted from the 32-channel EEG signals of each sample, specifically including mean, variance, first difference value, second difference value, standard deviation, and fuzzy entropy. Among them, Fuzzy Entropy was proposed by Chen et al. (<xref ref-type="bibr" rid="B5">2007</xref>) and applied to the representation of EMG signals. Fuzzy Entropy introduces the concept of fuzzy sets. Based on the exponential function and its shape, the similarity of vectors is vaguely defined in FuzzyEn, compared with ApEn and SampEn, the FuzzyEn is an effective measure algorithm for analyzing chaotic sequence complexity, it has better robustness and measure value continuities.</p>
<p>The soft continuous boundary of the fuzzy function ensures the continuity and effectiveness of the fuzzy entropy under small parameters, so the more details obtained by the fuzzy function also make the fuzzy entropy a more accurate definition of entropy. Assuming that the EEG signal of each channel is represented by s(T), t = 1, 2..., T, T is the signal length, which is 128&#x000D7;60 (frequency&#x000D7;second), the measure of the length of the EEG signal subsequence in fuzzy entropy m is 2. By reconstructing the original sequence, we can get</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>m</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>}</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Among them, i=1,2,...,N-m&#x0002B;1. <inline-formula><mml:math id="M2"><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> represents m consecutive s values, <italic>s</italic><sub>0</sub>(<italic>i</italic>) represents the average value, calculated as follows,</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mstyle displaystyle="true"><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Define the maximum difference <inline-formula><mml:math id="M4"><mml:msubsup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> between elements in two m-dimensional vectors <inline-formula><mml:math id="M5"><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M6"><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> as the distance between them,</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x02208;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mi>m</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo>&#x02223;</mml:mo><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02223;</mml:mo></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The similarity between <inline-formula><mml:math id="M8"><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M9"><mml:msubsup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> can be defined by a fuzzy function,</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>&#x003BC;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Structure &#x003C6;<sub><italic>m</italic></sub>(<italic>r</italic>) and &#x003C6;<sub><italic>m</italic>&#x0002B;1</sub>(<italic>r</italic>),</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mi>m</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msubsup><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mi>m</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msubsup><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Then can define the parameter <italic>FuzzyEn</italic>(<italic>m, r</italic>) of the time series as:</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>F</mml:mi><mml:mi>u</mml:mi><mml:mi>z</mml:mi><mml:mi>z</mml:mi><mml:mi>y</mml:mi><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>,</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo class="qopname">lim</mml:mo></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x02192;</mml:mo><mml:mo>-</mml:mo><mml:mi>&#x0221E;</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">ln</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Among them, when N is finite, it can be estimated by statistics,</p>
<disp-formula id="E8"><label>(8)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>F</mml:mi><mml:mi>u</mml:mi><mml:mi>z</mml:mi><mml:mi>z</mml:mi><mml:mi>y</mml:mi><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>,</mml:mo><mml:mi>r</mml:mi><mml:mo>,</mml:mo><mml:mi>N</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mo class="qopname">ln</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003C6;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Two emotion dimensions are used in the experiments. For each sample, if the self-assessment value of the arousal is greater than 5, the category label of the sample is set as the high arousal (HA), otherwise it is set as the low arousal (LA). In the valence-sentiment dimension, the same label division is used for samples, including the high valence (HV) and the low valence (LV).</p>
</sec>
</sec>
<sec id="s3">
<title>3. ResGAT emotion recognition framework</title>
<p>The structure of the proposed ResGAT model is described in <xref ref-type="fig" rid="F1">Figure 1</xref>. The framework includes feature extraction and feature mapping module, ResNet modules, GAT modules and Classification modules. The first part is feature extraction and feature mapping modules, which extracts 6 kinds of temporal features from 32 EEG signals. Then, a 2D electrode position mapping matrix and a 3D sparse feature matrixand are constructed according to the temporal features. The 2D electrode position mapping matrix is input into the GAT modules to extract high-level abstract features, which contain the intrinsic connection relationships between EEG channels located in different brain regions. The ResNet modules is employed to receive the 3D sparse feature matrix and generate high-level abstract features representing electrode spatial position information. Finally, the classification modules is utilized to fuse the two high-level abstract features and judge the emotional state.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>ResGAT emotion recognition framework.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0001.tif"/>
</fig>
<sec>
<title>3.1. Extraction of spatial domain information based on ResNet Module</title>
<p><xref ref-type="fig" rid="F2">Figure 2</xref> shows the international 10/20 system plan, 2D electrode position mapping matrix and 3D sparse feature matrix <italic>X</italic><sup><italic>R</italic></sup> &#x02208; &#x0211D;<sup><italic>h</italic>&#x000D7;<italic>w</italic>&#x000D7;<italic>c</italic></sup>. The values of parameter h and parameter w are both set to 9, and the value of parameter c is 6, indicating that the shape of the 3D feature matrix is. The left side of <xref ref-type="fig" rid="F2">Figure 2</xref> shows the International 10/20 system, where the EEG electrodes marked by green circles are the test points used in the DEAP dataset. Some researches (Li et al., <xref ref-type="bibr" rid="B25">2017</xref>; Chao and Dong, <xref ref-type="bibr" rid="B4">2020</xref>; Cui et al., <xref ref-type="bibr" rid="B6">2020</xref>) have found that spatial features of EEG channels can improve the performance of emotion recognition. In order to represent the spatial location information of all EEG signal channels, a feature matrix is constructed according to the positions of electrodes on the brain, and the spatial parts of different EEG signal channels are mined. In the feature matrix, the time-domain features extracted from different EEG channels are put into the corresponding positions in the matrix by name, and the positions of unused electrodes in the matrix are set to 0. Finally, a 9&#x000D7;9&#x000D7;6 three-dimensional feature matrix is constructed according to the six extracted features, as shown in the right of <xref ref-type="fig" rid="F2">Figure 2</xref>. For each 9&#x000D7;9&#x000D7;1 matrix, different time domains are arranged according to the mapping rules shown in <xref ref-type="fig" rid="F2">Figure 2</xref> mapping matrix. Finally, the extracted 3D sparse feature matrix is represented by <italic>X</italic><sup><italic>R</italic></sup>, <italic>h</italic> &#x0003D; <italic>w</italic> &#x0003D; 9, <italic>c</italic> &#x0003D; 6, indicating that the shape of the 3D feature matrix is 9&#x000D7;9&#x000D7;6.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>International 10/20 system, 9&#x000D7;9 mapping matrix and 3D sparse feature matrix.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0002.tif"/>
</fig>
<p>After constructing the three-dimensional sparse mapping matrix, it is input into the residual network to extract high-level abstract features. The residual network structure adopted is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. It is composed of multiple residual blocks. Each residual block is composed of multiple convolution layers, batch normalization layers and activation layers. The size of the convolution kernels used in this part is 3&#x000D7;3. The residual block is calculated as follows:</p>
<p>First, the 3D sparse feature map <italic>U</italic> &#x02208; &#x0211D;<sup><italic>h</italic>&#x02032;&#x000D7;<italic>w</italic>&#x02032;&#x000D7;<italic>c</italic>&#x02032;</sup> is obtained from <italic>X</italic><sup><italic>R</italic></sup> &#x02208; &#x0211D;<sup><italic>h</italic>&#x000D7;<italic>w</italic>&#x000D7;<italic>c</italic></sup> by transforming <italic>F</italic><sub><italic>tr</italic></sub>. For transform <italic>F</italic><sub><italic>tr</italic></sub>, it is a Convolution operation. Use <inline-formula><mml:math id="M15"><mml:mi>V</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> to represent the filter set, where <italic>v</italic><sub><italic>i</italic></sub> refers to the parameter of the <italic>i</italic><sub><italic>th</italic></sub> filter. The output is <inline-formula><mml:math id="M16"><mml:mi>U</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>, and</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002A;</mml:mo><mml:msup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Here, &#x0002A; means convolution, the filter can learn the spatial position information of electrodes and the interaction information between electrodes in local spatial position through convolution operation.</p>
<p>The normalized network response after batching is <inline-formula><mml:math id="M18"><mml:mi>Z</mml:mi><mml:mo>=</mml:mo><mml:mi>B</mml:mi><mml:mi>N</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
<p>Batch normalization can effectively prevent the gradient explosion and gradient disappearance in the network, and speed up the convergence speed of the network. Finally, the nonlinear interaction between the feature map channels is learned through the activation layer, and the complete dependence between the channels is obtained. It is expressed by the following formula:</p>
<disp-formula id="E10"><label>(10)</label><mml:math id="M19"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mi>W</mml:mi><mml:mi>Z</mml:mi><mml:mo>,</mml:mo><mml:mi>W</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Among them, &#x003B4; refers to the relu activation function. After multiple convolutions and activation calculations, the final EEG signal characteristics are expressed as <inline-formula><mml:math id="M20"><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec>
<title>3.2. Dynamic learning of the intrinsic connection relationship between EEG channels located in different brain regions</title>
<p>As the basis of ResGAT method, some basic knowledge about graph representation is introduced first. A directed connected graph can be defined as <italic>G</italic> &#x0003D; <italic>V, E, W</italic>, where <italic>V</italic> represents the node set with the number of |<italic>V</italic>| &#x0003D; <italic>N</italic>, and <italic>E</italic> represents the edge set connecting these nodes. Let <italic>W</italic> &#x02208; &#x0211D;<sup><italic>N</italic>&#x000D7;<italic>N</italic></sup> represents the adjacency matrix describing the connection between any two nodes in <italic>V</italic>, in which the entry of <italic>W</italic> in row <italic>i</italic> and column <italic>j</italic> measures the importance of the connection between node <italic>i</italic> and <italic>j</italic>. <xref ref-type="fig" rid="F3">Figure 3</xref> shows five nodes and edges connecting those nodes, as well as the adjacency matrix associated with the graph. The different colored arrows on the left side of the figure represent the edges connecting the source node and the target node, while the corresponding adjacency matrix is on the right side of the figure.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>A directed graph and the corresponding adjacency matrix.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0003.tif"/>
</fig>
<p>In the past, convolutional neural networks have been applied in many fields due to their powerful modeling capabilities, such as computer vision, speech recognition, and natural language processing. Due to its locality and translation invariance properties, it is very suitable for processing Euclidean data.However, many elements in the real world exist in the form of graph data, such as social networks, transportation networks, and drug discovery.The features extracted from the EEG signal and the distance between different electrodes are non-European data. Although the number of features on each signal channel is consistent, the distance between each adjacent electrode is uneven, and brain functional connectivity tends to capture global relationships among EEG channels. Therefore, the graph neural network is more suitable for learning the potential internal connections between different channels.At present, the graph attention network (GAT) (Veli&#x0010D;kovi&#x00107; et al., <xref ref-type="bibr" rid="B39">2017</xref>) is a widely used graph neural network. GAT achieves information aggregation in the spatial domain by introducing an attention mechanism, making the model pay more attention to the mutual influence between neighbor nodes, and applying it to EEG data to make the channels aggregate the characteristics of neighbor nodes and pay more attention to channels that are more relevant to themselves. Each EEG electrode can be regarded as a node of the graph, and the connection between the electrodes corresponds to the edge of the graph. The weights of all edges, which representing the functional relationship between electrodes, constitute the adjacency matrix of the graph. Therefore, GAT can learn the internal relationship between different EEG electrodes. As shown in the attention neural network in <xref ref-type="fig" rid="F1">Figure 1</xref>, although GAT can describe the connection between different nodes according to their spatial positions, the connection between EEG channels should be determined in advance before applying it to the construction of emotion recognition model. In addition, it should be noted that the spatial location connection between EEG channels is different from the functional connection between them. In other words, closer spatial relationships may not guarantee closer functional relationships.</p>
<p>The flow of processing EEG signal features with GAT is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. After data acquisition, preprocessing and feature extraction, EEG data are represented by undirected graph <italic>G</italic> &#x0003D; <italic>V, E, W</italic>. The data on can be represented as feature matrix <italic>X</italic><sup><italic>G</italic></sup> &#x02208; &#x0211D;<sup><italic>n</italic>&#x000D7;<italic>d</italic></sup>, where <italic>n</italic> represents the number of electrodes and <italic>d</italic> represents the number of features extracted on each electrode channel. The constructed initial adjacency matrix <italic>W</italic><sup><italic>G</italic></sup> &#x02208; &#x0211D;<sup><italic>n</italic>&#x000D7;<italic>n</italic></sup>, where <italic>n</italic> represents the number of electrode channels, characterizes the correlation between 2D space electrodes. Assume that each electrode channel has an internal relationship with the other 31 electrode channels, and is initialized as a diagonal matrix with the main diagonal of 0 and other values of 1. The feature combination extracted from each EEG channel is represented as a node in the graph neural network model, can be expressed as:</p>
<disp-formula id="E11"><label>(11)</label><mml:math id="M21"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>,</mml:mo><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>,</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>,</mml:mo><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>,</mml:mo><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In order to obtain sufficient expression ability to transform input features into higher-level features, at least one learnable linear transformation is needed. A shared <italic>H</italic> &#x02208; &#x0211D;<sup><italic>d</italic>&#x02032;&#x000D7;<italic>d</italic></sup> applies to all nodes to increase the expression ability of node features.</p>
<p>Then, a self attention mechanism is used on all nodes. At this time, the dimension of features on the nodes remains unchanged, which is &#x0211D;<sup><italic>d</italic>&#x02032;</sup>. The self attention mechanism is described as:</p>
<disp-formula id="E12"><label>(12)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>A</mml:mi><mml:mi>t</mml:mi><mml:mi>t</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>,</mml:mo><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>Att</italic> stands for self attention mechanism, and <italic>e</italic><sub><italic>ij</italic></sub> represents the importance of the characteristics of node <italic>j</italic> to <italic>i</italic>. Only the first-order neighbors of each node is calculated. In order to make the coefficients easy to compare between different nodes, the softmax function is used to normalize the attention coefficients of node <italic>j</italic> to other neighbor nodes.</p>
<disp-formula id="E13"><label>(13)</label><mml:math id="M23"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mi>o</mml:mi><mml:mi>f</mml:mi><mml:mi>t</mml:mi><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:msub><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mstyle><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>a</italic><sub><italic>ij</italic></sub> is the coefficient of attention mechanism. In fact, the attention mechanism a is composed of a single-layer feedforward neural network, and the leakyrelu activation function is used for non-linear processing. Finally, the coefficient of attention mechanism can be expressed as:</p>
<disp-formula id="E15"><label>(14)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:msub><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mstyle><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x02003;</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>L</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>y</mml:mi><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>&#x02225;</mml:mo><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msubsup></mml:mstyle><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>L</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>k</mml:mi><mml:mi>y</mml:mi><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>U</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>&#x02225;</mml:mo><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo><mml:mover><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mi>d</mml:mi></mml:mrow></mml:msup><mml:mtext>&#x02003;&#x000A0;</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x02225; indicates connection operation.</p>
<p>Then apply the normalized attention coefficient to the features corresponding to the node, and get the output after feature recalibration:</p>
<disp-formula id="E16"><label>(15)</label><mml:math id="M26"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mover><mml:mrow><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Veli&#x0010D;kovi&#x00107; et al. (<xref ref-type="bibr" rid="B39">2017</xref>) found that it is beneficial to use multi head attention mechanism in graph neural network. Using <italic>K</italic> independent attention mechanisms at the same time, Formula 14 will produce <italic>K</italic> outputs. Then splice the above <italic>K</italic> outputs together, as shown in the following formula:</p>
<disp-formula id="E17"><label>(16)</label><mml:math id="M27"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mover><mml:mrow><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x02225;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msubsup><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:msubsup><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mi>H</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msup><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The output of each node changes to <italic>Kd</italic>&#x02032;. In the experiment, the <italic>K</italic> is 2.</p>
<p>The aggregation process of multi head attention mechanism on nodes is shown in <xref ref-type="fig" rid="F4">Figure 4</xref>.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>(A)</bold> The attention mechanism <inline-formula><mml:math id="M28"><mml:mi>a</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover><mml:mo>,</mml:mo><mml:mi>H</mml:mi><mml:mover><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is parameterized by the weight vector <italic>a</italic> &#x02208; &#x0211D;<sup>2<italic>d</italic>&#x02032;</sup>. <bold>(B)</bold> Illustration of multi headed attention (k = 2) of node 1 in its neighborhood. Arrows of different colors indicate independent attention calculation. Aggregate features from each head are connected to obtain <inline-formula><mml:math id="M29"><mml:mover><mml:mrow><mml:msubsup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mrow></mml:mover></mml:math></inline-formula>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0004.tif"/>
</fig>
<p>The above is a complete graph convolution process. After multi-layer graph convolution, the EEG features will be further transmitted to the full connection layer, fused and classified with the extracted high-level abstract spatial features, and the <italic>S</italic><sup><italic>G</italic></sup> is obtained by batch normalization before full connection.</p>
</sec>
<sec>
<title>3.3. Feature fusion</title>
<p>The deep features extracted from convolution network and graph network are flattened and spliced, as shown below:</p>
<disp-formula id="E18"><label>(17)</label><mml:math id="M30"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>O</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mi>p</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>f</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mi>f</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Finally, the softmax function is used to output the emotional state. The loss function of this model is the cross-entropy function, and the loss function is minimized using the Adam optimizer with an initial learning rate of 0.0001.</p>
</sec>
</sec>
<sec id="s4">
<title>4. Experimental results and analysis</title>
<sec>
<title>4.1. Performance analysis</title>
<p>The emotion classification network in the experiment consists of residual network and graph attention neural network. The residual network consists of multiple blocks, and each block contains two convolutional layers. In order to increase the fitting ability of the network, an activation layer is added after all convolutional layers. The first two residual blocks employ 64 filters with a size of 3&#x000D7;3 for convolution calculations, and the last two residual blocks use 128 filters of the same size.</p>
<p>The effect of the proposed network is verified on the DEAP dataset. In order to make the experimental results more objective, 10-fold cross-validation technique is used.</p>
<p><xref ref-type="fig" rid="F5">Figures 5</xref>, <xref ref-type="fig" rid="F6">6</xref> show the training process of the proposed network on the dataset, <xref ref-type="fig" rid="F5">Figure 5</xref> shows the training process of the two emotional dimensions of arousal and valence on the DEAP data set, and <xref ref-type="fig" rid="F6">Figure 6</xref> shows the training process of the SEED data set process. Among them, when the training period is less than 750 in the DEAP dataset, the training accuracy and validation accuracy increase with the increase of the epoch. When the epoch is greater than 750, the training accuracy and validation longitude tend to be stable.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>The training process of the proposed network in the two dimensions of Arousal and Valence.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>The training process of the proposed network in the SEED-IV.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0006.tif"/>
</fig>
<p>The classification accuracy (Acc) and F1 score (F1) are used to evaluate the performance of the proposed model. The emotion recognition results are shown in <xref ref-type="fig" rid="F7">Figure 7</xref>, respectively. In the arousal dimension, the accuracy is 0.8706 and the F1 score is 0.8833. In the valence dimension, the recognition accuracy and F1 score are 0.8926 and 0.9042, respectively. In addition, 0.9773 Acc and 1.0 F1 were achieved on the four-category task of the SEED-IV dataset. The results of three classified tasks demonstrate the effectiveness of the proposed method.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Emotion recognition results of the proposed network for binary classification.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0007.tif"/>
</fig>
<p>The receiver operating curve (ROC) is also used to evaluate the performance of the proposed network. The ROC curve is located at the upper left triangle of the square, which reflects a more satisfactory classification rule. The higher the area under ROC Curve (AUC) value, the better the classification effect. <xref ref-type="fig" rid="F8">Figure 8</xref> shows the receiver operating curves on the two classifications of arousal and valence. The values of AUC in the two dimensions are 0.9378 and 0.9565, respectively. The relatively convex curve and high AUC value prove the excellent classification performance of the proposed classification network.</p>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>ROC curve of the proposed network.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0008.tif"/>
</fig>
</sec>
<sec>
<title>4.2. Comparison between the ensemble method and the single network</title>
<p>In the experiment, an independent GAT model and an independent ResNet model are constructed, respectively. The network structures of the independent GAT and the independent ResNet used in the experiment are consistent with those in the proposed ensemble ResGAT network. When these two independent models are used for emotion classification, the high-level abstract features are flattened and fed into a fully connected layer for classification. The 10-fold cross-validation technique are also used here, and other hyperparameters remain the same.</p>
<p>Firstly, the comparison is carried out on the emotion recognition accuracy. Compared with the GAT model, the proposed ResGAT improves the emotion recognition accuracies by 21.85% in the arousal dimension and 24.68% in the valence dimension. Compared with the ResNet model, the proposed ResGAT improves the emotion recognition accuracies by 1.64% in the arousal dimension and 2.99% in the valence dimension. Secondly, the comparison is carried out on the F1 scores. Compared with the GAT model, the proposed ResGAT improves the emotion recognition accuracies by 17.5% in the arousal dimension and 21.54% in the valence dimension. Compared with the ResNet model, the proposed ResGAT improves the emotion recognition accuracies by 0.9% in the arousal dimension and 2.26% in the valence dimension. The results show that the performance of the proposed ResGAT is obviously better than that of GAT, and it is also improved compared with ResNet.</p>
<p>In addition, it was also verified on the SEED-IV dataset, and the model recognition results are shown in <bold>Table 2</bold>. The experimental results in <xref ref-type="table" rid="T1">Tables 1</xref>, <xref ref-type="table" rid="T2">2</xref> show that the performance of the integrated network is better than that of a single network, because each network in the integrated network can extract different information, that is, the spatial position information of electrodes and the internal relationship of EEG channels in different brain regions. These two kinds of information complement each other to improve the model recognition performance.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>The results of ResGAT and the two single networks.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<th valign="top" align="center" colspan="7" rowspan="1"><bold>Recognition results</bold></th>
</tr>
</thead>
<tbody>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<td valign="top" align="left"><bold>Emotion dimension</bold></td>
<td valign="top" align="center" colspan="2"><bold>ResGAT</bold></td>
<td valign="top" align="center" colspan="2"><bold>GAT</bold></td>
<td valign="top" align="center" colspan="2"><bold>ResNet</bold></td>
</tr>
 <tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<td/>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
</tr> <tr>
<td valign="top" align="left">Arousal</td>
<td valign="top" align="center">0.8706</td>
<td valign="top" align="center">0.8833</td>
<td valign="top" align="center">0.6521</td>
<td valign="top" align="center">0.7083</td>
<td valign="top" align="center">0.8542</td>
<td valign="top" align="center">0.8743</td>
</tr> <tr>
<td valign="top" align="left">Valence</td>
<td valign="top" align="center">0.8926</td>
<td valign="top" align="center">0.9042</td>
<td valign="top" align="center">0.6458</td>
<td valign="top" align="center">0.6888</td>
<td valign="top" align="center">0.8627</td>
<td valign="top" align="center">0.8816</td>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>The results of ResGAT and the two single networks (SEED-IV).</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<th valign="top" align="center" colspan="7"><bold>Recognition results</bold></th>
</tr>
</thead>
<tbody>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<td valign="top" align="left"><bold>Emotion dimension</bold></td>
<td valign="top" align="center" colspan="2"><bold>ResGAT</bold></td>
<td valign="top" align="center" colspan="2"><bold>GAT</bold></td>
<td valign="top" align="center" colspan="2"><bold>ResNet</bold></td>
</tr>
 <tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<td/>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
<td valign="top" align="center"><bold>Accuracy</bold></td>
<td valign="top" align="center"><bold>F1</bold></td>
</tr> <tr>
<td valign="top" align="left">Four classification</td>
<td valign="top" align="center">0.9773</td>
<td valign="top" align="center">0.9973</td>
<td valign="top" align="center">0.9522</td>
<td valign="top" align="center">0.9635</td>
<td valign="top" align="center">0.9021</td>
<td valign="top" align="center">0.9251</td>
</tr></tbody>
</table>
</table-wrap>
<p>To further verify the complementarity between the electrode spatial position information and the intrinsic connection relationship between EEG channels, a variety of CNN networks, graph convolution network (GCN) (Kipf and Welling, <xref ref-type="bibr" rid="B19">2016</xref>), vision in transformer network (VIT) (Dosovitskiy et al., <xref ref-type="bibr" rid="B7">2020</xref>) and the ensemble models integrated by the above networks are constructed. The CNN networks constructed specifically include Alex (Krizhevsky et al., <xref ref-type="bibr" rid="B21">2017</xref>), VGGNet (Simonyan and Zisserman, <xref ref-type="bibr" rid="B33">2014</xref>), DenseNet (Huang et al., <xref ref-type="bibr" rid="B17">2017</xref>), and GoogLeNet (Szegedy et al., <xref ref-type="bibr" rid="B36">2015</xref>), which focus on extracting the electrode spatial position information. Similar to GAT, GCN and VIT focus on extracting the intrinsic connection relationship between EEG channels. An ensemble model is constructed by a CNN network and GCN, or by a CNN network and VIT, which means these ensemble models can capture both the electrode spatial position information and the intrinsic connection relationship between EEG channels.</p>
<p>In the CNN networks, the AlexNet structure is affected by the size of the input data. Compared with the structure in the reference (Krizhevsky et al., <xref ref-type="bibr" rid="B21">2017</xref>), the maximum pooling is removed, and the size of the convolution kernel is modified. Other structures remain unchanged. Compared with the residual network in this paper, the structure of VGG only removes the spanning connection. DenseNet employs 1&#x000D7;1 convolutions for better data representation, where the depth of the convolutional layers is 9. GoogLeNet contains a multi-branch convolution structure, which uses convolution kernels of 3&#x000D7;3, 5&#x000D7;5, and 7&#x000D7;7, respectively.</p>
<p>GCN is a natural extension of convolutions on graph structure. Because GCN is suitable for extracting structural features of graphs and can customize local receptive fields, it is widely used in network analysis, traffic prediction. and recommender systems. Inspired by the reference (Kipf and Welling, <xref ref-type="bibr" rid="B19">2016</xref>), a spectral domain-based graph convolutional network is constructed, which contains two convolutional layers. Transformer has been successfully applied in natural language processing and computer vision. Therefore, in the experiment, the standard transformer is directly applied to the EEG signal features with minimal modification. Similar to the reference (Dosovitskiy et al., <xref ref-type="bibr" rid="B7">2020</xref>), the 3D feature matrix is divided into small blocks, and the linear embedding sequence of these blocks is provided as the input of the transformer.</p>
<p>The recognition results of the above networks and the ensemble models are shown in <xref ref-type="fig" rid="F9">Figure 9</xref>. Most of the ensemble models have higher classification accuracy than the corresponding single network, which proves that the electrode spatial position information and the intrinsic connection relationship between EEG channels are complementary to emotion classification. ResNet has the highest classification accuracy in a single network, which achieves 85.42% classification accuracy in arousal dimension and 86.27% classification accuracy invalence dimension, respectively. In the integrated models, ResGAT achieves the highest classification accuracy in the valence dimension, and ResNet-VIT achieves the highest classification accuracy in the arousal dimension. Among the above ensemble models, the ensemble models including GAT performs well in emotion recognition tasks. Combining GAT with any kind of CNN, the classification accuracy can be improved. The experimental results show that GAT has better information capture ability than GCN and VIT, and is more suitable for combining with convolutional networks, which makes the extracted high-level abstract features contain relatively less redundant and irrelevant components.</p>
<fig id="F9" position="float">
<label>Figure 9</label>
<caption><p>Recognition results of the single networks and the ensemble models integrated by the single networks.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0009.tif"/>
</fig>
</sec>
<sec>
<title>4.3. ResGAT with different model structures</title>
<p>In addition to the ResGAT (ResGAT1) proposed in this paper, three other ResGAT models (ResGAT2, ResGAT3, and ResGAT4) are also constructed. In the ResGAT1, all 3&#x000D7;3 filters are used in the residual network, and the number of multi-head attention in all graph attention layers in GAT is 2. ResGAT2 sets the multi-head attention number of GAT to 4. ResGAT3 is twice as deep as ResGAT1. ResGAT4 uses 3&#x000D7;3 and 5&#x000D7;5 filters to cross the residual structure on the basis of increasing the network depth. The other parameters in ResGAT for the three comparisons remain unchanged. All samples of subjects and 10-fold cross-validation technique are also used here. The recognition results of ResGAT with different structures under the two sentiment annotation schemes are shown in <xref ref-type="fig" rid="F10">Figure 10</xref>.</p>
<fig id="F10" position="float">
<label>Figure 10</label>
<caption><p>Recognition results of ResGAT1, ResGAT2, ResGAT3, and ResGAT4.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0010.tif"/>
</fig>
<p>Compared with ResGAT2, ResGAT3, and ReGAT4, the recognition accuracy of the proposed ResGAT in the arousal dimension is improved by 1.02, 1.85, and 0.24%, respectively. In the dimension of valence, the recognition accuracy of the proposed ResGAT is improved by 0.59, 3.36, and 1.61%, respectively. It can be seen from the comparison results that increasing the complexity and depth of the network will not necessarily improve the accuracy, but will increase the calculation of the model. Therefore, it is very important to choose the appropriate network structure.</p>
</sec>
<sec>
<title>4.4. Sensitivity analysis</title>
<p>To further prove that the proposed network can extract the spatial domain information of EEG signal channels and learn the internal relationship of different EEG signal channels, the information extraction ability of the proposed model is analyzed.</p>
<p>In the proposed method, the three-dimensional sparse feature matrix and deep residual network are used to capture the dependence between local EEG signal channels. As a contrast, the deep residual network model is also used to deal with the same time-domain characteristics without mapping and arranging according to the international 10/20 standard. Six features of 32 channels can construct a two-dimensional feature matrix with a size of 32&#x000D7;6. The hyperparameters in the experiment remain unchanged. The recognition results using 3D feature matrix and 2D feature matrix, respectively, are shown in <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Recognition results using residual networks.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<th/>
<th valign="top" align="center"><bold>3D</bold></th>
<th valign="top" align="center"><bold>2D</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">DEAP-Arousal</td>
<td valign="top" align="center">0.8542</td>
<td valign="top" align="center">0.6203</td>
</tr> <tr>
<td valign="top" align="left">DEAP-Valence</td>
<td valign="top" align="center">0.8627</td>
<td valign="top" align="center">0.6136</td>
</tr> <tr>
<td valign="top" align="left">SEED-IV</td>
<td valign="top" align="center">0.9021</td>
<td valign="top" align="center">0.9020</td>
</tr></tbody>
</table>
</table-wrap>
<p>Compared with the two-dimensional feature matrix, the accuracy of emotion recognition of the three-dimensional feature matrix in the arousal dimension is increased by 23.39%, the accuracy of emotion recognition in the valence dimension is increased by 24.91%, 0.01% improvement on the SEED-IV dataset. The results show that three-dimensional feature matrix and deep residual network can effectively extract local dependency information of signal channels.</p>
<p>In order to illustrate the intrinsic connection relationship between EEG channels mined by GAT, the adjacency matrix learned during the training process is displayed. The adjacency matrix is affected by the input data. Input all training data into the GAT in turn to obtain the adjacency matrix corresponding to each sample. The average value of all adjacency matrices can construct a heat map, as shown in <xref ref-type="fig" rid="F11">Figures 11</xref>, <xref ref-type="fig" rid="F12">12</xref>.</p>
<fig id="F11" position="float">
<label>Figure 11</label>
<caption><p>Heatmap representation of adjacency matrices in GAT on DEAP-Arousal and DEAP-Valence affective dimensions.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0011.tif"/>
</fig>
<fig id="F12" position="float">
<label>Figure 12</label>
<caption><p>Heatmap representation of adjacency matrices in GAT on SEED-IV.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0012.tif"/>
</fig>
<p>It can be clearly seen that the graph neural network is not limited by distance when collecting neighbor node information in 32 electrode channels. In terms of arousal and valence emotion, C4 electrode channel pays more attention to FC5 channel when aggregating neighbor node information, and FC2 channel pays more attention to F7 and FC5 channels, and CP5 pays more attention to FP2 channel. In the SEED-IV dataset, all nodes focus more on the four channels CZ, CPZ, PZ, and POZ.</p>
</sec>
<sec>
<title>4.5. t-SNE analysis</title>
<p>In order to demonstrate the effectiveness of ResGAT in extracting high-level abstract features, the t-SNE tool is used to visually analyze the features in two-dimensional space, these features extract all data from a single person. As shown in <xref ref-type="fig" rid="F13">Figures 13</xref>, <xref ref-type="fig" rid="F14">14</xref>, the input data of the model and the high-level abstract features extracted by ResGAT are displayed in the two emotional dimensions of arousal and valence. The results in the figure demonstrate the effectiveness of the proposed ResGAT in extracting affective state discriminative features.</p>
<fig id="F13" position="float">
<label>Figure 13</label>
<caption><p>t-SNE analysis on the emotional dimensions of DEAP-Arousal and DEAP-Valence.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0013.tif"/>
</fig>
<fig id="F14" position="float">
<label>Figure 14</label>
<caption><p>t-SNE analysis on the emotional dimensions of SEED-IV.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-17-1135850-g0014.tif"/>
</fig>
</sec>
<sec>
<title>4.6. Comparison with existing methods</title>
<p>The recognition performance of the proposed method is compared with several existing studies. The dataset and labeling scheme are the same for all reported studies. <xref ref-type="table" rid="T4">Table 4</xref> details the features and classifiers used in the comparative study. Since recognition accuracy and F1 score are the most commonly used, these two indicators are adopted for comparison. As shown in <xref ref-type="table" rid="T4">Table 4</xref>, the performance of our approach is better than the comparison methods in both the arousal dimension and the valence dimension. The comparison results show that our approach is excellent in multichannel EEG emotion recognition.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Details of previous research.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<th valign="top" align="left"><bold>Study</bold></th>
<th valign="top" align="left"><bold>Feature</bold></th>
<th valign="top" align="left"><bold>Classifier</bold></th>
<th valign="top" align="center" colspan="2"><bold>DEAP</bold></th>
<th valign="top" align="center"><bold>SEED-IV</bold></th>
</tr>
</thead>
<tbody>
<tr style="background-color:&#x00023;919498;color:&#x00023;ffffff">
<td/>
<td/>
<td/>
<td valign="top" align="center"><bold>Arousal</bold></td>
<td valign="top" align="center"><bold>Valence</bold></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Samara et al. (<xref ref-type="bibr" rid="B32">2016</xref>)</td>
<td valign="top" align="left">Band power</td>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">0.7367</td>
<td valign="top" align="center">0.8599</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Guo et al. (<xref ref-type="bibr" rid="B12">2017</xref>)</td>
<td valign="top" align="left">DWT</td>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">0.6279</td>
<td valign="top" align="center">0.6021</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Alhagry et al. (<xref ref-type="bibr" rid="B1">2017</xref>)</td>
<td valign="top" align="left">Raw EEG signals</td>
<td valign="top" align="left">LSTM</td>
<td valign="top" align="center">0.8565</td>
<td valign="top" align="center">0.8545</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Yang and Liu (<xref ref-type="bibr" rid="B41">2019</xref>)</td>
<td valign="top" align="left">Differential entropy</td>
<td valign="top" align="left">TCN</td>
<td valign="top" align="center">0.7140</td>
<td valign="top" align="center">0.7440</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Tripathi et al. (<xref ref-type="bibr" rid="B37">2017</xref>)</td>
<td valign="top" align="left">Statistical parameters</td>
<td valign="top" align="left">DNN</td>
<td valign="top" align="center">0.7313</td>
<td valign="top" align="center">0.7578</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">0.7336</td>
<td valign="top" align="center">0.8141</td>
<td valign="top" align="center">0.8599</td>
</tr> <tr>
<td valign="top" align="left">Gao et al. (<xref ref-type="bibr" rid="B10">2022</xref>)</td>
<td valign="top" align="left">Differential entropy</td>
<td valign="top" align="left">GCN</td>
<td valign="top" align="center">0.8193</td>
<td valign="top" align="center">0.8177</td>
<td valign="top" align="center">-</td>
</tr> <tr>
<td valign="top" align="left">Zhong et al. (<xref ref-type="bibr" rid="B43">2020</xref>)</td>
<td valign="top" align="left">Differential entropy</td>
<td valign="top" align="left">RGNN</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">0.7750</td>
</tr> <tr>
<td valign="top" align="left">Du et al. (<xref ref-type="bibr" rid="B8">2022</xref>)</td>
<td valign="top" align="left">Differential entropy</td>
<td valign="top" align="left">MD-GCN</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">0.9083</td>
</tr> <tr>
<td valign="top" align="left">Li et al. (<xref ref-type="bibr" rid="B23">2023</xref>)</td>
<td valign="top" align="left">Differential entropy</td>
<td valign="top" align="left">FGCN</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">0.7714</td>
</tr> <tr>
<td valign="top" align="left">Vafaei et al. (<xref ref-type="bibr" rid="B38">2023</xref>)</td>
<td valign="top" align="left">Time domain features</td>
<td valign="top" align="left">SAETM</td>
<td valign="top" align="center">0.8037</td>
<td valign="top" align="center">0.8173</td>
<td valign="top" align="center">-</td>
</tr> <tr>
<td valign="top" align="left">The proposed method</td>
<td valign="top" align="left">Time domain features</td>
<td valign="top" align="left">ResGAT</td>
<td valign="top" align="center">0.8706</td>
<td valign="top" align="center">0.8926</td>
<td valign="top" align="center">0.9773</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="conclusions" id="s5">
<title>5. Conclusion</title>
<p>A novel ensemble deep learning framework is proposed in this work. In the framework, the residual network is employed to extract the spatial position information of the electrode channel through the 3D characteristic matrix. The graph neural network is utilized to learn the neural functional connections between different brain regions, and the multi-head self-attention mechanism is used to adaptively adjust the adjacency matrix in the network. The results show the proposed ResGAT framework makes full use of the electrode spatial position information and the intrinsic connection relationship between EEG channels located in different brain regions. Moreover, the emotion recognition performance of the proposed method is compared with some existing methods and shows advantages, which proves the feasibility and effectiveness of the proposed emotion recognition method.</p>
<p>The experiments in this manuscript were conducted on public datasets DEAP and SEED, and the proposed emotion recognition method demonstrated good performance. However, the number of subjects on the dataset is limited, and the effectiveness of its use in a large population needs further verification. The monitoring and regulation of emotional state is of great significance for the psychological and physiological health of individuals. For example, in clinical treatment, monitoring and regulating emotional states can help doctors better understand patients&#x00027; emotional states, thereby providing more personalized and effective treatment plans for patients. In daily life, monitoring and regulating emotional states can help individuals better manage their emotions and improve their quality of life. Moreover, the emotion recognition performance of the proposed method is compared with some existing methods and shows advantages, which proves the feasibility and effectiveness of the proposed emotion recognition method. In addition, the proposed emotion recognition classification model can also be applied in disease diagnosis, such as identification of patients with depression; issuing execution commands to control external devices, helping patients to carry out active rehabilitation training; diagnosis of schizophrenia; quantifying the neurophysiological changes associated with a variety of work-related physical activities (Ismail et al., <xref ref-type="bibr" rid="B18">2023</xref>).</p>
</sec>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="http://www.eecs.qmul.ac.uk/mmv/datasets/deap/">http://www.eecs.qmul.ac.uk/mmv/datasets/deap/</ext-link>.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>HC designed and organized the study and wrote this article. HC, YL, and YC collected and analyzed the data. All authors contributed to manuscript revision, read, and approved the submitted version.</p>
</sec>
</body>
<back>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>This work was supported by National Natural Science Research Fund (61872126), Key Scientific Research Project of Colleges and Universities in Henan Province (19A520004), the Fundamental Research Funds for the Central Universities (NSFRF1616), and Programs for Science and Technology Development of Henan province No. 222102210078.</p>
</sec>
<ack><p>We would like to thank all participants of this study.</p></ack>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alhagry</surname> <given-names>S.</given-names></name> <name><surname>Fahmy</surname> <given-names>A. A.</given-names></name> <name><surname>El-Khoribi</surname> <given-names>R. A.</given-names></name></person-group> (<year>2017</year>). <article-title>Emotion recognition based on EEG using LSTM recurrent neural network</article-title>. <source>Int. J. Adv. Comput. Sci. Appl.</source> <volume>8</volume>, <fpage>355</fpage>&#x02013;<lpage>358</lpage>. <pub-id pub-id-type="doi">10.14569/IJACSA.2017.081046</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alhaj</surname> <given-names>H.</given-names></name> <name><surname>Wisniewski</surname> <given-names>G.</given-names></name> <name><surname>McAllister-Williams</surname> <given-names>R. H.</given-names></name></person-group> (<year>2011</year>). <article-title>The use of the EEG in measuring therapeutic drug action: focus on depression and antidepressants</article-title>. <source>J. Psychopharmacol.</source> <volume>25</volume>, <fpage>1175</fpage>&#x02013;<lpage>1191</lpage>. <pub-id pub-id-type="doi">10.1177/0269881110388323</pub-id><pub-id pub-id-type="pmid">22807041</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Castelnovo</surname> <given-names>A.</given-names></name> <name><surname>Riedner</surname> <given-names>B. A.</given-names></name> <name><surname>Smith</surname> <given-names>R. F.</given-names></name> <name><surname>Tononi</surname> <given-names>G.</given-names></name> <name><surname>Boly</surname> <given-names>M.</given-names></name> <name><surname>Benca</surname> <given-names>R. M.</given-names></name></person-group> (<year>2016</year>). <article-title>Scalp and source power topography in sleepwalking and sleep terrors: a high-density EEG study</article-title>. <source>Sleep</source> <volume>39</volume>, <fpage>1815</fpage>&#x02013;<lpage>1825</lpage>. <pub-id pub-id-type="doi">10.5665/sleep.6162</pub-id><pub-id pub-id-type="pmid">27568805</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chao</surname> <given-names>H.</given-names></name> <name><surname>Dong</surname> <given-names>L.</given-names></name></person-group> (<year>2020</year>). <article-title>Emotion recognition using three-dimensional feature and convolutional neural network from multichannel EEG signals</article-title>. <source>IEEE Sens. J.</source> <volume>21</volume>, <fpage>2024</fpage>&#x02013;<lpage>2034</lpage>. <pub-id pub-id-type="doi">10.1109/JSEN.2020.3020828</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Xie</surname> <given-names>H.</given-names></name> <name><surname>Yu</surname> <given-names>W.</given-names></name></person-group> (<year>2007</year>). <article-title>Characterization of surface EMG signal based on fuzzy entropy</article-title>. <source>IEEE Trans. Neural Syst. Rehabil. Eng.</source> <volume>15</volume>, <fpage>266</fpage>&#x02013;<lpage>272</lpage>. <pub-id pub-id-type="doi">10.1109/TNSRE.2007.897025</pub-id><pub-id pub-id-type="pmid">17601197</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cui</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>A.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name></person-group> (<year>2020</year>). <article-title>EEG-based emotion recognition using an end-to-end regional-asymmetric convolutional neural network</article-title>. <source>Knowledge Based Syst.</source> <volume>205</volume>, <fpage>106243</fpage>. <pub-id pub-id-type="doi">10.1016/j.knosys.2020.106243</pub-id><pub-id pub-id-type="pmid">35892418</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dosovitskiy</surname> <given-names>A.</given-names></name> <name><surname>Beyer</surname> <given-names>L.</given-names></name> <name><surname>Kolesnikov</surname> <given-names>A.</given-names></name> <name><surname>Weissenborn</surname> <given-names>D.</given-names></name> <name><surname>Zhai</surname> <given-names>X.</given-names></name> <name><surname>Unterthiner</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>An image is worth 16x16 words: transformers for image recognition at scale</article-title>. <source>arXiv preprint arXiv:2010.11929</source>.</citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Du</surname> <given-names>G.</given-names></name> <name><surname>Su</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Su</surname> <given-names>K.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Teng</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>A multi-dimensional graph convolution network for EEG emotion recognition</article-title>. <source>IEEE Trans. Instrument. Meas.</source> <volume>71</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1109/TIM.2022.3204314</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>F&#x000FC;rbass</surname> <given-names>F.</given-names></name> <name><surname>Kural</surname> <given-names>M. A.</given-names></name> <name><surname>Gritsch</surname> <given-names>G.</given-names></name> <name><surname>Hartmann</surname> <given-names>M.</given-names></name> <name><surname>Kluge</surname> <given-names>T.</given-names></name> <name><surname>Beniczky</surname> <given-names>S.</given-names></name></person-group> (<year>2020</year>). <article-title>An artificial intelligence-based EEG algorithm for detection of epileptiform EEG discharges: validation against the diagnostic gold standard</article-title>. <source>Clin. Neurophysiol.</source> <volume>131</volume>, <fpage>1174</fpage>&#x02013;<lpage>1179</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2020.02.032</pub-id><pub-id pub-id-type="pmid">32299000</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>Y.</given-names></name> <name><surname>Fu</surname> <given-names>X.</given-names></name> <name><surname>Ouyang</surname> <given-names>T.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name></person-group> (<year>2022</year>). <article-title>EEG-GCN: spatio-temporal and self-adaptive graph convolutional networks for single and multi-view EEG-based emotion recognition</article-title>. <source>IEEE Signal Process. Lett.</source> <volume>29</volume>, <fpage>1574</fpage>&#x02013;<lpage>1578</lpage>. <pub-id pub-id-type="doi">10.1109/LSP.2022.3179946</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gogna</surname> <given-names>A.</given-names></name> <name><surname>Majumdar</surname> <given-names>A.</given-names></name> <name><surname>Ward</surname> <given-names>R.</given-names></name></person-group> (<year>2016</year>). <article-title>Semi-supervised stacked label consistent autoencoder for reconstruction and analysis of biomedical signals</article-title>. <source>IEEE Trans. Biomed. Eng.</source> <volume>64</volume>, <fpage>2196</fpage>&#x02013;<lpage>2205</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2016.2631620</pub-id><pub-id pub-id-type="pmid">27893378</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>K.</given-names></name> <name><surname>Candra</surname> <given-names>H.</given-names></name> <name><surname>Yu</surname> <given-names>H.</given-names></name> <name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Nguyen</surname> <given-names>H. T.</given-names></name> <name><surname>Su</surname> <given-names>S. W.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;EEG-based emotion classification using innovative features and combined SVM and HMM classifier,&#x0201D;</article-title> in <source>2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</source> (<publisher-name>IEEE</publisher-name>), <fpage>489</fpage>&#x02013;<lpage>492</lpage>. <pub-id pub-id-type="pmid">29059916</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guozhen</surname> <given-names>Z.</given-names></name> <name><surname>Jinjing</surname> <given-names>S.</given-names></name> <name><surname>Yan</surname> <given-names>G.</given-names></name> <name><surname>Yongjin</surname> <given-names>L.</given-names></name> <name><surname>Lin</surname> <given-names>Y.</given-names></name> <name><surname>Tao</surname> <given-names>W.</given-names></name></person-group> (<year>2016</year>). <article-title>Advances in emotion recognition based on physiological big data</article-title>. <source>J. Comput. Res. Dev.</source> <volume>53</volume>, <fpage>80</fpage>&#x02013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.7544/issn1000-1239.2016.20150636</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamilton</surname> <given-names>M.</given-names></name></person-group> (<year>1960</year>). <article-title>A rating scale for depression</article-title>. <source>J. Neurol. Neurosurg. Psychiatry</source> <volume>23</volume>, <fpage>56</fpage>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haufe</surname> <given-names>S.</given-names></name> <name><surname>Nikulin</surname> <given-names>V. V.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <name><surname>Nolte</surname> <given-names>G.</given-names></name></person-group> (<year>2013</year>). <article-title>A critical assessment of connectivity measures for EEG data: a simulation study</article-title>. <source>Neuroimage</source> <volume>64</volume>, <fpage>120</fpage>&#x02013;<lpage>133</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2012.09.036</pub-id><pub-id pub-id-type="pmid">27534393</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>F.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2019</year>). <article-title>Ten challenges for EEG-based affective computing</article-title>. <source>Brain Sci. Adv.</source> <volume>5</volume>, <fpage>1</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1177/2096595819896200</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>G.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name> <name><surname>Van Der Maaten</surname> <given-names>L.</given-names></name> <name><surname>Weinberger</surname> <given-names>K. Q.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Densely connected convolutional networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, <fpage>4700</fpage>&#x02013;<lpage>4708</lpage>.</citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ismail</surname> <given-names>L.</given-names></name> <name><surname>Karwowski</surname> <given-names>W.</given-names></name> <name><surname>Hancock</surname> <given-names>P. A.</given-names></name> <name><surname>Taiar</surname> <given-names>R.</given-names></name> <name><surname>Fernandez-Sumano</surname> <given-names>R.</given-names></name></person-group> (<year>2023</year>). <article-title>Electroencephalography (EEG) physiological indices reflecting human physical performance: a systematic review using updated prisma</article-title>. <source>J. Integr. Neurosci.</source> <volume>22</volume>, <fpage>62</fpage>. <pub-id pub-id-type="doi">10.31083/j.jin2203062</pub-id><pub-id pub-id-type="pmid">37258434</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kipf</surname> <given-names>T. N.</given-names></name> <name><surname>Welling</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>Semi-supervised classification with graph convolutional networks</article-title>. <source>arXiv preprint arXiv:1609.02907</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1609.02907</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Koelstra</surname> <given-names>S.</given-names></name> <name><surname>Muhl</surname> <given-names>C.</given-names></name> <name><surname>Soleymani</surname> <given-names>M.</given-names></name> <name><surname>Lee</surname> <given-names>J.-S.</given-names></name> <name><surname>Yazdani</surname> <given-names>A.</given-names></name> <name><surname>Ebrahimi</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2011</year>). <article-title>DEAP: a database for emotion analysis; using physiological signals</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>3</volume>, <fpage>18</fpage>&#x02013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1109/T-AFFC.2011.15</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krizhevsky</surname> <given-names>A.</given-names></name> <name><surname>Sutskever</surname> <given-names>I.</given-names></name> <name><surname>Hinton</surname> <given-names>G. E.</given-names></name></person-group> (<year>2017</year>). <article-title>ImageNet classification with deep convolutional neural networks</article-title>. <source>Commun. ACM</source> <volume>60</volume>, <fpage>84</fpage>&#x02013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1145/3065386</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>Z.</given-names></name> <name><surname>He</surname> <given-names>H.</given-names></name></person-group> (<year>2018a</year>). <article-title>Hierarchical convolutional neural networks for EEG-based emotion recognition</article-title>. <source>Cogn. Comput.</source> <volume>10</volume>, <fpage>368</fpage>&#x02013;<lpage>380</lpage>. <pub-id pub-id-type="doi">10.1007/s12559-017-9533-x</pub-id><pub-id pub-id-type="pmid">37015466</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M.</given-names></name> <name><surname>Qiu</surname> <given-names>M.</given-names></name> <name><surname>Kong</surname> <given-names>W.</given-names></name> <name><surname>Zhu</surname> <given-names>L.</given-names></name> <name><surname>Ding</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Fusion graph representation of EEG for emotion recognition</article-title>. <source>Sensors</source> <volume>23</volume>, <fpage>1404</fpage>. <pub-id pub-id-type="doi">10.3390/s23031404</pub-id><pub-id pub-id-type="pmid">36772444</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>X.</given-names></name> <name><surname>Lu</surname> <given-names>S.</given-names></name></person-group> (<year>2018b</year>). <article-title>Emotion recognition from multichannel EEG signals using k-nearest neighbor classification</article-title>. <source>Technol. Health Care</source> <volume>26</volume>, <fpage>509</fpage>&#x02013;<lpage>519</lpage>. <pub-id pub-id-type="doi">10.3233/THC-174836</pub-id><pub-id pub-id-type="pmid">29758974</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Huang</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>H.</given-names></name> <name><surname>Zhong</surname> <given-names>N.</given-names></name></person-group> (<year>2017</year>). <article-title>Human emotion recognition with electroencephalographic multidimensional features by hybrid deep neural networks</article-title>. <source>Appl. Sci.</source> <volume>7</volume>, <fpage>1060</fpage>. <pub-id pub-id-type="doi">10.3390/app7101060</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Zheng</surname> <given-names>W.</given-names></name> <name><surname>Zong</surname> <given-names>Y.</given-names></name> <name><surname>Cui</surname> <given-names>Z.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name> <name><surname>Zhou</surname> <given-names>X.</given-names></name></person-group> (<year>2018c</year>). <article-title>A bi-hemisphere domain adversarial neural network model for EEG emotion recognition</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>12</volume>, <fpage>494</fpage>&#x02013;<lpage>504</lpage>. <pub-id pub-id-type="doi">10.1109/TAFFC.2018.2885474</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mohammadi</surname> <given-names>M.</given-names></name> <name><surname>Al-Azab</surname> <given-names>F.</given-names></name> <name><surname>Raahemi</surname> <given-names>B.</given-names></name> <name><surname>Richards</surname> <given-names>G.</given-names></name> <name><surname>Jaworska</surname> <given-names>N.</given-names></name> <name><surname>Smith</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Data mining EEG signals in depression for their diagnostic value</article-title>. <source>BMC Med. Inform. Decis. Mak.</source> <volume>15</volume>, <fpage>108</fpage>. <pub-id pub-id-type="doi">10.1186/s12911-015-0227-6</pub-id><pub-id pub-id-type="pmid">26699540</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Niedermeyer</surname> <given-names>E.</given-names></name> <name><surname>da Silva</surname> <given-names>F. L.</given-names></name></person-group> (<year>2005</year>). <source>Electroencephalography: Basic Principles, Clinical Applications, and Related Fields</source>. <publisher-name>Lippincott Williams &#x00026; Wilkins</publisher-name>.</citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nolte</surname> <given-names>G.</given-names></name> <name><surname>Bai</surname> <given-names>O.</given-names></name> <name><surname>Wheaton</surname> <given-names>L.</given-names></name> <name><surname>Mari</surname> <given-names>Z.</given-names></name> <name><surname>Vorbach</surname> <given-names>S.</given-names></name> <name><surname>Hallett</surname> <given-names>M.</given-names></name></person-group> (<year>2004</year>). <article-title>Identifying true brain interaction from EEG data using the imaginary part of coherency</article-title>. <source>Clin. Neurophysiol.</source> <volume>115</volume>, <fpage>2292</fpage>&#x02013;<lpage>2307</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2004.04.029</pub-id><pub-id pub-id-type="pmid">15351371</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nolte</surname> <given-names>G.</given-names></name> <name><surname>Ziehe</surname> <given-names>A.</given-names></name> <name><surname>Nikulin</surname> <given-names>V. V.</given-names></name> <name><surname>Schl&#x000F6;gl</surname> <given-names>A.</given-names></name> <name><surname>Kr&#x000E4;mer</surname> <given-names>N.</given-names></name> <name><surname>Brismar</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2008</year>). <article-title>Robustly estimating the flow direction of information in complex physical systems</article-title>. <source>Phys. Rev. Lett.</source> <volume>100</volume>, <fpage>234101</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.100.234101</pub-id><pub-id pub-id-type="pmid">18643502</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramirez</surname> <given-names>P. M.</given-names></name> <name><surname>Desantis</surname> <given-names>D.</given-names></name> <name><surname>Opler</surname> <given-names>L. A.</given-names></name></person-group> (<year>2001</year>). <article-title>EEG biofeedback treatment of add: a viable alternative to traditional medical intervention?</article-title> <source>Ann. N. Y. Acad. Sci.</source> <volume>931</volume>, <fpage>342</fpage>&#x02013;<lpage>358</lpage>. <pub-id pub-id-type="doi">10.1111/j.1749-6632.2001.tb05789.x</pub-id><pub-id pub-id-type="pmid">11462752</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Samara</surname> <given-names>A.</given-names></name> <name><surname>Menezes</surname> <given-names>M. L. R.</given-names></name> <name><surname>Galway</surname> <given-names>L.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Feature extraction for emotion recognition and modelling using neurophysiological data,&#x0201D;</article-title> in <source>2016 15th International Conference on Ubiquitous Computing and Communications and 2016 International Symposium on Cyberspace and Security (IUCC-CSS)</source> (<publisher-name>IEEE</publisher-name>), <fpage>138</fpage>&#x02013;<lpage>144</lpage>.</citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2014</year>). <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv preprint arXiv:1409.1556</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1409.1556</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sung</surname> <given-names>M.</given-names></name> <name><surname>Marci</surname> <given-names>C.</given-names></name> <name><surname>Pentland</surname> <given-names>A.</given-names></name></person-group> (<year>2005</year>). <source>Objective Physiological and Behavioral Measures for Identifying and Tracking Depression State in Clinically Depressed Patients</source>. <publisher-name>Massachusetts Institute of Technology Media Laboratory, Cambridge, MA</publisher-name>.</citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Supp</surname> <given-names>G. G.</given-names></name> <name><surname>Schl&#x000F6;gl</surname> <given-names>A.</given-names></name> <name><surname>Trujillo-Barreto</surname> <given-names>N.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>M. M.</given-names></name> <name><surname>Gruber</surname> <given-names>T.</given-names></name></person-group> (<year>2007</year>). <article-title>Directed cortical information flow during human object recognition: analyzing induced EEG gamma-band responses in brain&#x00027;s source space</article-title>. <source>PLoS ONE</source> <volume>2</volume>, <fpage>e684</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0000684</pub-id><pub-id pub-id-type="pmid">17668062</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szegedy</surname> <given-names>C.</given-names></name> <name><surname>Liu</surname> <given-names>W.</given-names></name> <name><surname>Jia</surname> <given-names>Y.</given-names></name> <name><surname>Sermanet</surname> <given-names>P.</given-names></name> <name><surname>Reed</surname> <given-names>S.</given-names></name> <name><surname>Anguelov</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>&#x0201C;Going deeper with convolutions,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, <fpage>1</fpage>&#x02013;<lpage>9</lpage>.</citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tripathi</surname> <given-names>S.</given-names></name> <name><surname>Acharya</surname> <given-names>S.</given-names></name> <name><surname>Sharma</surname> <given-names>R. D.</given-names></name> <name><surname>Mittal</surname> <given-names>S.</given-names></name> <name><surname>Bhattacharya</surname> <given-names>S.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Using deep and convolutional neural networks for accurate emotion classification on deap dataset,&#x0201D;</article-title> in <source>Twenty-ninth IAAI conference</source>. <pub-id pub-id-type="pmid">33424547</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vafaei</surname> <given-names>E.</given-names></name> <name><surname>Nowshiravan Rahatabad</surname> <given-names>F.</given-names></name> <name><surname>Setarehdan</surname> <given-names>S. K.</given-names></name> <name><surname>Azadfallah</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Extracting a novel emotional EEG topographic map based on a stacked autoencoder network</article-title>. <source>J. Healthcare Eng.</source> <volume>2023</volume>, <fpage>9223599</fpage>. <pub-id pub-id-type="doi">10.1155/2023/9223599</pub-id><pub-id pub-id-type="pmid">36714412</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Veli&#x0010D;kovi&#x00107;</surname> <given-names>P.</given-names></name> <name><surname>Cucurull</surname> <given-names>G.</given-names></name> <name><surname>Casanova</surname> <given-names>A.</given-names></name> <name><surname>Romero</surname> <given-names>A.</given-names></name> <name><surname>Lio</surname> <given-names>P.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>2017</year>). <article-title>Graph attention networks</article-title>. <source>arXiv preprint arXiv:1710.10903</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1710.10903</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xing</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>Z.</given-names></name> <name><surname>Xu</surname> <given-names>T.</given-names></name> <name><surname>Shu</surname> <given-names>L.</given-names></name> <name><surname>Hu</surname> <given-names>B.</given-names></name> <name><surname>Xu</surname> <given-names>X.</given-names></name></person-group> (<year>2019</year>). <article-title>SAE&#x0002B; LSTM: a new framework for emotion recognition from multi-channel EEG</article-title>. <source>Front. Neurorobot.</source> <volume>13</volume>, <fpage>37</fpage>. <pub-id pub-id-type="doi">10.3389/fnbot.2019.00037</pub-id><pub-id pub-id-type="pmid">31244638</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;EEG-based emotion recognition using temporal convolutional network,&#x0201D;</article-title> in <source>2019 IEEE 8th Data Driven Control and Learning Systems Conference (DDCLS)</source> (<publisher-name>IEEE</publisher-name>), <fpage>437</fpage>&#x02013;<lpage>442</lpage>. <pub-id pub-id-type="pmid">34838262</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>W.-L.</given-names></name> <name><surname>Liu</surname> <given-names>W.</given-names></name> <name><surname>Lu</surname> <given-names>Y.</given-names></name> <name><surname>Lu</surname> <given-names>B.-L.</given-names></name> <name><surname>Cichocki</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Emotionmeter: a multimodal framework for recognizing human emotions</article-title>. <source>IEEE Trans. Cybernet.</source> <volume>49</volume>, <fpage>1110</fpage>&#x02013;<lpage>1122</lpage>. <pub-id pub-id-type="doi">10.1109/TCYB.2018.2797176</pub-id><pub-id pub-id-type="pmid">29994384</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhong</surname> <given-names>P.</given-names></name> <name><surname>Wang</surname> <given-names>D.</given-names></name> <name><surname>Miao</surname> <given-names>C.</given-names></name></person-group> (<year>2020</year>). <article-title>EEG-based emotion recognition using regularized graph neural networks</article-title>. <source>IEEE Trans. Affect. Comput.</source> <volume>13</volume>, <fpage>1290</fpage>&#x02013;<lpage>1301</lpage>. <pub-id pub-id-type="doi">10.1109/TAFFC.2020.2994159</pub-id></citation>
</ref>
</ref-list> 
</back>
</article>