<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Hum. Neurosci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Human Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Hum. Neurosci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1662-5161</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnhum.2026.1751058</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Advancing individual finger classification through a sandwich enhanced CBAM network with ultra-high-density EEG data</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Zhang</surname> <given-names>Xinguo</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhang</surname> <given-names>Yiman</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/375084"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Peng</surname> <given-names>Hong</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1542976"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Deng</surname> <given-names>Tao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3192830"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Key Laboratory of China&#x00027;s Ethnic Languages and Information Technology of Ministry of Education, Chinese National Information Technology Research Institute, Northwest Minzu University</institution>, <city>Lanzhou</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>School of Artificial Intelligence, Hebei Institute of Communications</institution>, <city>Shijiazhuang</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>School of Information Science and Engineering, Lanzhou University</institution>, <city>Lanzhou</city>, <country country="cn">China</country></aff>
<aff id="aff4"><label>4</label><institution>School of Mathematics and Computer Science, Northwest Minzu University</institution>, <city>Lanzhou</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Tao Deng, <email xlink:href="mailto:dttom@lzu.edu.cn">dttom@lzu.edu.cn</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-18">
<day>18</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>20</volume>
<elocation-id>1751058</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>03</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2026 Zhang, Zhang, Peng and Deng.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Zhang, Zhang, Peng and Deng</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-18">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Ultra-High-Density Electroencephalography (uHD EEG) has gained increasing attention for its potential in individual finger decoding. However, accurately classifying these movements remains challenging due to the subtle spatial overlaps in cortical activity, which standard architectures often fail to isolate.</p>
</sec>
<sec>
<title>Methods</title>
<p>To address this, we propose the Sandwich enhanced Convolutional Block Attention Module (SCBAM). The unique sandwich structure integrates dual attention mechanisms between convolutional layers, enabling the network to more effectively refine high-dimensional spatial features.</p>
</sec>
<sec>
<title>Results and discussion</title>
<p>The proposed network achieves an average accuracy of 78.63 (1.56)% in binary classification across ten finger pairs in five subjects, with the highest accuracy of 85% obtained at Thumb vs. Ring. The proposed network achieves an average accuracy of 61.12 (0.95)% in five-class classification across five subjects, with a highest accuracy of 62.36% on subject S2. The five-class classification is performed using 10 binary classifiers under a one-vs.-one strategy. Notably, five-class classification of individual fingers has not been extensively explored in the current literature, particularly with high-density EEG (HDEEG) data. This study addresses this gap, offering a valuable reference for future discussions. We conduct ablation studies to investigate the individual and synergistic effects of the modules in the proposed model. The results highlight the effects of two sequential attention mechanisms in this task. We conduct comparative experiments of our proposed model against six benchmark networks. The results from SCBAM significantly outperform these established models with FBCSP features. The proposed SCBAM significantly improves accuracy in binary finger classification compared to SVM and MLP using the same uHD EEG dataset. In summary, this study presents a high-performance hybrid network for individual finger classification and highlights the potential of uHD EEG for dexterous task decoding in Brain-Computer Interfaces (BCI).</p>
</sec></abstract>
<kwd-group>
<kwd>attention mechanism</kwd>
<kwd>brain-computer interfaces</kwd>
<kwd>Convolutional Block Attention Module</kwd>
<kwd>individual finger classification</kwd>
<kwd>ultra-high-density EEG</kwd>
</kwd-group>
<funding-group>
<award-group id="gs1">
<funding-source id="sp1">
<institution-wrap>
<institution>National Natural Science Foundation of China</institution>
<institution-id institution-id-type="doi" vocab="open-funder-registry" vocab-identifier="10.13039/open_funder_registry">10.13039/501100001809</institution-id>
</institution-wrap>
</funding-source>
</award-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported in part by the National Natural Science Foundation of China, China, under Grant No. 32060234, in part by the Longyuan Young Talents of Gansu Province [2023]11, China, and in part by the Lanshan Talents Project of Northwest Minzu University, China, under Grant No. xbmulsrc202411.</funding-statement>
</funding-group>
<counts>
<fig-count count="11"/>
<table-count count="8"/>
<equation-count count="6"/>
<ref-count count="56"/>
<page-count count="20"/>
<word-count count="13816"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Brain-Computer Interfaces</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Brain-computer interface (BCI) offers a promising communication solution for individuals with severe neuromuscular disorders (<xref ref-type="bibr" rid="B52">Wolpaw et al., 2002</xref>). A main task of BCI is motor imagery (<xref ref-type="bibr" rid="B36">Pfurtscheller et al., 2000</xref>). These systems provide high temporal resolution but offer limited spatial resolution that constrains the precise localization of neural sources (<xref ref-type="bibr" rid="B38">Pfurtscheller et al., 1997</xref>). Traditional EEG systems are according to standard layouts (e.g., 10-20 or 10-10 systems), with inter-electrode distances ranging roughly from 20 mm to 60 mm (<xref ref-type="bibr" rid="B3">Chatrian et al., 1985</xref>; <xref ref-type="bibr" rid="B8">Jasper, 1958</xref>), brain rhythms (alpha, beta, and gamma). Event-related synchronization and desynchronization are widely applied in motor imagery decoding (<xref ref-type="bibr" rid="B37">Pfurtscheller and Lopes da Silva, 1999</xref>). EEG decoding has been recognized as a pattern recognition task for decades (<xref ref-type="bibr" rid="B12">Krusienski et al., 2006</xref>). Spatial filters, such as ear-reference, common average reference (CAR), small Laplacian, and large Laplacian, have been widely applied in this context (<xref ref-type="bibr" rid="B26">McFarland et al., 1997</xref>).</p>
<p>The pioneering study of <xref ref-type="bibr" rid="B32">M&#x000FC;ller-Gerking et al. (1999)</xref> introduced CSP as a feature extraction method for single-trial EEG classification in movement tasks. <xref ref-type="bibr" rid="B23">Lotte and Guan (2011)</xref> unified different kinds of regularization CSP methods, which addressed problems of overfitting and improving generalization. Combining with the frequency band of the EEG signal, the Filter Bank Common Spatial Pattern (FBCSP) algorithm was shown to have superior classification accuracies on various datasets, including BCI Competition IV Datasets 2a and 2b (<xref ref-type="bibr" rid="B1">Ang et al., 2008</xref>; <xref ref-type="bibr" rid="B39">Quadrianto et al., 2007</xref>). The classification algorithms in the field of Brain-Computer Interface (BCI) have evolved significantly over the last two decades. According to a 2007 review (<xref ref-type="bibr" rid="B22">Lotte et al., 2007</xref>), the popular classifiers of that time included linear models, such as Linear Discriminant Analysis (LDA) and Support Vector Machines (SVM). Currently, the trend of classifier algorithms has shifted toward deep neural networks due to advancements in computational power and the availability of large datasets (<xref ref-type="bibr" rid="B21">Lotte et al., 2018</xref>).</p>
<p>The EEG signal has achieved significant accuracy in decoding prominent movements, such as the left hand, right hand, foot, and tongue, over the past few decades (<xref ref-type="bibr" rid="B27">McFarland et al., 2000</xref>; <xref ref-type="bibr" rid="B15">Lange et al., 2016</xref>). However, decoding dexterous individual finger movements has shown limited accuracy (<xref ref-type="bibr" rid="B19">Liao et al., 2014</xref>; <xref ref-type="bibr" rid="B34">Paek et al., 2014</xref>). For individual finger movements, where ECoG signals were considered more feasible (<xref ref-type="bibr" rid="B19">Liao et al., 2014</xref>). ECoG signals demonstrate high spatial resolution and a high signal-to-noise ratio. Researchers have put efforts into unveiling the properties of ECoG during motor tasks for decades (<xref ref-type="bibr" rid="B51">Volkova et al., 2019</xref>; <xref ref-type="bibr" rid="B54">Wu et al., 2024</xref>). High frequency components such as gamma and super gamma rhythms, together with the low frequency rhythms such as delta, alpha, and beta, have also been widely applied in decoding (<xref ref-type="bibr" rid="B13">Kub&#x000E1;nek et al., 2009</xref>). The properties of ECoG rhythms&#x02013;specifically phase (<xref ref-type="bibr" rid="B29">Miller et al., 2012</xref>), power (<xref ref-type="bibr" rid="B31">Miller K. J. et al., 2009</xref>), and modulation (<xref ref-type="bibr" rid="B28">Miller K. et al., 2009</xref>)&#x02014;have been extensively explored in motor decoding tasks. The individual fingers&#x00027; continuous flexion and extension trajectory decoding from human ECoG is a topic in BCI competition IV (<xref ref-type="bibr" rid="B49">Tangermann et al., 2012</xref>). The discrete classification of individual finger flexion decoding was reported in <xref ref-type="bibr" rid="B45">Shenoy et al. (2007)</xref>, <xref ref-type="bibr" rid="B13">Kub&#x000E1;nek et al. (2009)</xref>, and <xref ref-type="bibr" rid="B4">Chestek et al. (2013)</xref>. However, ECoG collection requires craniotomy to place electrodes on the surface of the cortex, making the procedure high-risk and limiting its use in daily applications (<xref ref-type="bibr" rid="B30">Miller et al., 2007</xref>; <xref ref-type="bibr" rid="B7">Hotson et al., 2016</xref>; <xref ref-type="bibr" rid="B9">Jiang et al., 2018</xref>).</p>
<p>Ultra-high-density EEG (uHD EEG) enhances spatial resolution by using a significantly larger number of electrodes (&#x0003E;256) across the scalp (<xref ref-type="bibr" rid="B11">Krishnan et al., 2018</xref>; <xref ref-type="bibr" rid="B25">Ma et al., 2022</xref>). This increased density allows for more precise localization of brain activity and detection of subtle neural signals, with applications in both clinical (<xref ref-type="bibr" rid="B14">Kumar et al., 2018</xref>) and research settings (<xref ref-type="bibr" rid="B44">Schreiner et al., 2023</xref>). These advances facilitate decoding of individual finger movements (<xref ref-type="bibr" rid="B19">Liao et al., 2014</xref>), which are challenging or impossible with conventional EEG (<xref ref-type="bibr" rid="B43">Schreiner et al., 2024</xref>). <xref ref-type="bibr" rid="B18">Lee et al. (2022)</xref> and <xref ref-type="bibr" rid="B33">Nemes et al. (2024)</xref> reported the binary classification accuracy of ten individual finger pairs. The study in <xref ref-type="bibr" rid="B18">Lee et al. (2022)</xref> achieved an average accuracy of 64.8% using Support Vector Machines (SVM), while <xref ref-type="bibr" rid="B33">Nemes et al. (2024)</xref> attained an average accuracy of 65.68% with Multi-Layer Perceptron (MLP). These results reveal there is still room for improvement in binary finger classification. Furthermore, to our knowledge, there is currently no literature addressing the five-class classification of the individual fingers with uHD EEG data, indicating a research gap in this area.</p>
<p>Deep learning has advanced the analysis and classification in BCI. Deep learning networks, such as Convolutional Neural Networks (CNN) (<xref ref-type="bibr" rid="B17">LeCun and Bengio, 1998</xref>), Recurrent Neural Networks (RNNs) (<xref ref-type="bibr" rid="B46">Sherstinsky, 2018</xref>), and attention mechanisms (<xref ref-type="bibr" rid="B50">Vaswani et al., 2017</xref>; <xref ref-type="bibr" rid="B35">Park et al., 2018</xref>), have been successfully used for neural decoding (<xref ref-type="bibr" rid="B56">Yao et al., 2024</xref>). Deep convolutional networks (ConvNets) are capable of extracting multiple levels of feature directly from raw EEG signals (<xref ref-type="bibr" rid="B10">Kern et al., 2019</xref>). <xref ref-type="bibr" rid="B2">Bashivan et al. (2016)</xref> developed the EEGLearn network, drawing inspiration from the popular Visual Geometry Group (VGG) architecture in the image processing field, which was then applied to the context of the mental load task. <xref ref-type="bibr" rid="B16">Lawhern et al. (2018)</xref> proposed EEGNet, which utilizes the temporal and spatial information from EEG signals through two convolutional layers and implements depthwise convolution in the spatial layer by assigning channel-wise filters. <xref ref-type="bibr" rid="B42">Schirrmeister et al. (2017)</xref> designed a shallow and a deep ConvNets for classifying movements of the left hand, right hand, foot, and rest with BCI Competition IV 2a and 2b EEG data. Shallow ConvNets combined with intermediate FBCSP features offer advantages of lightweight architecture, fewer parameters, and reduced tuning time, ultimately achieving comparable results (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>). The evolution of attention mechanisms has significantly advanced the capabilities of neural network architectures. Initially, attention mechanisms (<xref ref-type="bibr" rid="B50">Vaswani et al., 2017</xref>) allowed models to selectively focus on specific parts of input data, enhancing their ability to capture relevant features in natural language processing (NLP). This was further developed by the Convolutional Block Attention Module (CBAM) (<xref ref-type="bibr" rid="B53">Woo et al., 2018</xref>), a lightweight module that operates on intermediate feature maps for image processing. The key characteristic of CBAM is its ability to process channel and spatial information sequentially through two separate attention modules.</p>
<p>The combination of convolutional layers and attention mechanisms has been applied for EEG decoding tasks. <xref ref-type="bibr" rid="B5">Eldele et al. (2021)</xref> present a novel approach that combines a multi-resolution convolutional neural network (MRCNN) with a multi-head transformer for sleep stage classification from single-channel EEG data. They named their model the attensleep network. <xref ref-type="bibr" rid="B47">Shi et al. (2023)</xref> come up with a Multi-band EEG Transformer (MEET) model for brain state classification. In the MEET, a multi-head attention mechanism is applied across multiple frequency bands of images of EEG signals. The fusion of these frequency band images enhances MEET&#x00027;s ability to discern brain states. In a recent study reported by <xref ref-type="bibr" rid="B24">Ma et al. (2024)</xref>, a model combining a two-branch 3D convolutional network and a temporal attention mechanism has been proposed and applied to the motor imagery task with the BCI Competition IV 2a and 2b dataset. <xref ref-type="bibr" rid="B48">Song et al. (2023)</xref> proposed the EEG Conformer, a unified architecture that leverages CNNs for local feature learning and Transformers for global correlation extraction to improve EEG signal decoding. <xref ref-type="bibr" rid="B20">Liu et al. (2024)</xref> introduce MSVTNet, a novel end-to-end multi-scale vision transformer that integrates CNN-based local spatio-temporal extraction with multiscale feature tokenization to enhance MI-EEG classification. Deep Neural Networks (DNNs) are increasingly applied and crucial for advancing Brain-Computer Interface (BCI) technologies and rehabilitation devices (<xref ref-type="bibr" rid="B54">Wu et al., 2024</xref>). The main contributions of this study are as follows:</p>
<list list-type="bullet">
<list-item><p>This study presents a network to decode individual finger movements from uHD EEG data. We achieve an average accuracy of 78.63% in binary classification among 10 finger pairs and 61.12% in five-class classification across 5 subjects. To our knowledge, this is the first attempt at five-class classification of individual finger movements using uHD EEG, demonstrating the feasibility of this approach for the task.</p></list-item>
<list-item><p>We propose a model integrating convolutional layers and Convolutional Block Attention Module (CBAM) for this task. The model begins with a convolutional layer for low-level features, followed by two attention mechanisms that process preceding filters, spatial, and rhythm domain information. This model is completed by an additional convolutional layer for high-level feature extraction. We name this network Sandwich enhanced CBAM (SCBAM). Additionally, we conduct ablation studies to evaluate the contribution of each module and to analyze the synergistic effects of multiple modules.</p></list-item>
<list-item><p>We process the uHD EEG data with the Common Average Reference (CAR) method and utilize the FBCSP for feature extraction. The extracted features after the FBCSP are then fed into our proposed model for classification.</p></list-item>
<list-item><p>Comparative experiments are conducted between our SCBAM model and several state-of-the-art networks, including EEGNet (<xref ref-type="bibr" rid="B16">Lawhern et al., 2018</xref>), EEGLearn (<xref ref-type="bibr" rid="B2">Bashivan et al., 2016</xref>), Shallow ConvNets (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>), Deep ConvNets (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>), EEG Conformer (<xref ref-type="bibr" rid="B48">Song et al., 2023</xref>), and Msvt Net (<xref ref-type="bibr" rid="B20">Liu et al., 2024</xref>) to evaluate the effectiveness of the proposed model.</p></list-item>
<list-item><p>We compare our classification results obtained from uHD EEG data with results gained from current literature of the same task, including those decoded from uHD EEG, EEG, and ECoG. This analysis demonstrates that uHD EEG is feasible and has the potential to improve performance on dexterous movement tasks.</p></list-item>
</list>
<p>The rest of this study is organized as follows: Section 2 illustrates the methodology of data processing and network design. Section 3 presents experimental results and analysis, and Section 4 discusses the proposed networks&#x00027; performance and outlines future work. Finally, Section 5 provides a comprehensive conclusion to the study.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<label>2</label>
<title>Materials and methods</title>
<p>In this section, we reiterate the collecting experiments of the uHD EEG dataset from the (<xref ref-type="bibr" rid="B18">Lee et al., 2022</xref>). Then, we detail the classification method employed for uHD EEG. To enhance signal quality, we utilize a Common Average Reference (CAR) for spatial filtering and apply a notch filter to eliminate power line noise. We apply the Filter Bank Common Spatial Pattern (FBCSP) for feature extraction. We design an innovative network to perform classification tasks in two scenarios. In the binary classification scenario, we classify 10 movement pairs generated by 5 fingers. For the five-class classification scenario, we build upon the 10 binary classifiers using a one-vs-one strategy with majority voting. We conduct ablation experiments to investigate the contributions of the components in the proposed network. We conduct rigorous performance evaluations on various architectures and state-of-the-art models, including EEGNet (<xref ref-type="bibr" rid="B16">Lawhern et al., 2018</xref>), EEGLearn (<xref ref-type="bibr" rid="B2">Bashivan et al., 2016</xref>), Shallow ConvNets (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>), Deep ConvNets (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>), EEG Conformer (<xref ref-type="bibr" rid="B48">Song et al., 2023</xref>), and Msvt Net (<xref ref-type="bibr" rid="B20">Liu et al., 2024</xref>). The workflow of the uHD EEG is illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>The workflow of the uHD EEG.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0001.tif">
<alt-text content-type="machine-generated">Flowchart illustrating a neural network pipeline for finger classification using raw data. Steps include data preprocessing, feature extraction, sequential convolution layers with CBAM attention, max pooling, flattening, dense layer, and classifier output for thumb, index, ring, and little.</alt-text>
</graphic>
</fig>
<sec>
<label>2.1</label>
<title>uHD EEG dataset</title>
<p>The uHD EEG data collection was approved by the Institutional Review Board of Korea Advanced Institute of Science and Technology (KH2018-127), and the gained dataset was published on OSF (<xref ref-type="bibr" rid="B18">Lee et al., 2022</xref>). The uHD EEG was collected from a finger movement experiment involving five healthy subjects (S1-S5). The participants used their dominant hands for testing, whereas S1 was left-handed and S2-S5 were right-handed. Each participant completed 10 sequential runs. Each run began with a 30-second baseline period, followed by alternating rest and finger extension periods. The rest period lasted between 3 and 4 s, while the extension (task) period lasted 5 s. During the task, participants extended the finger indicated by a colored cue displayed on a screen. Each finger was cued five times in a pseudo-random order, resulting in 25 finger extensions per run. Researchers use a novel g. Pangolin system to measure brain signal. The system achieves high spatial resolution, with 16 electrode grids (16 electrodes for each grid, 256 electrodes in total) placed over the contralateral sensorimotor cortex. The electrode grids were attached sequentially in the downward direction laterally of the head, starting with the first electrode grid at the Cz position. Dense distribution and small-sized electrodes result in an inter-electrode distance of 8.6 mm (uHD EEG). The sampling rate is 600 Hz, leading to 3,000 samples of each extension trail. The high number of electrodes and the high sample rate ensure that the data contain rich spatial and temporal information. This high sample rate also captures sufficient frequency bands for brain rhythms, which are key information carriers. Thus, the collected uHD EEG data encompass multi-domain information.</p>
</sec>
<sec>
<label>2.2</label>
<title>Data processing</title>
<p>The data processing consists of two stages: preprocessing and feature extraction. In preprocessing, the raw data are initially standardized using a z-score, and then undergo common average referencing (CAR) to reduce interference from neighboring electrodes. Subsequently, a 4th-order Butterworth notch filter cascade is applied to eliminate interference of line power at 60 Hz and its harmonics (120 Hz, 180 Hz, and 240 Hz). We apply the FBCSP to extract spatial and frequency features. The FBCSP includes two steps:</p>
<p>Firstly, we utilize the Common Spatial Pattern (CSP) method to decompose the joint covariance matrices of the uHD EEG signals from two classes. CSP identifies spatial filters that lead to optimal discrimination between two classes:</p>
<disp-formula id="EQ1"><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>J</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mi>w</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(1)</label></disp-formula>
<p>Specifically, the CSP components refer to the eigenvector corresponding to the large eigenvalue of the joint covariance matrix (<inline-formula><mml:math id="M2"><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:math></inline-formula>) associated with the first class, while the last CSP component corresponds to the small eigenvalue of the joint covariance matrix reflecting the large eigenvalue of the second class. To determine the optimal number of CSP components, experiments are conducted with the number of components ranging from 2 to 20. Ultimately, it is determined that a 12 (6 pairs) CSP components provides the best performance for the classification tasks, with 6 components corresponding to the first class and 6 to the second class for binary classification. Two CSP component patterns for movement pairs (thumb vs ring) are depicted in <xref ref-type="fig" rid="F2">Figure 2</xref> as an example.</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>The topology patterns of the CSP components. <bold>(a)</bold> The pattern of the first CSP component. <bold>(b)</bold> The pattern of the last CSP component. The X-axis and Y-axis represent the medial-lateral and anterior-posterior directions of the human scalp, with the coordinates indicating the locations of each electrode. The color bar illustrates the value associated with each electrode, where red represents positive values and blue represents negative values. The intensity of the color indicates the absolute values, with high intensity representing large absolute values.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0002.tif">
<alt-text content-type="machine-generated">Contour plots compare the spatial distributions of values for the first and last common spatial pattern (CSP) components. Both plots display a range of values indicated by a color bar legend, with axes labeled as X-axis and Y-axis. The left plot shows a wider value range from negative to positive, while the right plot presents a narrower range. Each plot is labeled below: &#x0201C;a. First CSP component&#x0201D; and &#x0201C;b. Last CSP component.&#x0201D; Data point indices are shown with small numbers throughout each plot.</alt-text>
</graphic>
</fig>
<p>Secondly, multiple 4th-order Butterworth band-pass filters are utilized to extract EEG rhythms: delta (1&#x02013;4 Hz), alpha (8&#x02013;12 Hz), beta (12&#x02013;30 Hz), gamma (30&#x02013;50 Hz), high gamma (75&#x02013;115 Hz), and super high gamma (125&#x02013;150 Hz). For CSP filtered rhythms, the centered mean power is calculated by subtracting the respective mean, squaring to get the power. The powers are log-transformed to enhance Gaussianity. Automated artifact rejection was performed to ensure signal integrity. For each channel, data segments with Z-scores greater than 6 were marked as artifacts. Entire experimental runs were rejected if more than 10% of the channels were classified as bad. All preprocessing parameters (<italic>z</italic>-score &#x003BC;/&#x003C3;), and all label-supervised spatial filters FBCSP were fitted using training runs only (runs 1-8) and then applied unchanged to test runs (runs 9&#x02013;10). Hyperparameters were selected using validation within the training runs; test runs were used only once for final reporting. Notably, the percentage of data reduction from FBCSP is approximately 28%, and the output of the FBCSP will be fed into the proposed network.</p>
</sec>
<sec>
<label>2.3</label>
<title>Network design</title>
<p>In this study, we investigate the network performance by integrating convolutional layers with multiple current popular attention mechanisms, including self-attention mechanism, the Convolutional Block Attention Module (CBAM), and multi-head transformers. We define the network in a sandwich configuration, placing convolutional layers both before and after the attention modules to facilitate multi-scale feature extraction and summarization. For the CBAM module, which contains two sequential attention mechanisms, with the first attention mechanism deals with the filter domain information from the preceding convolutional layer, and the second attention mechanism processes 2D information from the spatial and rhythm domains. We then switch the two attention modules in CBAM to study the effect of the order by placing the first one to handle 2D spatial and rhythm information, and the second to process filter information. We also conduct a thorough investigation into the optimal number of convolutional layers surrounding the CBAM module by adding layers both before and after the CBAM. The classification results indicate that increasing the number of convolutional layers does not yield any significant benefits for final accuracy. We name the winning model as Sandwich enhanced CBAM (SCBAM). The flowchart of the SCBAM is indicated in <xref ref-type="fig" rid="F3">Figure 3</xref>. The CBAM module is treated as a component in <xref ref-type="fig" rid="F3">Figure 3a</xref> of the overall data flow, and the details of the CBAM are then elaborated in <xref ref-type="fig" rid="F3">Figure 3b</xref>. <xref ref-type="fig" rid="F4">Figure 4</xref> presents the structure and computational process within the network. It should be noted that the input of the SCBAM network is the features extracted from FBCSP.</p>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>The flowchart of the proposed network and CBAM module. <bold>(a)</bold> The flowchart of the proposed network. <bold>(b)</bold> The workflow of the CBAM. CBAM, Convolutional Block Attention Module; GAP, Global Average Pooling; MP, Max Pooling.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0003.tif">
<alt-text content-type="machine-generated">Flowchart-style diagram with two sections: on the left, the proposed network pipeline shows sequential modules including input, five-by-five Conv2D with one hundred twenty-eight filters, a CBAM module, three-by-three Conv2D with thirty-two filters, max pooling, flatten and dense layers, softmax, and output; on the right, a detailed CBAM module diagram splits into filters attention (via GAP, MP, MLP, and dense layers) and spatial and rhythm attention (using convolution and weighted feature generation).</alt-text>
</graphic>
</fig>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>The architecture of the proposed network.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0004.tif">
<alt-text content-type="machine-generated">Flowchart diagram illustrating the CBAM architecture for neural networks, depicting sequential modules including two convolutional layers, filter attention and spatial-rhythm attention modules, max pooling, global average pooling, multi-layer perceptron, and classifier, with arrows indicating feature extraction and refinement process using input of shape six hundred by twelve by six.</alt-text>
</graphic>
</fig>
<sec>
<label>2.3.1</label>
<title>Network architecture and parameters</title>
<p>This SCBAM effectively integrates convolutional layers with the attention mechanisms by structuring the model to begin and end with convolutional layers surrounding the CBAM. The first important decision is how to represent the input <italic>X</italic><sup><italic>j</italic></sup>, where <italic>j</italic> is the number of trials. To this end, we represent the input as <italic>X</italic><sup><italic>j</italic></sup> &#x02208; &#x1D540;<sup><italic>T</italic>&#x000B7;<italic>C</italic>&#x000B7;<italic>R</italic></sup> 3D-array with the number of time samples is <italic>T</italic>, the number of CSP components <italic>C</italic>, and the number of rhythms <italic>R</italic>. The first convolutional layer performs 2D convolution across the time domain and spatial dimensions (CSP components) using 128 filters with a kernel size of (5, 5) and a stride of (1, 1). Filters are applied to slide the data to capture temporal and spatial information. The output feature map will be fed into the CBAM to further extract fine-scale features.</p>
<p>An intermediate feature map (128, 12, 6) as input for the subsequent CBAM module. The first attention mechanism in CBAM generates a 1D attention map for the filter domain. The second attention mechanism generates a 2D attention map hyperplane expanded by the spatial (CSP components) and rhythm domains. The computation of the two attention modules is as follows:</p>
<disp-formula id="EQ2"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x02297;</mml:mo><mml:mi>M</mml:mi><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02033;</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x02297;</mml:mo><mml:msup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(2)</label></disp-formula>
<p>where <italic>M</italic> represents the input feature map, <italic>A</italic><sub><italic>f</italic></sub> is the attention map derived from the filter domain, generated by the preceding convolutional layer, <italic>A</italic><sub><italic>cr</italic></sub> is the attention feature map that is expanded by the CSP components and rhythm domain. The refined outputs are <italic>M</italic>&#x02032; and <italic>M</italic>&#x02033;, &#x02297; denotes the element-wise multiplication. Notably, the <italic>A</italic><sub><italic>cr</italic></sub> is a 2D attention map where rhythm domain information is first extracted.</p>
<p>The filter attention mechanism generates a 1D attention map through a global average pooling (GAP) and a max pooling (MP) layers, followed by a multi-layer perceptron (MLP). The GAP and MP operations generate average-pooled features and max-pooled features: <inline-formula><mml:math id="M4"><mml:mrow><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">gap</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M5"><mml:mrow><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">max</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> respectively. Both descriptors are then fed into a MLP module to produce an attention map <inline-formula><mml:math id="M6"><mml:mrow><mml:msub><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>. The MLP is composed of a hidden layer and a dense layer. To reduce parameter overhead, the hidden layer reduces the dimensionality by a reduction ratio <italic>r</italic> &#x0003D; 16, leading to a hidden size &#x0211D;<sup><italic>F</italic>/<italic>r</italic>&#x000D7;1&#x000D7;1</sup>. Specifically, the 128 input filters are first compressed to 8 and then reconstructed to 128 to generate filter-wise weights via a sigmoid function. The dense layer restores the original size 128 and applies a sigmoid activation function to generate attention weights. The filter attention is computed as:</p>
<disp-formula id="EQ3"><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mi>L</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>M</mml:mi><mml:mi>L</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">gap</mml:mtext></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">max</mml:mtext></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(3)</label></disp-formula>
<p>where the &#x003C3; is the sigmoid activation function, <italic>W</italic><sub>0</sub> and <italic>W</italic><sub>1</sub> are the weights of MLP layers.</p>
<p>In the second attention mechanism, we first apply GAP and MP oprations, then concatenate the results to create an efficient feature descriptor <inline-formula><mml:math id="M8"><mml:mrow><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">gap</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mi>C</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>. This attention module generates a 2D attention map via a Conv 2D layer, using 32 filters with a kernel size of (3, 3) and a stride of (1, 1). The second attention is computed as :</p>
<disp-formula id="EQ4"><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mi>A</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>;</mml:mo><mml:mi>M</mml:mi><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">gap</mml:mtext></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">max</mml:mtext></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(4)</label></disp-formula>
<p>where the &#x003C3; is the sigmoid activation function, <italic>f</italic><sup>3&#x000D7;3</sup> denotes the convolution operation.</p>
<p>After the attention modules, another convolutional layer is introduced for final feature map refinement. This layer utilizes 32 filters with a kernel size of (3, 3), a stride of (1, 1), and employs a ReLU activation function. This layer processes the features enhanced by the attention mechanisms from the filter, spatial, and rhythm domains. After that, a max-pooling layer with a pool size of (2, 2) and a stride of (2, 2) is applied to reduce computational complexity. With subsequent flattening and dense layers, this network effectively provides the discriminative feature maps for the final classification task. Besides, we use a dropout layer with a dropout rate of 0.3 to mitigate overfitting.</p>
</sec>
<sec>
<label>2.3.2</label>
<title>Ablation studies</title>
<p>We conduct ablation experiments on the SCBAM model to assess the isolated effect of each key component by systematically removing: the first convolutional layer, the two attention modules in CBAM, respectively, and the second convolutional layer. To assess the effects of the first convolutional layer, we evaluate the results of the CBAM module alongside the second convolutional layer (CBAMs). For the second convolutional layer, we examine the results by combining the first convolutional layer with the CBAM module (fCBAM). For attention mechanisms, we keep the sandwich shape generated by the two convolutional layers and combine it with the filter attention module (SA1) and spatial and rhythm attention module (SA2), respectively. To explore the cooperation of multiple components, we investigate the performance of the CBAM module to study the synergistic effects of the two attention mechanisms, as well as a network comprising only two convolutional layers (two convs).</p>
<p>To systematically evaluate the contribution of individual data processing components and ensure a rigorous performance analysis, we conducted two targeted ablation studies. Specifically, we assessed the impact of Filter Bank Common Spatial Pattern (FBCSP) and Common Average Reference (CAR) by removing them individually from the pipeline.</p>
</sec>
</sec>
<sec>
<label>2.4</label>
<title>Training and classification</title>
<p>In this section, we provide details of the training process and the classification strategy. Given that the number of trials is relatively small for effective tuning of a deep neural network, we chop the trials from the 3000 samples into 600 samples per trial, with a 50% (300) overlap. We concatenate the chopped trials from 10 runs, resulting in a total of 2,250 trials, 80% of which are used for training and 20% for testing. To guarantee the separation of the training and testing data, the data from the first 8 runs are used for training, while the last 2 runs are used for testing. Furthermore, we utilize 10% of the training data for validation purposes. To improve the model&#x00027;s generalization capabilities, we will shuffle the training data along with the corresponding training labels. The proposed network architecture is configured to run for 150 epochs, allowing sufficient learning from the training data. It employs a batch size of 32, which balances memory efficiency and stability in gradient updates. A learning rate of 0.001 is chosen to facilitate effective learning and prevent overshooting during optimization. Additionally, the weights and biases of the filters in each layer are trained jointly. To ensure robust evaluation, we determined trial-wise accuracy using a majority voting scheme. For each 5-s trial, we aggregated the predictions of its nine constituent windows, assigning the most frequent class label to the entire event.</p>
<p>We initially conduct binary classification to assess the model&#x00027;s discrimination ability corresponding to each movement pair, leading to 10 classifiers in total, and subsequently balance memory efficiency and stability in gradient updates. The five-class classification is implemented using a One vs. One (OvO) strategy. The final results of the five-class classification are obtained through a majority voting mechanism involving 10 binary classifiers. Each binary classifier cast a vote for its predicted class, and the class with the most votes was declared the winner. The class receiving the highest number of votes is selected as the final prediction. If two or more classes receive an equal number of votes, the tie is first addressed by conducting a sub-vote among the tied candidates. If the tie remains unresolved, the final decision is determined by aggregating the softmax class probabilities from the corresponding pairwise classifiers; specifically, the class with the highest total cumulative probability across all relevant binary comparisons is assigned as the final label. Each binary classification is trained for 150 epochs, resulting in a total of 1,500 epochs (150 &#x000D7; 10 classifiers).</p>
</sec>
<sec>
<label>2.5</label>
<title>Evaluation</title>
<p>We use the accuracy metric for classification results evaluation:</p>
<disp-formula id="EQ5"><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">Accuracy (Acc)</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(5)</label></disp-formula>
<p>where TP = True Positives, FP = False Positives, FN = False Negatives, and TN = True Negatives.</p>
<p>To provide a more detailed assessment of model performance, we also introduce the kappa statistic. Kappa contributes by quantifying the agreement between predicted and true classifications while accounting for chance agreement, thus offering insights into the model&#x00027;s reliability. Specifically, it highlights how well the model distinguishes between classes beyond random guessing, enabling a clearer evaluation of performance across both binary and multi-class scenarios.</p>
<disp-formula id="E6"><mml:math id="M11"><mml:mrow><mml:mi>&#x003BA;</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mrow></mml:math></disp-formula>
<p>where the <italic>P</italic><sub><italic>o</italic></sub> is the observed agreement, representing the proportion of instances where the raters agree. <italic>P</italic><sub><italic>e</italic></sub> is the expected possibility, indicating the proportion of instances where agreement will be expected by chance.</p>
<p>To evaluate classification significance against the empirical chance level, a non-parametric permutation test was performed with 5,000 iterations, where class labels were randomly shuffled at the trial level. The empirical <italic>p</italic>-value was calculated as the proportion of permutations where the shuffled accuracy was greater than or equal to the observed accuracy. To evaluate the performance difference between models, we conducted a paired t-test on the results obtained from the binary classification of 10 finger pairs at three significance levels (<italic>p</italic> &#x0003C; 0.05, <italic>p</italic> &#x0003C; 0.01, and <italic>p</italic> &#x0003C; 0.001). Furthermore, the stability of the results was quantified by calculating 95% confidence intervals via a non-parametric bootstrap procedure across subjects. For each model, subject-wise accuracies were resampled with replacement 1,000 times to estimate the distribution of the mean accuracy.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Results and analysis</title>
<p>In this section, we present a comprehensive analysis of our findings, structured into four key subsections. First, we detail the classification performance of our proposed model. Second, we discuss the results of our network design. Third, we show the classification results of ablation studies. Finally, we report on experiments conducted with several state-of-the-art models applied to our dataset. Each subsection provides in-depth insights into our methodology and results, reinforcing the effectiveness of our proposed approach.</p>
<sec>
<label>3.1</label>
<title>The results of the SCBAM</title>
<p>The binary and five-class classification results of individual fingers by the proposed SCBAM are presented in <xref ref-type="table" rid="T1">Table 1</xref>. The five subjects (S1 to S5) achieve accuracies of 77.38%, 79.40%, 79.69%, 78.57%, and 78.12% in binary classification, respectively, resulting in a mean accuracy of 78.63% (SD 1.56%; 95% CI [77.26, 80.00]%). Based on the results of binary classification, we observe that ten pairs of activities can be roughly categorized into three types: easy, moderate, and difficult.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>The classification results of the proposed network in accuracy.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>S1(%)</bold></th>
<th valign="top" align="center"><bold>S2(%)</bold></th>
<th valign="top" align="center"><bold>S3(%)</bold></th>
<th valign="top" align="center"><bold>S4(%)</bold></th>
<th valign="top" align="center"><bold>S5(%)</bold></th>
<th valign="top" align="center"><bold>mean (SD) (%)</bold></th>
<th valign="top" align="center"><bold><italic>P</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. Index</td>
<td valign="top" align="center">83.99</td>
<td valign="top" align="center">83.38</td>
<td valign="top" align="center">85.08</td>
<td valign="top" align="center">83.33</td>
<td valign="top" align="center">82.71</td>
<td valign="top" align="center">83.70 (0.90)</td>
<td valign="top" align="center"><bold>&#x0003C;0.001</bold></td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Middle</td>
<td valign="top" align="center">83.66</td>
<td valign="top" align="center">85.17</td>
<td valign="top" align="center">83.38</td>
<td valign="top" align="center">81.67</td>
<td valign="top" align="center">83.05</td>
<td valign="top" align="center">83.39 (1.26)</td>
<td valign="top" align="center"><bold>&#x0003C;0.001</bold></td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Ring</td>
<td valign="top" align="center">84.74</td>
<td valign="top" align="center">85.76</td>
<td valign="top" align="center">83.39</td>
<td valign="top" align="center">84.66</td>
<td valign="top" align="center">86.44</td>
<td valign="top" align="center">85.00 (1.16)</td>
<td valign="top" align="center"><bold>&#x0003C;0.001</bold></td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Little</td>
<td valign="top" align="center">81.35</td>
<td valign="top" align="center">86.10</td>
<td valign="top" align="center">84.13</td>
<td valign="top" align="center">82.06</td>
<td valign="top" align="center">81.03</td>
<td valign="top" align="center">82.93 (2.14)</td>
<td valign="top" align="center">0.0041</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Middle</td>
<td valign="top" align="center">85.33</td>
<td valign="top" align="center">81.69</td>
<td valign="top" align="center">83.33</td>
<td valign="top" align="center">86.33</td>
<td valign="top" align="center">83.00</td>
<td valign="top" align="center">83.94 (1.87)</td>
<td valign="top" align="center"><bold>&#x0003C;0.001</bold></td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Ring</td>
<td valign="top" align="center">81.35</td>
<td valign="top" align="center">84.66</td>
<td valign="top" align="center">84.67</td>
<td valign="top" align="center">81.67</td>
<td valign="top" align="center">82.67</td>
<td valign="top" align="center">83.00 (1.59)</td>
<td valign="top" align="center">0.0035</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Little</td>
<td valign="top" align="center">72.53</td>
<td valign="top" align="center">77.00</td>
<td valign="top" align="center">77.89</td>
<td valign="top" align="center">78.04</td>
<td valign="top" align="center">77.89</td>
<td valign="top" align="center">76.67 (2.35)</td>
<td valign="top" align="center">0.0192</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. Ring</td>
<td valign="top" align="center">73.62</td>
<td valign="top" align="center">74.23</td>
<td valign="top" align="center">72.11</td>
<td valign="top" align="center">73.09</td>
<td valign="top" align="center">72.63</td>
<td valign="top" align="center">73.14 (0.83)</td>
<td valign="top" align="center">0.0241</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. Little</td>
<td valign="top" align="center">63.09</td>
<td valign="top" align="center">69.99</td>
<td valign="top" align="center">68.42</td>
<td valign="top" align="center">68.76</td>
<td valign="top" align="center">64.74</td>
<td valign="top" align="center">67.00 (2.93)</td>
<td valign="top" align="center">0.0339</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. Little</td>
<td valign="top" align="center">64.12</td>
<td valign="top" align="center">66.00</td>
<td valign="top" align="center">64.58</td>
<td valign="top" align="center">66.06</td>
<td valign="top" align="center">67.03</td>
<td valign="top" align="center">65.96 (1.46)</td>
<td valign="top" align="center">0.0376</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center">77.38</td>
<td valign="top" align="center">79.40</td>
<td valign="top" align="center">79.69</td>
<td valign="top" align="center">78.57</td>
<td valign="top" align="center">78.12</td>
<td valign="top" align="center">78.63 (1.56)</td>
<td valign="top" align="center">0.0122</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>S1(%)</bold></td>
<td valign="top" align="center"><bold>S2(%)</bold></td>
<td valign="top" align="center"><bold>S3(%)</bold></td>
<td valign="top" align="center"><bold>S4(%)</bold></td>
<td valign="top" align="center"><bold>S5(%)</bold></td>
<td valign="top" align="center"><bold>mean (SD) (%)</bold></td>
<td valign="top" align="center"><italic>P</italic></td>
</tr>
<tr>
<td valign="top" align="left">Accuracy</td>
<td valign="top" align="center">60.73</td>
<td valign="top" align="center">62.36</td>
<td valign="top" align="center">60.00</td>
<td valign="top" align="center">61.81</td>
<td valign="top" align="center">60.72</td>
<td valign="top" align="center">61.12 (0.95)</td>
<td valign="top" align="center">0.0423</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<p>The six pairs from Thumb vs. Index to Index vs. Ring in <xref ref-type="table" rid="T1">Table 1</xref> are categorized as easy type, with accuracies above 80%. The Index vs. Little and Middle vs. Ring are recognized as moderate type, with accuracies between 70% and 80%. Meanwhile, the Middle vs. Little and Ring vs. Little are considered as difficult types, with accuracies ranging from 60% to 70%. The classification results indicate varying levels of performance among the subjects, with S2 achieving the best performance. Moreover, all classification pairs for all subjects performed significantly above chance (<italic>P</italic> &#x0003C; 0.05). The thumb vs. the Ring finger performs the best overall at 85.00 (1.16%) in mean accuracy across five subjects. The finger pairs Thumb vs. Index, Thumb vs. Middle, Thumb vs. Ring, and Index vs. Middle achieve a highly significant result at the level of <italic>P</italic> &#x0003C; 0.001. In binary classification, finger pairs yield distinct results based on the included fingers: pairs including the Thumb (Thumb vs. Index, Thumb vs. Middle, Thumb vs. Ring, Thumb vs. Little) achieve an average accuracy of 83.76% across five subjects. Additionally, pairs including the index, middle, ring, and little fingers achieve average accuracies of 81.83%, 76.87%, 76.78%, and 73.14%, respectively, across the same subjects.</p>
<p>The five subjects (S1 to S5) achieve accuracies of 60.73%, 62.36%, 60.00%, 61.81%, and 60.72% in five-class classification, respectively, resulting in a mean accuracy of 61.12% (SD 0.95%; 95% CI [60.29, 61.95]%). From the confusion matrix depicted in <xref ref-type="fig" rid="F5">Figure 5</xref>, the SCBAM achieves average accuracies of 66.67% for Thumb, 64.71% for Index, 62.50% for Middle, 62.50% for Ring, and 59.46% for Little fingers across five subjects. From this confusion matrix, it is clear that adjacent fingers tend to confuse with each other. For example, the thumb is misclassified as the index at a rate of 13.33%, and the ring finger is classified as the little finger at a rate of 15.62%. The confusion matrix reveals distinct patterns, indicating that the thumb, index, and middle fingers form one group with significant confusion tendencies, while the middle, ring, and little fingers constitute another one. The confusion clusters observed in <xref ref-type="fig" rid="F5">Figure 5</xref> (orange boxes) align with the somatotopic overlap in the primary motor cortex. Specifically, the confusion between the middle, ring, and little fingers reflects the known muscular and neural coupling termed enslavement, where these fingers share common multi-tendon muscle groups. This makes the neural signatures for these fingers more similar than those of the thumb and index finger. The higher classification accuracy for the Thumb (66.67%) and Index finger (64.71%) compared to the Little finger (59.46%) is consistent with the principle of cortical magnification. The thumb and index finger have larger, more specialized representations in the motor homunculus. Our SCBAM model effectively captures this increased neural real estate through its attention mechanism, resulting in more robust feature extraction for these digits.</p>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>The confusion matrix of the five-class classification (%). The yellow rectangle in the top-left corner highlights the confusion between adjacent fingers (the Thumb, Index, and Middle fingers), indicating that these gestures are often misclassified. The yellow rectangle in the bottom-right corner highlights that the gestures of the adjacent fingers (Middle, Ring, and Little fingers), indicating that these gestures are often misclassified.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0005.tif">
<alt-text content-type="machine-generated">Confusion matrix heatmap showing the classification accuracy for identifying thumb, index, middle, ring, and little fingers, with highest percentages along the diagonal. Darker green indicates higher values, representing correct predictions.</alt-text>
</graphic>
</fig>
<p>The <xref ref-type="fig" rid="F6">Figure 6</xref> depicts the loss and accuracy curves for binary and five-class classification tasks on training and validation sets. For binary classification, <xref ref-type="fig" rid="F6">Figure 6a</xref> shows that the training loss and validation loss decrease and converge to 0.15 and 0.25, respectively. The training accuracy rapidly increases during the first 40 epochs, ultimately converging at 93%, indicating robust learning from the training dataset. In parallel, the validation accuracy rises quickly over the first 45 epochs, stabilizing at 80%, which suggests a commendable ability to generalize, albeit with a notable gap compared to training performance. For five-class classification, <xref ref-type="fig" rid="F6">Figure 6b</xref> presents the training loss and validation loss exhibit a downward trend, but stabilize at around 0.20 for training loss and approximately 0.45 for validation loss. The training accuracy reaches about 78%, while validation accuracy is approximately 62%. Overall, the figure supports the assertion that 150 epochs are ample for ensuring sufficient learning of the model, as indicated by the convergence of both accuracy and loss metrics. Continuing training beyond this point is unlikely to enhance performance and may lead to overfitting.</p>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>Training and validation learning curves for binary and five-class classification across epochs. The learning curve was measured using two metrics: accuracy on the left y-axis and loss on the right y-axis. In the binary classification scenario <bold>(a)</bold>, the graph displays training loss (yellow), validation loss (blue), training accuracy (green), and validation accuracy (red), with horizontal dashed lines marking accuracy thresholds (80% and 93%) and loss thresholds (0.15 and 0.25). The five-class classification scenario <bold>(b)</bold> presents analogous metrics, illustrating training loss (yellow), validation loss (blue), training accuracy (green), and validation accuracy (red), accompanied by accuracy thresholds (60% and 80%) and loss thresholds (0.20 and 0.40) to evaluate model performance.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0006.tif">
<alt-text content-type="machine-generated">Two line charts compare training accuracy, validation accuracy, training loss, and validation loss over 150 epochs for binary classification (panel a, left) and averaged five-class classification (panel b, right). Both charts show training and validation accuracy increasing over epochs, while training and validation loss decrease. Panel b displays slightly lower accuracy and higher loss compared to panel a. Color key: green represents training loss, yellow validation loss, red training accuracy, and blue validation accuracy.</alt-text>
</graphic>
</fig>
<p>As our study includes both binary classification and five-class classification, we employ the Kappa statistic as a complementary measure to accuracy, providing a more nuanced classification result by accounting for chance agreement. The results are shown in <xref ref-type="table" rid="T2">Table 2</xref>. It achieves an average Kappa of 0.7898 for binary classification across 10 pairs in five subjects. For five-class classification, the average Kappa is 0.6130.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>The classification results of the proposed model in Kappa.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>S1</bold></th>
<th valign="top" align="center"><bold>S2</bold></th>
<th valign="top" align="center"><bold>S3</bold></th>
<th valign="top" align="center"><bold>S4</bold></th>
<th valign="top" align="center"><bold>S5</bold></th>
<th valign="top" align="center"><bold>Mean</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. index</td>
<td valign="top" align="center">0.8312</td>
<td valign="top" align="center">0.8298</td>
<td valign="top" align="center">0.8501</td>
<td valign="top" align="center">0.8250</td>
<td valign="top" align="center">0.8225</td>
<td valign="top" align="center">0.8217</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. middle</td>
<td valign="top" align="center">0.8350</td>
<td valign="top" align="center">0.8199</td>
<td valign="top" align="center">0.8317</td>
<td valign="top" align="center">0.8123</td>
<td valign="top" align="center">0.8205</td>
<td valign="top" align="center">0.8239</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. ring</td>
<td valign="top" align="center">0.8403</td>
<td valign="top" align="center">0.8612</td>
<td valign="top" align="center">0.8120</td>
<td valign="top" align="center">0.8340</td>
<td valign="top" align="center">0.8605</td>
<td valign="top" align="center">0.8416</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. little</td>
<td valign="top" align="center">0.8107</td>
<td valign="top" align="center">0.8172</td>
<td valign="top" align="center">0.8285</td>
<td valign="top" align="center">0.8265</td>
<td valign="top" align="center">0.8080</td>
<td valign="top" align="center">0.8182</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. middle</td>
<td valign="top" align="center">0.8476</td>
<td valign="top" align="center">0.8071</td>
<td valign="top" align="center">0.8275</td>
<td valign="top" align="center">0.8369</td>
<td valign="top" align="center">0.8064</td>
<td valign="top" align="center">0.8251</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. ring</td>
<td valign="top" align="center">0.7964</td>
<td valign="top" align="center">0.8372</td>
<td valign="top" align="center">0.8174</td>
<td valign="top" align="center">0.8035</td>
<td valign="top" align="center">0.8133</td>
<td valign="top" align="center">0.8136</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. little</td>
<td valign="top" align="center">0.7280</td>
<td valign="top" align="center">0.7535</td>
<td valign="top" align="center">0.7565</td>
<td valign="top" align="center">0.7400</td>
<td valign="top" align="center">0.7681</td>
<td valign="top" align="center">0.7512</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. ring</td>
<td valign="top" align="center">0.7196</td>
<td valign="top" align="center">0.7248</td>
<td valign="top" align="center">0.7035</td>
<td valign="top" align="center">0.7167</td>
<td valign="top" align="center">0.7162</td>
<td valign="top" align="center">0.7162</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. little</td>
<td valign="top" align="center">0.6390</td>
<td valign="top" align="center">0.6643</td>
<td valign="top" align="center">0.6725</td>
<td valign="top" align="center">0.6673</td>
<td valign="top" align="center">0.6280</td>
<td valign="top" align="center">0.6542</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. little</td>
<td valign="top" align="center">0.6375</td>
<td valign="top" align="center">0.6219</td>
<td valign="top" align="center">0.6405</td>
<td valign="top" align="center">0.6408</td>
<td valign="top" align="center">0.6790</td>
<td valign="top" align="center">0.6434</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average kappa</bold></td>
<td valign="top" align="center">0.7785</td>
<td valign="top" align="center">0.7937</td>
<td valign="top" align="center">0.7940</td>
<td valign="top" align="center">0.7903</td>
<td valign="top" align="center">0.7923</td>
<td valign="top" align="center">0.7898</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>S1</bold></td>
<td valign="top" align="center"><bold>S2</bold></td>
<td valign="top" align="center"><bold>S3</bold></td>
<td valign="top" align="center"><bold>S4</bold></td>
<td valign="top" align="center"><bold>S5</bold></td>
<td valign="top" align="center"><bold>Mean</bold></td>
</tr>
<tr>
<td valign="top" align="left">Accuracy</td>
<td valign="top" align="center">0.6174</td>
<td valign="top" align="center">0.6306</td>
<td valign="top" align="center">0.5987</td>
<td valign="top" align="center">0.6053</td>
<td valign="top" align="center">0.6129</td>
<td valign="top" align="center">0.6130</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
<sec>
<label>3.2</label>
<title>The results of network design</title>
<p>In our network design, we conduct a thorough study of the parameter settings, including the number of convolutional layers, the number of attention modules, and the order of attention module placement. By comparing the average accuracy across five subjects from different model candidates reported in Section 2, we identify the CBAM as the most suitable attention mechanism for our classification task. First, the self-attention mechanism between the initial and the final convolutional layer (CSA) model achieves an average accuracy of 63.98% for binary classification and 58.95% for five-class classification. Next, two convolutional layers integrating with a CBAM module in sandwich shape (SCBAM) model achieve an accuracy of 78.63% for binary classification and 61.12% for five-class classification. Integrating with a multi-head transformer (CTF) achieves an average accuracy of 64.89% for binary classification and 59.34% for five-class classification. Results from these three models indicate that embedding the CBAM module yields the best classification accuracy for our task. Switching the two attention mechanisms results in a decrease of 23.00% in binary classification accuracy and a drop of 24.40% in five-class classification accuracy. These results reveal that the order of attention modules is significant in network design. Attempts to add extra convolutional layers before and after the SCBAM model (C-SCBAM and SCBAM-C) all resulted in a drop in classification accuracy as reported in <xref ref-type="table" rid="T3">Table 3</xref>. These findings demonstrate that having one layer before the attention modules and one layer after is sufficient for our task. Excessive layers led to a model that was too complex for the limited training data, resulting in overfitting. Besides, the significance test of the binary classification results for all variants in the network design is presented in <xref ref-type="fig" rid="F7">Figure 7a</xref>, the ROC curves further illustrate the trade-offs between sensitivity and specificity for each model as depicted in <xref ref-type="fig" rid="F7">Figure 7b</xref>.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>The average accuracy of model candidates across five subjects.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>SCBAM (%)</bold></th>
<th valign="top" align="center"><bold>CTF (%)</bold></th>
<th valign="top" align="center"><bold>CSA (%)</bold></th>
<th valign="top" align="center"><bold>SA2A1 (%)</bold></th>
<th valign="top" align="center"><bold>C-SCBAM (%)</bold></th>
<th valign="top" align="center"><bold>SCBAM-C (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. index</td>
<td valign="top" align="center"><bold>83.70</bold></td>
<td valign="top" align="center">81.01</td>
<td valign="top" align="center">74.24</td>
<td valign="top" align="center">56.95</td>
<td valign="top" align="center">69.31</td>
<td valign="top" align="center">71.58</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. middle</td>
<td valign="top" align="center"><bold>83.39</bold></td>
<td valign="top" align="center">61.03</td>
<td valign="top" align="center">59.31</td>
<td valign="top" align="center">48.97</td>
<td valign="top" align="center">69.27</td>
<td valign="top" align="center">72.06</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. ring</td>
<td valign="top" align="center"><bold>85.00</bold></td>
<td valign="top" align="center">67.12</td>
<td valign="top" align="center">55.93</td>
<td valign="top" align="center">56.61</td>
<td valign="top" align="center">54.13</td>
<td valign="top" align="center">70.53</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. little</td>
<td valign="top" align="center">82.93</td>
<td valign="top" align="center"><bold>86.44</bold></td>
<td valign="top" align="center">82.37</td>
<td valign="top" align="center">54.58</td>
<td valign="top" align="center">52.54</td>
<td valign="top" align="center">69.47</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. middle</td>
<td valign="top" align="center"><bold>83.94</bold></td>
<td valign="top" align="center">48.47</td>
<td valign="top" align="center">54.24</td>
<td valign="top" align="center">55.25</td>
<td valign="top" align="center">63.05</td>
<td valign="top" align="center">67.89</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. ring</td>
<td valign="top" align="center"><bold>83.00</bold></td>
<td valign="top" align="center">73.33</td>
<td valign="top" align="center">60.00</td>
<td valign="top" align="center">54.00</td>
<td valign="top" align="center">52.20</td>
<td valign="top" align="center">65.15</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. little</td>
<td valign="top" align="center"><bold>76.67</bold></td>
<td valign="top" align="center">63.00</td>
<td valign="top" align="center">70.00</td>
<td valign="top" align="center">54.00</td>
<td valign="top" align="center">57.28</td>
<td valign="top" align="center">56.91</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. ring</td>
<td valign="top" align="center"><bold>73.14</bold></td>
<td valign="top" align="center">58.31</td>
<td valign="top" align="center">72.88</td>
<td valign="top" align="center">56.95</td>
<td valign="top" align="center">54.67</td>
<td valign="top" align="center">58.45</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. little</td>
<td valign="top" align="center"><bold>67.00</bold></td>
<td valign="top" align="center">50.50</td>
<td valign="top" align="center">50.85</td>
<td valign="top" align="center">56.61</td>
<td valign="top" align="center">49.83</td>
<td valign="top" align="center">53.09</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. little</td>
<td valign="top" align="center"><bold>65.96</bold></td>
<td valign="top" align="center">59.66</td>
<td valign="top" align="center">60.00</td>
<td valign="top" align="center">56.33</td>
<td valign="top" align="center">52.66</td>
<td valign="top" align="center">51.08</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center"><bold>78.63</bold></td>
<td valign="top" align="center">64.89</td>
<td valign="top" align="center">63.98</td>
<td valign="top" align="center">55.63</td>
<td valign="top" align="center">57.49</td>
<td valign="top" align="center">63.62</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>61.12</bold></td>
<td valign="top" align="center">59.34</td>
<td valign="top" align="center">58.95</td>
<td valign="top" align="center">36.72</td>
<td valign="top" align="center">48.58</td>
<td valign="top" align="center">48.95</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>CTF, the transformer module between the first and the second convolutional layers; CSA, the self-attention mechanism between the first and the second convolutional layers; SA2A1, the first convolutional layer &#x0002B; spatial and rhythm attention &#x0002B; filters attention &#x0002B; the second convolutional layer; C-SCBAM, add an extra convolutional layer before the SCBAM; SCBAM-C, add an extra convolutional layer after the SCBAM. The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<fig position="float" id="F7">
<label>Figure 7</label>
<caption><p>The classification and significance test results and ROC curves of SCBAM with the model variants in network design. <bold>(a)</bold> The classification and significantce test results of the SCBAM and its network design models. <bold>(b)</bold> The ROC curves of the network design. <sup>&#x0002A;&#x0002A;&#x0002A;</sup>represents the <italic>P</italic> &#x0003C; 0.001, <sup>&#x0002A;&#x0002A;</sup>denotes 0.001 &#x0003C; <italic>P</italic> &#x0003C; 0.01, <sup>&#x0002A;</sup> is 0.01 &#x0003C; <italic>P</italic> &#x0003C; 0.05. The red dashed line is the chance level.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0007.tif">
<alt-text content-type="machine-generated">Bar chart comparing classification accuracy percentages of six models with statistical significance annotations, and a ROC curve plot showing SCBAM outperforms others in AUC, with legend specifying AUC values for each model.</alt-text>
</graphic>
</fig>
</sec>
<sec>
<label>3.3</label>
<title>The ablation studies of the proposed network</title>
<p>The ablation studies of SCBAM performed across five subjects are reported in <xref ref-type="table" rid="T4">Table 4</xref>, involving a series of model variants. By removing two convolution layers, only the CBAM module achieves an average decrease in accuracy of 16.94% for the binary classification task and a decrease of 12.55% for the five-class classification task compared to the SCBAM. This significant decrease in accuracy (at a level of &#x0003C;0.001) illustrates that the two convolutional layers also play a crucial role in performance, although not as profound as that of the CBAM module. By removing the first convolutional layer, the CBAM model shows an average accuracy decrease of 15.12% for the binary classification task compared to the proposed model. Furthermore, for the five-class classification, removing the first convolutional layer results in an accuracy drop of 4.93%. The last convolutional layer also plays an important role, as its removal leads to a drop of 15.90% for the binary classification and a decrease of 1.12% for the five-class classification task. By conducting a t-test between the two model variants and the proposed SCBAM, the results indicate that both individual convolutional layers are significant contributors to the model&#x00027;s performance. By removing the attention modules, the two convolution layers (two convs) yield a value of 28.10% for five-class classification, whereas binary classification values range from 46.32% to 65.76%. This results in an average accuracy decrease of 21.62% for binary classification and 33.02% for five-class classification compared to the proposed model. These results demonstrate that the CBAM plays a crucial role in enhancing the model&#x00027;s performance. By removing the second attention module, the SA1 shows an average accuracy decline of 16.87% for the binary classification and a decrease of 9.57% for the five-class classification compared to SCBAM. By removing the first attention module, the SA2 model exhibits an average accuracy decrease of 9.77% for the binary classification and a decrease of 2.67% for the five-class classification compared to the proposed network. From the accuracy comparison, the first attention mechanism has the weakest effect on classification performance. This may be due to the fact that it operates in the filter domain generated by the preceding convolutional layer, meaning it refines the obtained features rather than extracting new features from the data. On the contrary, the second attention mechanism has a greater effect on the model than the other modules. This is because it is the key component for handling rhythm domain information, which is not addressed by any other modules.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>The results of ablation studies across five subjects.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>SCBAM</bold></th>
<th valign="top" align="center"><bold>fCBAM</bold></th>
<th valign="top" align="center"><bold>CBAMs</bold></th>
<th valign="top" align="center"><bold>SA1</bold></th>
<th valign="top" align="center"><bold>SA2</bold></th>
<th valign="top" align="center"><bold>CBAM</bold></th>
<th valign="top" align="center"><bold>two convs</bold></th>
<th valign="top" align="center"><bold>NFBCSP</bold></th>
<th valign="top" align="center"><bold>NCAR</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. index</td>
<td valign="top" align="center"><bold>83.70</bold></td>
<td valign="top" align="center">78.01</td>
<td valign="top" align="center">74.67</td>
<td valign="top" align="center">66.90</td>
<td valign="top" align="center">75.90</td>
<td valign="top" align="center">73.46</td>
<td valign="top" align="center">61.34</td>
<td valign="top" align="center">64.12</td>
<td valign="top" align="center">69.45</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. middle</td>
<td valign="top" align="center"><bold>83.39</bold></td>
<td valign="top" align="center">72.12</td>
<td valign="top" align="center">69.18</td>
<td valign="top" align="center">62.14</td>
<td valign="top" align="center">71.58</td>
<td valign="top" align="center">73.75</td>
<td valign="top" align="center">65.76</td>
<td valign="top" align="center">56.91</td>
<td valign="top" align="center">69.39</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. ring</td>
<td valign="top" align="center"><bold>85.00</bold></td>
<td valign="top" align="center">67.38</td>
<td valign="top" align="center">59.96</td>
<td valign="top" align="center">67.59</td>
<td valign="top" align="center">77.95</td>
<td valign="top" align="center">75.36</td>
<td valign="top" align="center">62.63</td>
<td valign="top" align="center">62.06</td>
<td valign="top" align="center">63.62</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. little</td>
<td valign="top" align="center"><bold>82.93</bold></td>
<td valign="top" align="center">66.34</td>
<td valign="top" align="center">72.03</td>
<td valign="top" align="center">65.52</td>
<td valign="top" align="center">72.82</td>
<td valign="top" align="center">63.36</td>
<td valign="top" align="center">59.47</td>
<td valign="top" align="center">55.15</td>
<td valign="top" align="center">64.74</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. middle</td>
<td valign="top" align="center"><bold>83.94</bold></td>
<td valign="top" align="center">67.84</td>
<td valign="top" align="center">64.31</td>
<td valign="top" align="center">66.00</td>
<td valign="top" align="center">75.38</td>
<td valign="top" align="center">63.57</td>
<td valign="top" align="center">56.84</td>
<td valign="top" align="center">53.09</td>
<td valign="top" align="center">68.42</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. ring</td>
<td valign="top" align="center"><bold>83.00</bold></td>
<td valign="top" align="center">62.83</td>
<td valign="top" align="center">70.00</td>
<td valign="top" align="center">64.83</td>
<td valign="top" align="center">74.00</td>
<td valign="top" align="center">64.59</td>
<td valign="top" align="center">59.47</td>
<td valign="top" align="center">52.06</td>
<td valign="top" align="center">61.72</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. little</td>
<td valign="top" align="center"><bold>76.67</bold></td>
<td valign="top" align="center">60.00</td>
<td valign="top" align="center">58.73</td>
<td valign="top" align="center">59.47</td>
<td valign="top" align="center">69.33</td>
<td valign="top" align="center">54.00</td>
<td valign="top" align="center">58.95</td>
<td valign="top" align="center">51.72</td>
<td valign="top" align="center">60.03</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. ring</td>
<td valign="top" align="center"><bold>73.14</bold></td>
<td valign="top" align="center">48.13</td>
<td valign="top" align="center">62.33</td>
<td valign="top" align="center">56.19</td>
<td valign="top" align="center">65.33</td>
<td valign="top" align="center">44.59</td>
<td valign="top" align="center">48.89</td>
<td valign="top" align="center">53.09</td>
<td valign="top" align="center">52.56</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. little</td>
<td valign="top" align="center"><bold>67.00</bold></td>
<td valign="top" align="center">50.00</td>
<td valign="top" align="center">53.40</td>
<td valign="top" align="center">56.91</td>
<td valign="top" align="center">54.74</td>
<td valign="top" align="center">52.39</td>
<td valign="top" align="center">46.32</td>
<td valign="top" align="center">51.03</td>
<td valign="top" align="center">51.43</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. little</td>
<td valign="top" align="center"><bold>65.96</bold></td>
<td valign="top" align="center">54.64</td>
<td valign="top" align="center">50.52</td>
<td valign="top" align="center">52.06</td>
<td valign="top" align="center">51.55</td>
<td valign="top" align="center">51.83</td>
<td valign="top" align="center">50.50</td>
<td valign="top" align="center">48.28</td>
<td valign="top" align="center">51.72</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center"><bold>78.63</bold></td>
<td valign="top" align="center">62.73</td>
<td valign="top" align="center">63.51</td>
<td valign="top" align="center">61.76</td>
<td valign="top" align="center">68.86</td>
<td valign="top" align="center">61.69</td>
<td valign="top" align="center">57.01</td>
<td valign="top" align="center">54.75</td>
<td valign="top" align="center">61.31</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>61.12</bold></td>
<td valign="top" align="center">60.00</td>
<td valign="top" align="center">56.19</td>
<td valign="top" align="center">51.55</td>
<td valign="top" align="center">58.45</td>
<td valign="top" align="center">48.57</td>
<td valign="top" align="center">28.10</td>
<td valign="top" align="center">49.09</td>
<td valign="top" align="center">53.18</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>fCBAM, the first convolutional layer with the CBAM module; CBAMs, the CBAM module alongside the second convolutional layer; SA1, two convolution layers (sandwich) and filters attention; SA2, two convolution layers (sandwich) and spatial and rhythm attention; CBAM, only Convolutional Block Attention Module; two convs, only two convolution layers; NFBCSP, the proposed method without FBCSP; NCAR, the proposed method without CAR. The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<p>Besides, ablation experiments are conducted to verify the effectiveness of FBCSP and CAR. Compared to the proposed method, if we remove the FBCSP from the data processing pipeline, the accuracy drops by 23.88% in binary classification and 12.03% in five-class classification. If we remove the CAR, the accuracy declines by 17.32% in binary classification and by 7.94% in five-class classification, respectively. These results confirm that both steps are indispensable for achieving optimal classification performance. It is worth noting that, through the significance test of the binary classification results for all variants within the proposed SCBAM data pipeline, as presented in <xref ref-type="fig" rid="F8">Figure 8a</xref>, none of the modules or processing steps can be removed, as they all play significant roles in the final performance at a level of &#x0003C;0.01. Besides, the ROC curves further validate the trade-offs between sensitivity and specificity for each model as illustrated in <xref ref-type="fig" rid="F8">Figure 8b</xref>.</p>
<fig position="float" id="F8">
<label>Figure 8</label>
<caption><p>The classification and significance test results and ROC curves of SCBAM with the ablation variants in the pipeline. <bold>(a)</bold> The classification and significance test results of the SCBAM and ablation variants. <bold>(b)</bold> The ROC curves of the SCBAM and ablation variants. &#x0002A;&#x0002A;&#x0002A; represents the <italic>P</italic> &#x0003C; 0.001, &#x0002A;&#x0002A; denotes 0.001 &#x0003C; <italic>P</italic> &#x0003C; 0.01, &#x0002A; is 0.01 &#x0003C; <italic>P</italic> &#x0003C; 0.05. The red dashed line is the chance level.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0008.tif">
<alt-text content-type="machine-generated">Bar chart compares accuracy percentages for eight models, with SCBAM achieving the highest accuracy and several statistically significant differences marked by asterisks. ROC curves on a line graph display true positive versus false positive rates for each model, with SCBAM having the highest area under the curve value of zero point eight zero zero.</alt-text>
</graphic>
</fig>
</sec>
<sec>
<label>3.4</label>
<title>The results of comparasion experiments</title>
<p>We compare the classification performance of the proposed network with state-of-the-art models, including EEGNet, EEGLearn, Shallow ConvNets, Deep ConvNets, EEG Conformer, and Msvt Net with FBCSP features and raw data independently. The classification results with FBCSP features are presented in <xref ref-type="table" rid="T5">Table 5</xref>. For binary classification, SCBAM exceeds the performance of the EEGNet, EEG Learn, Shallow ConvNets, Deep ConvNets, EEG Conformer, and Msvt Net by 25.69%, 12.93%, 14.37%, 13.31%, 2.59%, and 13.25%, respectively. A paired <italic>t</italic>-test is conducted to assess the significance of the differences in binary classification accuracy between the proposed model and six benchmark models, as demonstrated in <xref ref-type="fig" rid="F9">Figure 9a</xref>. Specipically, the SCBAM surpasses the EEG Conformer and Deep ConvNets at the 0.05 significance level. The SCBAM significantly outperforms EEGLearn, Shallow ConvNets, and Msvt Net at the 0.01 level. SCBAM exhibits a highly significant performance advantage over EEGNet at the 0.001 level. For five-class classification, SCBAM achieves an average classification accuracy of 61.12%, while the EEGNet, EEGLearn, Shallow ConvNets, Deep ConvNets, EEG Conformer, and Msvt Net exhibit accuracies in the range from 45% to 58%. The ROC curves further illustrate the trade-offs between sensitivity and specificity for each model as presented in <xref ref-type="fig" rid="F9">Figure 9c</xref>. The accuracy and loss learning curves during training are depicted in <xref ref-type="fig" rid="F10">Figures 10a</xref>, <xref ref-type="fig" rid="F10">c</xref>. The accuracy of SCBAM rises sharply until it reaches 94.00% around the 45th epoch, surpassing benchmark models in both convergence speed and final accuracy. Subsequently, the accuracy of EEG Conformer and Deep ConvNets stabilizes around 90% at 120th epoch. While Shallow ConvNets, EEG Learn, and Msvt Net acheive approximately 80% accuracy at the 120th epoch. Among the six benchmark models, Shallow ConvNets converge most rapidly, reaching stability by the 80th epoch. The loss of SCBAM exhibits a sharp decline during the first 40 epochs, reaching a low value of approximately 0.25 by the 80th epoch. The loss of EEG Conformer and Deep ConvNets reaches a turning point around the 60th epoch, achieving a low value of approximately 0.4 by the 120th epoch. Shallow ConvNets converges more rapidly reaching its stability loss value of approximately 0.5 by the 100th epoch. The results indicate that SCBAM significantly surpasses the performance and efficiency of these six models.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Classification results of proposed model against state-of-the-art techniques with FBCSP features.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>SCBAM</bold></th>
<th valign="top" align="center"><bold>EEGNet (<xref ref-type="bibr" rid="B16">Lawhern et al., 2018</xref>)</bold></th>
<th valign="top" align="center"><bold>EEG learn (<xref ref-type="bibr" rid="B2">Bashivan et al., 2016</xref>)</bold></th>
<th valign="top" align="center"><bold>Shallow convs (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>)</bold></th>
<th valign="top" align="center"><bold>Deep convs (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>)</bold></th>
<th valign="top" align="center"><bold>EEG conformer (<xref ref-type="bibr" rid="B48">Song et al., 2023</xref>)</bold></th>
<th valign="top" align="center"><bold>Msvt net (<xref ref-type="bibr" rid="B20">Liu et al., 2024</xref>)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. index</td>
<td valign="top" align="center"><bold>83.70</bold></td>
<td valign="top" align="center">50.52</td>
<td valign="top" align="center">72.58</td>
<td valign="top" align="center">72.11</td>
<td valign="top" align="center">81.90</td>
<td valign="top" align="center">82.76</td>
<td valign="top" align="center">79.44</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. middle</td>
<td valign="top" align="center"><bold>83.39</bold></td>
<td valign="top" align="center">54.14</td>
<td valign="top" align="center">73.68</td>
<td valign="top" align="center">71.58</td>
<td valign="top" align="center">82.76</td>
<td valign="top" align="center">81.90</td>
<td valign="top" align="center">77.54</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. ring</td>
<td valign="top" align="center"><bold>85.00</bold></td>
<td valign="top" align="center">52.54</td>
<td valign="top" align="center">71.58</td>
<td valign="top" align="center">73.62</td>
<td valign="top" align="center">78.42</td>
<td valign="top" align="center">83.05</td>
<td valign="top" align="center">76.58</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. little</td>
<td valign="top" align="center"><bold>82.93</bold></td>
<td valign="top" align="center">51.19</td>
<td valign="top" align="center">70.53</td>
<td valign="top" align="center">72.60</td>
<td valign="top" align="center">79.90</td>
<td valign="top" align="center">82.11</td>
<td valign="top" align="center">73.09</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. middle</td>
<td valign="top" align="center"><bold>83.94</bold></td>
<td valign="top" align="center">62.03</td>
<td valign="top" align="center">69.47</td>
<td valign="top" align="center">68.95</td>
<td valign="top" align="center">76.58</td>
<td valign="top" align="center">81.69</td>
<td valign="top" align="center">72.00</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. ring</td>
<td valign="top" align="center"><bold>83.00</bold></td>
<td valign="top" align="center">48.67</td>
<td valign="top" align="center">72.06</td>
<td valign="top" align="center">67.37</td>
<td valign="top" align="center">73.62</td>
<td valign="top" align="center">77.89</td>
<td valign="top" align="center">64.12</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. little</td>
<td valign="top" align="center"><bold>76.67</bold></td>
<td valign="top" align="center">58.33</td>
<td valign="top" align="center">64.12</td>
<td valign="top" align="center">58.95</td>
<td valign="top" align="center">70.53</td>
<td valign="top" align="center">73.68</td>
<td valign="top" align="center">63.12</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. ring</td>
<td valign="top" align="center"><bold>73.14</bold></td>
<td valign="top" align="center">51.86</td>
<td valign="top" align="center">63.09</td>
<td valign="top" align="center">59.47</td>
<td valign="top" align="center">69.47</td>
<td valign="top" align="center">71.58</td>
<td valign="top" align="center">62.58</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. little</td>
<td valign="top" align="center"><bold>67.00</bold></td>
<td valign="top" align="center">47.80</td>
<td valign="top" align="center">49.48</td>
<td valign="top" align="center">50.00</td>
<td valign="top" align="center">58.11</td>
<td valign="top" align="center">63.12</td>
<td valign="top" align="center">52.14</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. little</td>
<td valign="top" align="center"><bold>65.96</bold></td>
<td valign="top" align="center">52.33</td>
<td valign="top" align="center">51.38</td>
<td valign="top" align="center">48.95</td>
<td valign="top" align="center">57.38</td>
<td valign="top" align="center">62.58</td>
<td valign="top" align="center">53.21</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center"><bold>78.63</bold></td>
<td valign="top" align="center">52.94</td>
<td valign="top" align="center">65.70</td>
<td valign="top" align="center">64.26</td>
<td valign="top" align="center">72.87</td>
<td valign="top" align="center">76.04</td>
<td valign="top" align="center">65.38</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>61.12</bold></td>
<td valign="top" align="center">45.68</td>
<td valign="top" align="center">47.52</td>
<td valign="top" align="center">53.26</td>
<td valign="top" align="center">56.73</td>
<td valign="top" align="center">58.55</td>
<td valign="top" align="center">54.00</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<fig position="float" id="F9">
<label>Figure 9</label>
<caption><p>The classification and significance test results and ROC curves of the SCBAM and the other six benchmark models with FBCSP features and raw data. <bold>(a)</bold> The classification results and significance test of the SCBAM and the other six models with FBCSP features. <bold>(b)</bold> The classification results and significance test of the SCBAM and the other six models with raw data. <bold>(c)</bold> The ROC curves of the SCBAM and the other six models with FBCSP features. <bold>(d)</bold> The ROC curves of the SCBAM and the other six models with raw data. &#x0002A;&#x0002A;&#x0002A; represents the <italic>P</italic> &#x0003C; 0.001, &#x0002A;&#x0002A; denotes 0.001 &#x0003C; <italic>P</italic> &#x0003C; 0.01, &#x0002A; is 0.01 &#x0003C; <italic>P</italic> &#x0003C; 0.05. The red dashed line is the chance level.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0009.tif">
<alt-text content-type="machine-generated">Bar charts compare accuracy percentages of EEG classification models, highlighting statistically significant differences, while ROC curves below assess model performance using FBCSP features and raw data, showing SCBAM outperforming others in both evaluations.</alt-text>
</graphic>
</fig>
<fig position="float" id="F10">
<label>Figure 10</label>
<caption><p>The learning curves of the SCBAM and the other six benchmark models with FBCSP features and raw data. <bold>(a)</bold> Accuracy curves with FBCSP features. <bold>(b)</bold> Accuracy curves with raw data. <bold>(c)</bold> Loss curves with FBCSP features. <bold>(d)</bold> Loss curves with raw data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0010.tif">
<alt-text content-type="machine-generated">Four line charts compare deep learning models using FBCSP features and raw data on accuracy and loss across epochs. SCBAM consistently achieves the highest accuracy and lowest loss, outperforming other methods.</alt-text>
</graphic>
</fig>
<p>The classification results with raw data are summarized in <xref ref-type="table" rid="T6">Table 6</xref>, SCBAM consistently outperforms benchmark models across both binary and five-class classification tasks. In binary classification, SCBAM achieves accuracy improvements of 20.06%, 14.42%, 9.95%, 5.03%, 0.99%, and 7.47% over EEGNet, EEGLearn, Shallow ConvNets, Deep ConvNets, EEG Conformer, and Msvt Net, respectively. In the five-class task, SCBAM surpasses these benchmark models by 18.03%, 16.50%, 13.13%, 9.12%, 4.61%, and 10.99%, respectively. A paired <italic>t</italic>-test is conducted to assess the significance of the differences in binary classification accuracy between the proposed model and six benchmark models, the results are demonstrated in <xref ref-type="fig" rid="F9">Figure 9b</xref>. SCBAM achieves highly significant improvements over EEGNet and EEGLearn (<italic>P</italic> &#x0003C; 0.001), and significantly outperforms Shallow ConvNets and Msvt Net (<italic>P</italic> &#x0003C; 0.01) and Deep ConvNets at the 0.05 significance level. It is worth noting that the EEG Conformer demonstrates strong classification performance that is statistically comparable to our proposed SCBAM (<italic>P</italic> &#x0003D; 0.43). The ROC curves further illustrate the trade-offs between sensitivity and specificity for each model as presented in <xref ref-type="fig" rid="F9">Figure 9d</xref>. Besides, the accuracy and loss learning curves during training are depicted in <xref ref-type="fig" rid="F10">Figures 10b</xref>, <xref ref-type="fig" rid="F10">d</xref>. SCBAM stabilizes at an accuracy of 81% around 130th epoch and a steady loss value of 0.4 around 150th epoch, outperforming other models.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Classification results of proposed model against state-of-the-art techniques with raw data.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>SCBAM</bold></th>
<th valign="top" align="center"><bold>EEGNet (<xref ref-type="bibr" rid="B16">Lawhern et al., 2018</xref>)</bold></th>
<th valign="top" align="center"><bold>EEG learn (<xref ref-type="bibr" rid="B2">Bashivan et al., 2016</xref>)</bold></th>
<th valign="top" align="center"><bold>Shallow convs (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>)</bold></th>
<th valign="top" align="center"><bold>Deep convs (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>)</bold></th>
<th valign="top" align="center"><bold>EEG conformer (<xref ref-type="bibr" rid="B48">Song et al., 2023</xref>)</bold></th>
<th valign="top" align="center"><bold>Msvt net (<xref ref-type="bibr" rid="B20">Liu et al., 2024</xref>)</bold></th>
</tr>
<tr>
<th valign="top" align="left"><bold>Thumb vs. index</bold></th>
<th valign="top" align="center"><bold>78.42</bold></th>
<th valign="top" align="center"><bold>51.05</bold></th>
<th valign="top" align="center"><bold>60.51</bold></th>
<th valign="top" align="center"><bold>69.61</bold></th>
<th valign="top" align="center"><bold>72.53</bold></th>
<th valign="top" align="center"><bold>76.58</bold></th>
<th valign="top" align="center"><bold>70.00</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. middle</td>
<td valign="top" align="center">73.09</td>
<td valign="top" align="center">48.89</td>
<td valign="top" align="center">54.13</td>
<td valign="top" align="center">70.32</td>
<td valign="top" align="center">71.03</td>
<td valign="top" align="center"><bold>74.55</bold></td>
<td valign="top" align="center">68.58</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. ring</td>
<td valign="top" align="center"><bold>79.44</bold></td>
<td valign="top" align="center">52.50</td>
<td valign="top" align="center">61.18</td>
<td valign="top" align="center">70.78</td>
<td valign="top" align="center">70.52</td>
<td valign="top" align="center">74.84</td>
<td valign="top" align="center">69.47</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. little</td>
<td valign="top" align="center"><bold>74.12</bold></td>
<td valign="top" align="center">51.05</td>
<td valign="top" align="center">52.54</td>
<td valign="top" align="center">69.58</td>
<td valign="top" align="center">72.06</td>
<td valign="top" align="center">71.48</td>
<td valign="top" align="center">66.74</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. middle</td>
<td valign="top" align="center"><bold>72.26</bold></td>
<td valign="top" align="center">56.84</td>
<td valign="top" align="center">62.03</td>
<td valign="top" align="center">63.09</td>
<td valign="top" align="center">72.00</td>
<td valign="top" align="center">72.09</td>
<td valign="top" align="center">68.05</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. ring</td>
<td valign="top" align="center"><bold>70.54</bold></td>
<td valign="top" align="center">50.50</td>
<td valign="top" align="center">58.66</td>
<td valign="top" align="center">61.38</td>
<td valign="top" align="center">69.38</td>
<td valign="top" align="center">69.19</td>
<td valign="top" align="center">65.43</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. little</td>
<td valign="top" align="center"><bold>71.05</bold></td>
<td valign="top" align="center">50.00</td>
<td valign="top" align="center">58.33</td>
<td valign="top" align="center">52.50</td>
<td valign="top" align="center">61.03</td>
<td valign="top" align="center">66.74</td>
<td valign="top" align="center">61.05</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. ring</td>
<td valign="top" align="center"><bold>68.75</bold></td>
<td valign="top" align="center">53.68</td>
<td valign="top" align="center">52.33</td>
<td valign="top" align="center">56.84</td>
<td valign="top" align="center">61.04</td>
<td valign="top" align="center">65.43</td>
<td valign="top" align="center">57.37</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. little</td>
<td valign="top" align="center"><bold>59.38</bold></td>
<td valign="top" align="center">46.32</td>
<td valign="top" align="center">51.86</td>
<td valign="top" align="center">51.05</td>
<td valign="top" align="center">51.38</td>
<td valign="top" align="center">58.47</td>
<td valign="top" align="center">51.53</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. little</td>
<td valign="top" align="center"><bold>56.42</bold></td>
<td valign="top" align="center">42.11</td>
<td valign="top" align="center">47.79</td>
<td valign="top" align="center">48.89</td>
<td valign="top" align="center">52.20</td>
<td valign="top" align="center">54.26</td>
<td valign="top" align="center">50.53</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center"><bold>70.35</bold></td>
<td valign="top" align="center">50.29</td>
<td valign="top" align="center">55.93</td>
<td valign="top" align="center">60.40</td>
<td valign="top" align="center">65.32</td>
<td valign="top" align="center">69.36</td>
<td valign="top" align="center">62.88</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>56.75</bold></td>
<td valign="top" align="center">38.72</td>
<td valign="top" align="center">40.25</td>
<td valign="top" align="center">43.62</td>
<td valign="top" align="center">47.63</td>
<td valign="top" align="center">52.14</td>
<td valign="top" align="center">45.76</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<p>In summary, the classification results of the SCBAM and benchmark models with FBCSP features significantly outperformed those obtained from raw data. This underscores the importance of integrating well-established features with deep learning networks, as it facilitates a balance between performance and efficiency. Consequently, we investigated the efficiency analysis of the seven models with FBCSP features. As described in Section 2.4, we conduct 2,250 trials from 10 experimental runs, where 1,800 trials from the first 8 runs are used for training, and 450 trials from the last 2 runs are used for testing. The assignment of training and test data is consistent for both binary and five-class classification scenarios. It is worth noting that the FBCSP step takes 314 s (around 5 min). The time consumption of seven models are indicated in <xref ref-type="table" rid="T7">Table 7</xref>. For binary classification, SCBAM exhibits the most efficient performance with a per-epoch time of 1 s, resulting in a total training time of 25 min and a testing time of just 10 s. Shallow ConvNets, though slightly slower with a per-epoch time of 1.8 s, maintain a competitive training time of 45 min and a testing time of 14 s. Deep ConvNets demonstrate a more pronounced increase in computational demands, featuring a per-epoch time of 3 s and a total training time of 74.5 min, with a testing time of 28 s. EEG Conformer performs a per-epoch time of 3.3 s and a total training time of 80 min, with a testing time of 24 s. EEGNet, Msvt Net, and EEGLearn exhibit even greater time consumption. For five-class classification, SCBAM exhibits the most efficient performance with a per epoch time of 1.5 s, resulting in a total training time of 37.5 min ((150 &#x000D7; 10 epochs &#x000D7; 1.5 s) / 60) and a testing time of just 17 s, which is lower compared to the other six models. This analysis underscores SCBAM&#x00027;s efficiency, making it the optimal choice for scenarios requiring quick training and inference times. It is worth noting that both the SCBAM and Shallow ConvNets have two convolutional layers. Deep ConvNets, EEG Conformer, EEGNet, EEGLearn, and Msvt Net rely on a larger number of convolutional layers and deeper networks for feature extraction, require significantly longer training times.</p>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>The runtime of all models with FBCSP features.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Binary classification</bold></th>
<th valign="top" align="center"><bold>Per epoch time (s)</bold></th>
<th valign="top" align="center"><bold>Training time (min)</bold></th>
<th valign="top" align="center"><bold>Testing time (s)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">SCBAM</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">25</td>
<td valign="top" align="center">10</td>
</tr>
<tr>
<td valign="top" align="left">Shallow ConvNets</td>
<td valign="top" align="center">1.8</td>
<td valign="top" align="center">45</td>
<td valign="top" align="center">14</td>
</tr>
<tr>
<td valign="top" align="left">Deep ConvNets</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">74.5</td>
<td valign="top" align="center">28</td>
</tr>
<tr>
<td valign="top" align="left">EEG Conformer</td>
<td valign="top" align="center">3.2</td>
<td valign="top" align="center">80</td>
<td valign="top" align="center">24</td>
</tr>
<tr>
<td valign="top" align="left">EEGNet</td>
<td valign="top" align="center">3.5</td>
<td valign="top" align="center">87.5</td>
<td valign="top" align="center">25</td>
</tr>
<tr>
<td valign="top" align="left">Msvt Net</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">30</td>
</tr>
<tr>
<td valign="top" align="left">EEGLearn</td>
<td valign="top" align="center">4.2</td>
<td valign="top" align="center">105</td>
<td valign="top" align="center">34</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Five-class classification</bold></td>
<td valign="top" align="center"><bold>Per epoch time (s)</bold></td>
<td valign="top" align="center"><bold>Training time (min)</bold></td>
<td valign="top" align="center"><bold>Testing time (s)</bold></td>
</tr>
<tr>
<td valign="top" align="left">SCBAM</td>
<td valign="top" align="center">1.5</td>
<td valign="top" align="center">37.5 (10 classifiers)</td>
<td valign="top" align="center">17</td>
</tr>
<tr>
<td valign="top" align="left">Shallow ConvNets</td>
<td valign="top" align="center">2.4</td>
<td valign="top" align="center">60 (10 classifiers)</td>
<td valign="top" align="center">19</td>
</tr>
<tr>
<td valign="top" align="left">Deep ConvNets</td>
<td valign="top" align="center">3.3</td>
<td valign="top" align="center">82.5 (10 classifiers)</td>
<td valign="top" align="center">33</td>
</tr>
<tr>
<td valign="top" align="left">EEG Conformer</td>
<td valign="top" align="center">4.3</td>
<td valign="top" align="center">107.5 (10 classifiers)</td>
<td valign="top" align="center">29</td>
</tr>
<tr>
<td valign="top" align="left">EEGNet</td>
<td valign="top" align="center">4.5</td>
<td valign="top" align="center">112.5 (10 classifiers)</td>
<td valign="top" align="center">42</td>
</tr>
<tr>
<td valign="top" align="left">Msvt Net</td>
<td valign="top" align="center">4.8</td>
<td valign="top" align="center">120 (10 classifiers)</td>
<td valign="top" align="center">50</td>
</tr>
<tr>
<td valign="top" align="left">EEGLearn</td>
<td valign="top" align="center">5</td>
<td valign="top" align="center">125 (10 classifiers)</td>
<td valign="top" align="center">56</td>
</tr></tbody>
</table>
</table-wrap>
<p>Comparing the time consumed in the binary and five-class classification scenarios of binary and five-class classification, it is evident that the five-class classification always consumes longer time in each epoch and total training and test time. This is due to the one-vs-one strategy employed in the five-class classification. In this scenario, we utilize 10 binary classifiers for each trial, leading to a cumulative processing time that may suggest a 10-fold increase relative to a single binary classification. In contrast, the binary classification task only requires distinguishing between two classes, which necessitates computational resources for just one binary classifier. Additionally, in the binary classification framework, each movement is represented in four pairs. For instance, the thumb is compared against the index, middle, ring, and little fingers. Consequently, each trial is read and classified four times overall. In summary, in five-class classification, each data trial is read once but classified 10 times using binary classifiers. In contrast, in the binary classification scenario, the data for each trial is read four times and classified through four different binary classifiers. These factors contribute to the observation that while the training time for five-class classification is longer than that for binary classification, it is not as extensive as 10 times the duration of binary classification.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>In this section, we summarize the main contributions of this study and discuss our results in relation to those reported in the current literature. In addition to analyzing the individual finger classification outcomes using the uHD EEG dataset, we also consider results from EEG and ECoG data. By contextualizing our findings within the broader landscape of brain signals, we aim to highlight the advantages and limitations of uHD EEG, paving the way for future research in this area. At the same time, we examine the decoding algorithms utilized in this field over the past two decades and identify emerging trends, such as the integration of deep learning modules. Our study suggests a promising direction aimed at enhancing classification accuracy while managing computational complexity, ensuring that computation times remain sufficiently short for online experiments and real-time applications.</p>
<sec>
<label>4.1</label>
<title>Individual finger classification with electrical brain signals</title>
<p>First, we compare our results with those from recent literature that utilized the same uHD EEG dataset, including one study that employed an SVM classifier (<xref ref-type="bibr" rid="B18">Lee et al., 2022</xref>) and another that used an MLP (<xref ref-type="bibr" rid="B33">Nemes et al., 2024</xref>). The results are summarized in the <xref ref-type="table" rid="T8">Table 8</xref>, indicating varying levels of accuracy across different finger pairs. The SCBAM method consistently outperforms these two approaches, achieving the highest average accuracy of 78.63% across all finger pairs and subjects. In contrast, the SVM and MLP methods recorded average accuracies of 64.8% and 65.68%, respectively. As depicted in <xref ref-type="fig" rid="F11">Figure 11</xref>, statistical analysis using the t-test reveals that SCBAM significantly improved performance compared to both SVM and MLP at a significance level of 0.001. These findings highlight the accuracy of the SCBAM achieved for individual finger binary classification, suggesting it as a superior choice for this task. One thing to note is that in both studies, the ring finger is reported to have greater accuracy and has been regarded as easier to distinguish from the others. This is attributed to the fact that the four movement pairs, including the ring finger, rank among the top four in accuracy of all ten finger pairs in their results, which was also inferred in Table 1 in (<xref ref-type="bibr" rid="B18">Lee et al. 2022</xref>) and Table 3 in (<xref ref-type="bibr" rid="B33">Nemes et al. 2024</xref>). However, our results indicate that the ring finger does not stand out significantly. Instead, it is more reasonable to assert that the thumb and index fingers demonstrate greater classification accuracy in our findings. The binary classification between five individual fingers with EEG data has also been reported in (<xref ref-type="bibr" rid="B19">Liao et al. 2014</xref>). Their findings indicate an average accuracy of 77.11%. In their results, even though the Ring vs. Little pair achieves the highest accuracy, the Ring vs. Thumb, Ring vs. Index, and Ring vs. Middle pairs are ranked 5th, 7th, and 9th in all ten pairs, respectively [as indicated in Table 2 in <xref ref-type="bibr" rid="B19">Liao et al. (2014)</xref>]. The overall classification performance of the ring finger pairs appears to be less pronounced.</p>
<table-wrap position="float" id="T8">
<label>Table 8</label>
<caption><p>uHD EEG classification results from current literature and proposed model in accuracy (mean (SD) %).</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Data</bold></th>
<th valign="top" align="center" colspan="3">uHD EEG</th>
<th valign="top" align="center"><bold>EEG</bold></th>
</tr>
<tr>
<th valign="top" align="left"><bold>Activity</bold></th>
<th valign="top" align="center"><bold>SCBAM</bold></th>
<th valign="top" align="center"><bold>MLP (<xref ref-type="bibr" rid="B33">Nemes et al., 2024</xref>)</bold></th>
<th valign="top" align="center"><bold>SVM (<xref ref-type="bibr" rid="B18">Lee et al., 2022</xref>)</bold></th>
<th valign="top" align="center"><bold>SVM (<xref ref-type="bibr" rid="B19">Liao et al., 2014</xref>)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Thumb vs. Index</td>
<td valign="top" align="center"><bold>83.70 (0.90)</bold></td>
<td valign="top" align="center">64.00 (6.50)</td>
<td valign="top" align="center">63.30 (3.50)</td>
<td valign="top" align="center">71.30 (13.30)</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Middle</td>
<td valign="top" align="center"><bold>83.39 (1.26)</bold></td>
<td valign="top" align="center">65.00 (5.50)</td>
<td valign="top" align="center">64.00 (4.10)</td>
<td valign="top" align="center">74.30 (1.14)</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Ring</td>
<td valign="top" align="center"><bold>85.00 (1.16)</bold></td>
<td valign="top" align="center">72.40 (10.10)</td>
<td valign="top" align="center">69.80 (7.10)</td>
<td valign="top" align="center">77.60 (12.60)</td>
</tr>
<tr>
<td valign="top" align="left">Thumb vs. Little</td>
<td valign="top" align="center"><bold>82.93 (2.14)</bold></td>
<td valign="top" align="center">62.40 (2.90)</td>
<td valign="top" align="center">61.80 (3.40)</td>
<td valign="top" align="center">78.20 (13.50)</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Middle</td>
<td valign="top" align="center"><bold>83.94 (1.87)</bold></td>
<td valign="top" align="center">62.80 (4.10)</td>
<td valign="top" align="center">60.20 (4.40)</td>
<td valign="top" align="center">69.20 (13.60)</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Ring</td>
<td valign="top" align="center"><bold>83.00 (1.59)</bold></td>
<td valign="top" align="center">65.60 (13.40)</td>
<td valign="top" align="center">68.60 (9.70)</td>
<td valign="top" align="center">73.50 (8.80)</td>
</tr>
<tr>
<td valign="top" align="left">Index vs. Little</td>
<td valign="top" align="center"><bold>76.67</bold> (2.35)</td>
<td valign="top" align="center">60.60 <bold>(1.50)</bold></td>
<td valign="top" align="center">61.80 (5.00)</td>
<td valign="top" align="center">79.60 (13.30)</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. Ring</td>
<td valign="top" align="center"><bold>73.14 (0.83)</bold></td>
<td valign="top" align="center">69.60 (7.50)</td>
<td valign="top" align="center">70.60 (9.40)</td>
<td valign="top" align="center">71.20 (12.20)</td>
</tr>
<tr>
<td valign="top" align="left">Middle vs. Little</td>
<td valign="top" align="center"><bold>67.00</bold> (2.93)</td>
<td valign="top" align="center">62.80 <bold>(2.40)</bold></td>
<td valign="top" align="center">62.60 (2.60)</td>
<td valign="top" align="center">77.90 (14.10)</td>
</tr>
<tr>
<td valign="top" align="left">Ring vs. Little</td>
<td valign="top" align="center">65.96 <bold>(1.40)</bold></td>
<td valign="top" align="center"><bold>70.40</bold> (9.30)</td>
<td valign="top" align="center">65.20 (3.90)</td>
<td valign="top" align="center">80.20 (10.80)</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Average accuracy</bold></td>
<td valign="top" align="center"><bold>78.63 (1.56)</bold></td>
<td valign="top" align="center">65.52 (6.32)</td>
<td valign="top" align="center">64.80 (5.31)</td>
<td valign="top" align="center">75.30 (11.33)</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The bold values represent the best classification performance for each respective action.</p>
</table-wrap-foot>
</table-wrap>
<fig position="float" id="F11">
<label>Figure 11</label>
<caption><p>The classification and significance test results of the SCBAM and literature models on uHD EEG. 1. <sup>&#x0002A;&#x0002A;&#x0002A;</sup> represents the <italic>P</italic> &#x0003C; 0.001, <sup>&#x0002A;&#x0002A;</sup> denotes 0.001 &#x0003C; <italic>P</italic> &#x0003C; 0.01, <sup>&#x0002A;</sup> is 0.01 &#x0003C; <italic>P</italic> &#x0003C; 0.05. 2. The red dashed line is the chance level.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-20-1751058-g0011.tif">
<alt-text content-type="machine-generated">Bar chart comparing the accuracy percentages of three models: SCBAM (approximately 78 percent), MLP (about 66 percent), and SVM (about 65 percent). SCBAM has the highest accuracy with error bars, and statistical significance between SCBAM and both other models is indicated by three asterisks above the bars. A red dashed line marks the 50 percent accuracy level.</alt-text>
</graphic>
</fig>
<p>A few studies have reported the five-class classification of individual fingers, particularly using ECoG signals. From the study of Kubanek (<xref ref-type="bibr" rid="B13">Kub&#x000E1;nek et al., 2009</xref>), the five-class individual finger classification achieved 90.6%, 89.9%, 76.4%, 68.1%, and 76.7% for their five subjects, respectively, with an average of 80.3 (8.66)% across subjects. <xref ref-type="bibr" rid="B4">Chestek et al. (2013)</xref> employed a Naive Bayes classifier, which achieved classification accuracies of 74%, 98%, and 66% for each subject, with an average of 79.33 (13.49)% across three subjects. Their study emphasizes the effect of gamma band features. Both of these studies utilize macro ECoG signals, which have a spatial resolution of around 10 mm. For <xref ref-type="bibr" rid="B7">Hotson et al. (2016)</xref>, who employed a micro ECoG with an electrode resolution of 2.3 mm, they achieved a five-class classification accuracy of 76% using Linear Discriminant Analysis (LDA) with only one subject. However, there are currently no literature reports on five-class classification results of individual finger flexion and extension tasks using uHD EEG data. In our study, we achieved five-class classification accuracies of 60.73%, 62.36%, 60.00%, 61.81%, and 60.72% for our five subjects, respectively. This results in an average accuracy of 61.12% with a small standard deviation of 0.95% across subjects. Compared to ECoG, our classification accuracy is lower; however, it is important to highlight that our results exhibit greater stability across subjects. The high variance observed in ECoG decoding is likely due to the cantotomy surgery, which is primarily designed to address clinical situations. This often results in electrode placements varying significantly among individuals, leading to configurations that may be either far from or close to the motor cortex. In contrast, uHD EEG collection does not depend on clinical surgery for electrode implantation, allowing for convenient electrode placements and calibrations. This property contributes to the stable results observed in our study. <xref ref-type="bibr" rid="B55">Xiao and Ding (2013)</xref> also demonstrated that with EEG signals, five-class classification reached an accuracy of 45.2% using three principal component features. The significant difference in accuracy between uHD EEG and EEG data underscores that spatial resolution is a crucial factor for individual finger decoding. What&#x00027;s more, from our confusion matrix and the confusion matrix derived from ECoG results, it is evident that adjacent fingers tend to confuse each other, while fingers that are farther apart are easier to distinguish. This aligns with the common understanding: fingers that are distanced from each other have less overlapping musculature and are less likely to interfere with one another during movement. Additionally, their corresponding cortical areas are situated farther apart, making them comparatively easier to differentiate in classification tasks. The observed grouping of the Thumb, Index, and Middle fingers in <xref ref-type="fig" rid="F5">Figure 5</xref> reflects both the common innervation by the median nerve and the functional synergies required for precision handling (<xref ref-type="bibr" rid="B41">Schieber and Santello, 2004</xref>). Conversely, the confusion between the Ring and Little fingers aligns with ulnar nerve distribution and the enslavement phenomenon, where peripheral mechanical and neural coupling limit independent digit control (<xref ref-type="bibr" rid="B6">H&#x000E4;ger-Ross and Schieber, 2000</xref>). Our model&#x00027;s higher accuracy for the thumb and index is further supported by the principle of cortical magnification, as these digits possess significantly larger and more complex representations in the motor cortex (<xref ref-type="bibr" rid="B40">Schieber, 2001</xref>). This finding is crucial for understanding the underlying neural mechanisms of finger movements.</p>
</sec>
<sec>
<label>4.2</label>
<title>Deep neural network application</title>
<p>In our study, we design a Sandwich-enhanced CBAM (SCBAM) network that integrates convolutional layers with a CBAM module. The two attention mechanisms in CBAM process information from different domains for uHD EEG signals. The filter attention highlights time-frequency patterns, prioritizing the most informative frequency bands relevant to finger movements. Spatial attention focuses on the most active electrodes, enhancing sensitivity to subtle differences in motor patterns. The multidimensional feature extraction capability of the attentional mechanisms in CBAM aligns well with the multi-domain features of uHD EEG signals, leading to significant improvements in both performance and efficiency. This is also validated by the results in Section 3.4, the proposed SCBAM model shows significant improvements in performance over several state-of-the-art models, including EEGNet, EEGLearn, Shallow ConvNets, Deep ConvNets, EEG Conformer, and Msvt Net in both binary and five-class classification tasks, <xref ref-type="table" rid="T5">Tables 5</xref>, <xref ref-type="table" rid="T6">6</xref>. It also requires much less inference time compared to these models in <xref ref-type="table" rid="T7">Table 7</xref>. This is due to the attention mechanisms in SCBAM, which are also validated in ablation studies.</p>
<p>The classification results indicate that while convolutional layers are highly effective at extracting multi-level features for movement discrimination, they eventually reach a performance saturation point. Simply increasing the depth of the model, as seen in architectures like EEGNet (<xref ref-type="bibr" rid="B16">Lawhern et al., 2018</xref>) or EEGlearn (<xref ref-type="bibr" rid="B2">Bashivan et al., 2016</xref>), does not necessarily lead to breakthroughs, creating a performance upper bound for both binary and five-class classification. The introduction of attention mechanisms, such as those in the EEG Conformer (<xref ref-type="bibr" rid="B48">Song et al., 2023</xref>), represents a strategic shift in model design. These mechanisms operate along the preceding convolutional filter domain to enhance both classification performance and computational efficiency. It is worth noting that the EEG Conformer demonstrates a strong classification ability that is statistically comparable to the proposed SCBAM (<italic>P</italic> &#x0003D; 0.43) when fed with raw data. However, the proposed SCBAM further refines this approach by embedding dual attention modules to manage innovative discriminative features from the filters, spatial, and rhythm domains. This dual attention module plays a crucial role in automatically extracting discriminative features from the input data in the proposed SCBAM network. The input of the proposed SCBAM network is the features extracted by the Filter Bank Common Spatial Pattern (FBCSP). These features have been thoroughly studied for decades (<xref ref-type="bibr" rid="B23">Lotte and Guan, 2011</xref>), and it is widely accepted that they are key features in neural decoding. We drew the inspiration of combining traditional features with neural networks from the Shallow ConvNets proposed by Tonnio Ball (<xref ref-type="bibr" rid="B42">Schirrmeister et al., 2017</xref>). This two-step design integrates established signal processing principles with the data-driven capabilities of deep neural networks, resulting in a lightweight architecture that achieves statistically significant results in individual finger classification. While the five-class accuracy of 61.3% notably above the 20% chance level, highlights the inherent difficulty of decoding fine-grained finger movements, our analysis of the error structure reveals that misclassifications are primarily concentrated among anatomically adjacent fingers. This performance provides a robust foundation for future dexterous BCI research, where these specific decoding patterns can be further refined through larger cohorts and online experiment.</p>
</sec>
<sec>
<label>4.3</label>
<title>Future study</title>
<p>While the current study utilizes an open-access dataset (<xref ref-type="bibr" rid="B18">Lee et al., 2022</xref>) consisting of five subjects. Our laboratory is currently initiating experiments using identical equipment and experimental paradigms with more than 20 subjects. This expansion will facilitate a more comprehensive investigation of inter-subject variability. Furthermore, we plan to bridge the gap between offline classification and practical utility by transitioning to closed-loop online experiments. Future research will also prioritize optimizing real-time performance metrics, such as Information Transfer Rate (ITR) and system latency, to meet the benchmarks necessary for intuitive and responsive BCI control.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="s5">
<label>5</label>
<title>Conclusion</title>
<p>In this study, we propose a Sandwich enhanced CBAM (SCBAM) network for decoding individual finger movements utilizing the uHD EEG dataset collected by g. Pangolin system, achieving an average accuracy of 78.63 (1.56)% in binary classification and 61.12 (0.95)% in five-class classification. The results significantly outperform the existing methods of SVM and MLP for binary finger classification with the same dataset. Comparative experiments are conducted between our proposed model and several state-of-the-art neural networks, including EEGNet, EEGLearn, Shallow ConvNets, and Deep ConvNet. These results indicate that the hybrid model incorporating attention mechanisms with convolutional layers is an effective way to improve performance for the task of individual finger classification.</p>
<p>From the network design process and the ablation experiments, the proposed SCBAM model can break the accuracy saturation of networks that consist only of convolutional layers. Combined with a well-established FBCSP in the BCI field, SCBAM takes full advantage of the brain rhythms and spatial patterns. This combination allows the network to maintain a small number of parameters while ensuring satisfactory classification performance.</p>
<p>Compared to the EEG signals for the individual finger task, uHD EEG shows significant improvements in both binary and five-class classification, attributed to its high spatial resolution. In comparison with ECoG signals, uHD EEG exhibits comparatively lower accuracy for five-class classification. However, uHD EEG demonstrates surprising stability, with the five subjects achieving a small standard deviation of 0.95%. This stability is attributed to the flexibility of electrode placement and the convenience of calibration during experiments, both of which are crucial for real-life scenarios. In conclusion, this study proposes a lightweight yet highly effective network for individual finger classification, emphasizing the potential of uHD EEG for real-time dexterous tasks in the field of brain-computer interfaces (BCIs).</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://osf.io/4dwjt?view_only=d23acfd50655427fbaae381a17cbfbcc">https://osf.io/4dwjt?view_only=d23acfd50655427fbaae381a17cbfbcc</ext-link>. Repository: OSF (Open Science Framework).</p>
</sec>
<sec sec-type="ethics-statement" id="s7">
<title>Ethics statement</title>
<p>The studies involving humans were approved by Institutional Review Board of Korea Advanced Institute of Science and Technology (KH2018-127). The studies were conducted in accordance with the local legislation and institutional requirements. The human samples used in this study were acquired from a uHD EEG dataset published on OSF (<ext-link ext-link-type="uri" xlink:href="https://osf.io/4dwjt/?view_only=d23acfd50655427fbaae381a17cbfbcc">https://osf.io/4dwjt/?view_only=d23acfd50655427fbaae381a17cbfbcc</ext-link>). This dataset was originally collected by a previous study of Hyemin S. Lee. Their work has received ethical approval from Institutional Review Board of Korea Advanced Institute of Science and Technology (KH2018-127) and the participants have provided their written informed consent to participate. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants&#x00027; legal guardians/next of kin in accordance with the national legislation and institutional requirements.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>XZ: Writing &#x02013; original draft, Conceptualization, Writing &#x02013; review &#x00026; editing, Validation, Investigation, Software, Methodology. YZ: Conceptualization, Supervision, Writing &#x02013; review &#x00026; editing, Methodology, Writing &#x02013; original draft, Formal analysis. HP: Validation, Supervision, Writing &#x02013; review &#x00026; editing. TD: Funding acquisition, Writing &#x02013; original draft, Resources, Writing &#x02013; review &#x00026; editing, Methodology.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ang</surname> <given-names>K. K.</given-names></name> <name><surname>Chin</surname> <given-names>Z. Y.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <name><surname>Guan</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Filter bank common spatial pattern (FBCSP) in brain-computer interface,&#x0201D;</article-title> in. <italic>IEEE World Congress on Computational Intelligence</italic> (Hong Kong: IEEE).</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bashivan</surname> <given-names>P.</given-names></name> <name><surname>Rish</surname> <given-names>I.</given-names></name> <name><surname>Yeasin</surname> <given-names>M.</given-names></name> <name><surname>Codella</surname> <given-names>N. C. F.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Learning representations from EEG with deep recurrent-convolutional neural networks,&#x0201D;</article-title> in <source>International Conference on Learning Representations</source>.</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chatrian</surname> <given-names>G. E.</given-names></name> <name><surname>Lettich</surname> <given-names>E.</given-names></name> <name><surname>Nelson</surname> <given-names>P. L.</given-names></name></person-group> (<year>1985</year>). <article-title>Ten percent electrode system for topographic studies of spontaneous and evoked EEG activities</article-title>. <source>Am. J. Electroneurodiagnostic Technol</source>. <volume>25</volume>, <fpage>1</fpage>&#x02013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1080/00029238.1985.11080163</pub-id></mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chestek</surname> <given-names>C. A.</given-names></name> <name><surname>Gilja</surname> <given-names>V.</given-names></name> <name><surname>Blabe</surname> <given-names>C. H.</given-names></name> <name><surname>Foster</surname> <given-names>B. L.</given-names></name> <name><surname>Shenoy</surname> <given-names>K. V.</given-names></name> <name><surname>Parvizi</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Hand posture classification using electrocorticography signals in the gamma band over human sensorimotor brain areas</article-title>. <source>J. Neural Eng</source>. <volume>10</volume>:<fpage>026002</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2560/10/2/026002</pub-id><pub-id pub-id-type="pmid">23369953</pub-id></mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Eldele</surname> <given-names>E.</given-names></name> <name><surname>Chen</surname> <given-names>Z.</given-names></name> <name><surname>Liu</surname> <given-names>C.</given-names></name> <name><surname>Wu</surname> <given-names>M.</given-names></name> <name><surname>Kwoh</surname> <given-names>C. K.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>An attention-based deep learning approach for sleep stage classification with single-channel EEG</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitat. Eng</source>.<volume>29</volume>, <fpage>809</fpage>&#x02013;<lpage>818</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TNSRE.2021.3076234</pub-id><pub-id pub-id-type="pmid">33909566</pub-id></mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>H&#x000E4;ger-Ross</surname> <given-names>C.</given-names></name> <name><surname>Schieber</surname> <given-names>M. H.</given-names></name></person-group> (<year>2000</year>). <article-title>Quantifying the independence of human finger movements: comparisons of digits, hands, and movement frequencies</article-title>. <source>J. Neurosci</source>. <volume>20</volume>, <fpage>8542</fpage>&#x02013;<lpage>8550</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.20-22-08542.2000</pub-id><pub-id pub-id-type="pmid">11069962</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hotson</surname> <given-names>G.</given-names></name> <name><surname>McMullen</surname> <given-names>D. P.</given-names></name> <name><surname>Fifer</surname> <given-names>M. S.</given-names></name> <name><surname>Johannes</surname> <given-names>M. S.</given-names></name> <name><surname>Katyal</surname> <given-names>K. D.</given-names></name> <name><surname>Para</surname> <given-names>M. P.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Individual finger control of a modular prosthetic limb using high-density electrocorticography in a human subject</article-title>. <source>J. Neural Eng</source>. <volume>13</volume>:<fpage>026017</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2560/13/2/026017</pub-id><pub-id pub-id-type="pmid">26863276</pub-id></mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jasper</surname> <given-names>H. H.</given-names></name></person-group> (<year>1958</year>). <article-title>The ten-twenty electrode system of the international federation. Electroencephalogr</article-title>. <source>Clin. Neurophysiol</source>. <volume>10</volume>, <fpage>371</fpage>&#x02013;<lpage>375</lpage></mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>T.</given-names></name> <name><surname>Liu</surname> <given-names>S.</given-names></name> <name><surname>Pellizzer</surname> <given-names>G.</given-names></name> <name><surname>Aydoseli</surname> <given-names>A.</given-names></name> <name><surname>Karamursel</surname> <given-names>S.</given-names></name> <name><surname>Sabanci</surname> <given-names>P. A.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Characterization of hand clenching in human sensorimotor cortex using high- and ultra-high frequency band modulations of electrocorticogram</article-title>. <source>Front. Neurosci</source>. <volume>12</volume>:<fpage>110</fpage> doi: <pub-id pub-id-type="doi">10.3389/fnins.2018.00110</pub-id><pub-id pub-id-type="pmid">29535603</pub-id></mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kern</surname> <given-names>M.</given-names></name> <name><surname>Bert</surname> <given-names>S.</given-names></name> <name><surname>Glanz</surname> <given-names>O.</given-names></name> <name><surname>Schulze-Bonhage</surname> <given-names>A.</given-names></name> <name><surname>Ball</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Human motor cortex relies on sparse and action-specific activation during laughing, smiling, and speech production</article-title>. <source>Commun. Biol</source>. <volume>2</volume>:<fpage>339</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s42003-019-0593-1</pub-id></mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Krishnan</surname> <given-names>A.</given-names></name> <name><surname>Kumar</surname> <given-names>R.</given-names></name> <name><surname>Etienne</surname> <given-names>A.</given-names></name> <name><surname>Robinson</surname> <given-names>A. K.</given-names></name> <name><surname>Kelly</surname> <given-names>S. K.</given-names></name> <name><surname>Behrmann</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>&#x0201C;Challenges and opportunities in instrumentation and use of high-density EEG for underserved regions,&#x0201D;</article-title> in <source>Innovation and Interdisciplinary Solutions for Underserved Areas</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>).</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Krusienski</surname> <given-names>D. J.</given-names></name> <name><surname>Sellers</surname> <given-names>E. W.</given-names></name> <name><surname>Cabestaing</surname> <given-names>F.</given-names></name> <name><surname>Bayoudh</surname> <given-names>S.</given-names></name> <name><surname>McFarland</surname> <given-names>D. J.</given-names></name> <name><surname>Vaughan</surname> <given-names>T. M.</given-names></name> <etal/></person-group>. (<year>2006</year>). <article-title>A comparison of classification techniques for the P300 speller</article-title>. <source>J. Neural Eng</source>. <volume>3</volume>:<fpage>299</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2560/3/4/007</pub-id><pub-id pub-id-type="pmid">17124334</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kub&#x000E1;nek</surname> <given-names>J.</given-names></name> <name><surname>Miller</surname> <given-names>K. J.</given-names></name> <name><surname>Ojemann</surname> <given-names>J. G.</given-names></name> <name><surname>Wolpaw</surname> <given-names>J. R.</given-names></name> <name><surname>Schalk</surname> <given-names>G.</given-names></name></person-group> (<year>2009</year>). <article-title>Decoding flexion of individual fingers using electrocorticographic signals in humans</article-title>. <source>J. Neural Eng</source>. <volume>6</volume>:<fpage>066001</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2560/6/6/066001</pub-id><pub-id pub-id-type="pmid">19794237</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kumar</surname> <given-names>R.</given-names></name> <name><surname>Venkatesh</surname> <given-names>P.</given-names></name> <name><surname>Sun</surname> <given-names>R.</given-names></name> <name><surname>Mohankumar</surname> <given-names>G.</given-names></name> <name><surname>Antony</surname> <given-names>A.</given-names></name> <name><surname>Richardson</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>F123. Ultra-high-density scalp EEG outperforms localized invasive ECoG grids in inferring depth of seizure foci</article-title>. <source>Clin. Neurophysiol</source>., 129:e113. doi: <pub-id pub-id-type="doi">10.1016/j.clinph.2018.04.286</pub-id></mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lange</surname> <given-names>G.</given-names></name> <name><surname>Low</surname> <given-names>C. Y.</given-names></name> <name><surname>Johar</surname> <given-names>K.</given-names></name> <name><surname>Hanapiah</surname> <given-names>F. A.</given-names></name> <name><surname>Kamaruzaman</surname> <given-names>F.</given-names></name></person-group> (<year>2016</year>). <article-title>Classification of electroencephalogram data from hand grasp and release movements for BCI controlled prosthesis</article-title>. <source>Procedia Technol</source>. <volume>25</volume>, <fpage>1170</fpage>&#x02013;<lpage>1174</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.protcy.2016.08.048</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lawhern</surname> <given-names>V. J.</given-names></name> <name><surname>Solon</surname> <given-names>A. J.</given-names></name> <name><surname>Waytowich</surname> <given-names>N. R.</given-names></name> <name><surname>Gordon</surname> <given-names>S. M.</given-names></name> <name><surname>Hung</surname> <given-names>C. P.</given-names></name> <name><surname>Lance</surname> <given-names>B. J.</given-names></name></person-group> (<year>2018</year>). <article-title>EEGNet: a compact convolutional neural network for EEG-based brain-computer interfaces</article-title>. <source>J. Neural Eng</source>. <volume>15</volume>:<fpage>056013</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2552/aace8c</pub-id><pub-id pub-id-type="pmid">29932424</pub-id></mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>LeCun</surname> <given-names>Y.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>1998</year>). <article-title>&#x0201C;Convolutional networks for images, speech, and time series,&#x0201D;</article-title> in <source>The Handbook of Brain Theory and Neural Networks</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>H. S.</given-names></name> <name><surname>Schreiner</surname> <given-names>L.</given-names></name> <name><surname>Jo</surname> <given-names>S.-H.</given-names></name> <name><surname>Sieghartsleitner</surname> <given-names>S.</given-names></name> <name><surname>Jordan</surname> <given-names>M. I.</given-names></name> <name><surname>Pretl</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Individual finger movement decoding using a novel ultra-high-density electroencephalography-based brain-computer interface system</article-title>. <source>Front. Neurosci</source>. <volume>16</volume>:<fpage>1009878</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnins.2022.1009878</pub-id><pub-id pub-id-type="pmid">36340769</pub-id></mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liao</surname> <given-names>K.</given-names></name> <name><surname>Xiao</surname> <given-names>R.</given-names></name> <name><surname>Gonzalez</surname> <given-names>J.</given-names></name> <name><surname>Ding</surname> <given-names>L.</given-names></name></person-group> (<year>2014</year>). <article-title>Decoding individual finger movements from one hand using human EEG signals</article-title>. <source>PLOS ONE</source> <volume>9</volume>:<fpage>e85192</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0085192</pub-id><pub-id pub-id-type="pmid">24416360</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>K.</given-names></name> <name><surname>Yang</surname> <given-names>T.</given-names></name> <name><surname>Yu</surname> <given-names>W.</given-names></name> <name><surname>Yi</surname> <given-names>W.</given-names></name> <name><surname>Yu</surname> <given-names>H.</given-names></name> <name><surname>Wnag</surname> <given-names>G.</given-names></name></person-group> (<year>2024</year>). <article-title>MSVTNet: multi-scale vision transformer neural network for EEG-based motor imagery decoding</article-title>. <source>IEEE J. Biomed. Health Informat</source>. <volume>28</volume>, <fpage>7126</fpage>&#x02013;<lpage>7137</lpage>. doi: <pub-id pub-id-type="doi">10.1109/JBHI.2024.3450753</pub-id><pub-id pub-id-type="pmid">39190517</pub-id></mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lotte</surname> <given-names>F.</given-names></name> <name><surname>Bougrain</surname> <given-names>L.</given-names></name> <name><surname>Cichocki</surname> <given-names>A.</given-names></name> <name><surname>Clerc</surname> <given-names>M.</given-names></name> <name><surname>Congedo</surname> <given-names>M.</given-names></name> <name><surname>Rakotomamonjy</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>A review of classification algorithms for EEG-based brain-computer interfaces: a 10-year update</article-title>. <source>J. Neural Eng</source>. <volume>15</volume>:<fpage>AAB2F</fpage>2. doi: <pub-id pub-id-type="doi">10.1088/1741-2552/aab2f2</pub-id></mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lotte</surname> <given-names>F.</given-names></name> <name><surname>Congedo</surname> <given-names>M.</given-names></name> <name><surname>L&#x000E9;cuyer</surname> <given-names>A.</given-names></name> <name><surname>Lamarche</surname> <given-names>F.</given-names></name> <name><surname>Arnaldi</surname> <given-names>B.</given-names></name></person-group> (<year>2007</year>). <article-title>A review of classification algorithms for EEG-based brain-computer interfaces</article-title>. <source>J. Neural Eng</source>. <volume>4</volume>:<fpage>R01</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1741-2560/4/2/R01</pub-id><pub-id pub-id-type="pmid">17409472</pub-id></mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lotte</surname> <given-names>F.</given-names></name> <name><surname>Guan</surname> <given-names>C.</given-names></name></person-group> (<year>2011</year>). <article-title>Regularizing common spatial patterns to improve BCI designs: unified theory and new algorithms</article-title>. <source>IEEE Trans. Biomed. Eng</source>. <volume>58</volume>, <fpage>2270</fpage>&#x02013;<lpage>2280</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TBME.2010.2082539</pub-id><pub-id pub-id-type="pmid">20889426</pub-id></mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Pei</surname> <given-names>Z.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>J.</given-names></name></person-group> (<year>2024</year>). <article-title>Attention-based convolutional neural network with multi-modal temporal information fusion for motor imagery EEG decoding</article-title>. <source>Comp. Biol. Med</source>. <volume>175</volume>:<fpage>108504</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108504</pub-id><pub-id pub-id-type="pmid">38701593</pub-id></mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>Z.</given-names></name> <name><surname>Xu</surname> <given-names>M.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name> <name><surname>Ming</surname> <given-names>D.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;Decoding of individual finger movement on one hand using ultra high-density EEG,&#x0201D;</article-title> <italic>2022 16th ICME International Conference on Complex Medical Engineering (CME)</italic> (Zhongshan: IEEE). doi: <pub-id pub-id-type="doi">10.1109/CME55444.2022.10063299</pub-id></mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McFarland</surname> <given-names>D. J.</given-names></name> <name><surname>McCane</surname> <given-names>L. M.</given-names></name> <name><surname>David</surname> <given-names>S. V.</given-names></name> <name><surname>Wolpaw</surname> <given-names>J. R.</given-names></name></person-group> (<year>1997</year>). <article-title>Spatial filter selection for EEG-based communication</article-title>. <source>Electroencephalogr. Clin. Neurophysiol</source>. <volume>103</volume>:<fpage>386</fpage>&#x02013;<lpage>394</lpage> doi: <pub-id pub-id-type="doi">10.1016/S0013-4694(97)00022-2</pub-id><pub-id pub-id-type="pmid">9305287</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McFarland</surname> <given-names>D. J.</given-names></name> <name><surname>Miner</surname> <given-names>L. A.</given-names></name> <name><surname>Vaughan</surname> <given-names>T. M.</given-names></name> <name><surname>Wolpaw</surname> <given-names>J. R.</given-names></name></person-group> (<year>2000</year>). <article-title>Mu and beta rhythm topographies during motor imagery and actual movements</article-title>. <source>Brain Topogr</source>. <volume>12</volume>, <fpage>177</fpage>&#x02013;<lpage>186</lpage> doi: <pub-id pub-id-type="doi">10.1023/A:1023437823106</pub-id><pub-id pub-id-type="pmid">10791681</pub-id></mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>K.</given-names></name> <name><surname>Zanos</surname> <given-names>S.</given-names></name> <name><surname>Fetz</surname> <given-names>E.</given-names></name> <name><surname>den Nijs</surname> <given-names>M.</given-names></name> <name><surname>Ojemann</surname> <given-names>J.</given-names></name></person-group> (<year>2009</year>). <article-title>Decoupling the cortical power spectrum reveals real-time representation of individual finger movements in humans</article-title>. <source>J. Neurosci</source>. <volume>29</volume>:<fpage>3132</fpage>&#x02013;<lpage>3137</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.5506-08.2009</pub-id><pub-id pub-id-type="pmid">19279250</pub-id></mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>K. J.</given-names></name> <name><surname>Hermes</surname> <given-names>D.</given-names></name> <name><surname>Honey</surname> <given-names>C. J.</given-names></name> <name><surname>Hebb</surname> <given-names>A. O.</given-names></name> <name><surname>Ramsey</surname> <given-names>N. F.</given-names></name> <name><surname>Knight</surname> <given-names>R. T.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Human motor cortical activity is selectively phase-entrained on underlying rhythms</article-title>. <source>PLOS Comput. Biol</source>. <volume>8</volume>:<fpage>e1002655</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002655</pub-id><pub-id pub-id-type="pmid">22969416</pub-id></mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>K. J.</given-names></name> <name><surname>Makeig</surname> <given-names>S.</given-names></name> <name><surname>Hebb</surname> <given-names>A.</given-names></name> <name><surname>Rao</surname> <given-names>R.</given-names></name> <name><surname>den Nijs</surname> <given-names>M.</given-names></name> <name><surname>Ojemann</surname> <given-names>J.</given-names></name></person-group> (<year>2007</year>). <article-title>Cortical electrode localization from X-rays and simple mapping for electrocorticographic research: the &#x0201C;Location on Cortex&#x0201D; (LOC) package for MATLAB</article-title>. <source>J. Neurosci. Methods</source> <volume>162</volume>, <fpage>303</fpage>&#x02013;<lpage>308</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jneumeth.2007.01.019</pub-id><pub-id pub-id-type="pmid">17343918</pub-id></mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>K. J.</given-names></name> <name><surname>Sorensen</surname> <given-names>L. B.</given-names></name> <name><surname>Ojemann</surname> <given-names>J. G.</given-names></name> <name><surname>den Nijs</surname> <given-names>M.</given-names></name></person-group> (<year>2009</year>). <article-title>Power-law scaling in the brain surface electric potential</article-title>. <source>PLOS Comput. Biol</source>. <volume>5</volume>:<fpage>e1000609</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pcbi.1000609</pub-id><pub-id pub-id-type="pmid">20019800</pub-id></mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>M&#x000FC;ller-Gerking</surname> <given-names>J.</given-names></name> <name><surname>Pfurtscheller</surname> <given-names>G.</given-names></name> <name><surname>Flyvbjerg</surname> <given-names>H.</given-names></name></person-group> (<year>1999</year>). <article-title>Designing optimal spatial filters for single-trial EEG classification in a movement task</article-title>. <source>Clin. Neurophysiol</source>. <volume>110</volume>, <fpage>787</fpage>&#x02013;<lpage>798</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S1388-2457(98)00038-8</pub-id><pub-id pub-id-type="pmid">10400191</pub-id></mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nemes</surname> <given-names>&#x000C1;.</given-names></name> <name><surname>Eigner</surname> <given-names>G.</given-names></name> <name><surname>Shi</surname> <given-names>P.</given-names></name></person-group> (<year>2024</year>). <article-title>Application of deep learning to enhance finger movement classification accuracy from UHD-EEG signals</article-title>. <source>IEEE</source> <volume>Access13</volume>, <fpage>6154</fpage>&#x02013;<lpage>6154</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2024.3407690</pub-id></mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Paek</surname> <given-names>A.</given-names></name> <name><surname>Agashe</surname> <given-names>H. A.</given-names></name> <name><surname>Contreras-Vidal</surname> <given-names>J. L.</given-names></name></person-group> (<year>2014</year>). <article-title>Decoding repetitive finger movements with brain activity acquired via non-invasive electroencephalography</article-title>. <source>Front. Neuroeng</source>. <volume>7</volume>:<fpage>3</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fneng.2014.00003</pub-id><pub-id pub-id-type="pmid">24659964</pub-id></mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>J.</given-names></name> <name><surname>Woo</surname> <given-names>S.</given-names></name> <name><surname>Lee</surname> <given-names>J.-Y.</given-names></name> <name><surname>Kweon</surname> <given-names>I. S.</given-names></name></person-group> (<year>2018</year>). <article-title>BAM: Bottleneck attention module</article-title>. <source>arXiv</source> [preprint] arXiv:1807.06514. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1807.06514</pub-id></mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pfurtscheller</surname> <given-names>G.</given-names></name> <name><surname>Guger</surname> <given-names>C.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>G.</given-names></name> <name><surname>Krausz</surname> <given-names>G.</given-names></name> <name><surname>Neuper</surname> <given-names>C.</given-names></name></person-group> (<year>2000</year>). <article-title>Brain oscillations control hand orthosis in a tetraplegic</article-title>. <source>Neurosci. Lett</source>. <volume>292</volume>, <fpage>211</fpage>&#x02013;<lpage>214</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0304-3940(00)01471-3</pub-id><pub-id pub-id-type="pmid">11018314</pub-id></mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pfurtscheller</surname> <given-names>G.</given-names></name> <name><surname>Lopes da Silva</surname> <given-names>F. H.</given-names></name></person-group> (<year>1999</year>). <article-title>Event-related EEG/MEG synchronization and desynchronization: basic principles</article-title>. <source>Clin. Neurophysiol</source>. <volume>110</volume>, <fpage>1842</fpage>&#x02013;<lpage>1857</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S1388-2457(99)00141-8</pub-id><pub-id pub-id-type="pmid">10576479</pub-id></mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pfurtscheller</surname> <given-names>G.</given-names></name> <name><surname>Neuper</surname> <given-names>C.</given-names></name> <name><surname>Flotzinger</surname> <given-names>D.</given-names></name> <name><surname>Pregenzer</surname> <given-names>M.</given-names></name></person-group> (<year>1997</year>). <article-title>EEG-based discrimination between imagination of right and left hand movement</article-title>. <source>Electroencephalogr. Clin. Neurophysiol</source>. <volume>105</volume>, <fpage>281</fpage>&#x02013;<lpage>288</lpage>. <pub-id pub-id-type="pmid">9546492</pub-id></mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Quadrianto</surname> <given-names>N.</given-names></name> <name><surname>Guan</surname> <given-names>C.</given-names></name> <name><surname>Dat</surname> <given-names>T.</given-names></name> <name><surname>Xue</surname> <given-names>P.</given-names></name></person-group> (<year>2007</year>). <article-title>&#x0201C;Sub-band common spatial pattern (SBCSP) for brain-computer interface,&#x0201D;</article-title> in <source>2007 3rd International IEEE/EMBS Conference on Neural Engineering</source> (<publisher-loc>Kohala Coast, HI</publisher-loc>: <publisher-name>IEEE</publisher-name>), 204-207.</mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schieber</surname> <given-names>M. H.</given-names></name></person-group> (<year>2001</year>). <article-title>Constraints on somatotopic organization in the primary motor cortex</article-title>. <source>J. Neurophysiol</source>. <volume>86</volume>, <fpage>2125</fpage>&#x02013;<lpage>2143</lpage>. doi: <pub-id pub-id-type="doi">10.1152/jn.2001.86.5.2125</pub-id><pub-id pub-id-type="pmid">11698506</pub-id></mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schieber</surname> <given-names>M. H.</given-names></name> <name><surname>Santello</surname> <given-names>M.</given-names></name></person-group> (<year>2004</year>). <article-title>Hand function: peripheral and central constraints on performance</article-title>. <source>J. Appl. Physiol</source>. <volume>96</volume>, <fpage>2293</fpage>&#x02013;<lpage>2300</lpage>. doi: <pub-id pub-id-type="doi">10.1152/japplphysiol.01063.2003</pub-id><pub-id pub-id-type="pmid">15133016</pub-id></mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schirrmeister</surname> <given-names>R. T.</given-names></name> <name><surname>Springenberg</surname> <given-names>J. T.</given-names></name> <name><surname>Fiederer</surname> <given-names>L. D. J.</given-names></name> <name><surname>Glasstetter</surname> <given-names>M.</given-names></name> <name><surname>Eggensperger</surname> <given-names>K.</given-names></name> <name><surname>Tangermann</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Deep learning with convolutional neural networks for EEG decoding and visualization</article-title>. <source>arXiv</source> [preprint] arXiv:1703.05051. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1703.05051</pub-id><pub-id pub-id-type="pmid">28782865</pub-id></mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schreiner</surname> <given-names>L.</given-names></name> <name><surname>Jordan</surname> <given-names>M.</given-names></name> <name><surname>Sieghartsleitner</surname> <given-names>S.</given-names></name> <name><surname>Kapeller</surname> <given-names>C.</given-names></name> <name><surname>Pretl</surname> <given-names>H.</given-names></name> <name><surname>Kamada</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Mapping of the central sulcus using non-invasive ultra-high-density brain recordings</article-title>. <source>Sci. Rep</source>. <volume>14</volume>:<fpage>57167</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-57167-y</pub-id><pub-id pub-id-type="pmid">38499709</pub-id></mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Schreiner</surname> <given-names>L.</given-names></name> <name><surname>Sieghartsleitner</surname> <given-names>S.</given-names></name> <name><surname>Mayr</surname> <given-names>K.</given-names></name> <name><surname>Pretl</surname> <given-names>H.</given-names></name> <name><surname>Guger</surname> <given-names>C.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Hand gesture decoding using ultra-high-density EEG,&#x0201D;</article-title> in <source>2023 11th International IEEE/EMBS Conference on Neural Engineering (NER)</source> (<publisher-loc>Baltimore, MD</publisher-loc>: <publisher-name>IEEE</publisher-name>).</mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Shenoy</surname> <given-names>P.</given-names></name> <name><surname>Miller</surname> <given-names>K. J.</given-names></name> <name><surname>Ojemann</surname> <given-names>J. G.</given-names></name> <name><surname>Rao</surname> <given-names>R. P. N.</given-names></name></person-group> (<year>2007</year>). <article-title>&#x0201C;Finger movement classification for an electrocorticographic BCI,&#x0201D;</article-title> in <source>2007 3rd International IEEE/EMBS Conference on Neural Engineering</source> (<publisher-loc>Kohala Coast, HI</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>192</fpage>&#x02013;<lpage>195</lpage>.</mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sherstinsky</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network</article-title>. <source>arXiv</source> [preprint] arXiv:1808.03314. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1808.03314</pub-id></mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>E.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name> <name><surname>Kang</surname> <given-names>Y.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Zhao</surname> <given-names>L.</given-names></name> <name><surname>Zhu</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>MEET: A multi-band EEG transformer for brain states decoding</article-title>. <source>IEEE Trans. Biomed. Eng</source>. <volume>71</volume>, <fpage>1442</fpage>&#x02013;<lpage>1453</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TBME.2023.3339892</pub-id><pub-id pub-id-type="pmid">38055365</pub-id></mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>Y.</given-names></name> <name><surname>Zheng</surname> <given-names>Q.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name> <name><surname>Gao</surname> <given-names>X.</given-names></name></person-group> (<year>2023</year>). <article-title>EEG conformer: convolutional transformer for EEG decoding and visualization</article-title>. <source>IEEE Trans. Neural Syst. Rehabilitat. Eng</source>. <volume>31</volume>, <fpage>710</fpage>&#x02013;<lpage>719</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TNSRE.2022.3230250</pub-id><pub-id pub-id-type="pmid">37015413</pub-id></mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tangermann</surname> <given-names>M.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <name><surname>Aertsen</surname> <given-names>A.</given-names></name> <name><surname>Birbaumer</surname> <given-names>N.</given-names></name> <name><surname>Braun</surname> <given-names>C.</given-names></name> <name><surname>Brunner</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Review of the BCI competition IV</article-title>. <source>Front. Neurosci</source>. <volume>5</volume>:<fpage>55</fpage> doi: <pub-id pub-id-type="doi">10.3389/fnins.2012.00055</pub-id><pub-id pub-id-type="pmid">22811657</pub-id></mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A.</given-names></name> <name><surname>Shazeer</surname> <given-names>N.</given-names></name> <name><surname>Parmar</surname> <given-names>N.</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J.</given-names></name> <name><surname>Jones</surname> <given-names>L.</given-names></name> <name><surname>Gomez</surname> <given-names>A. N.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>&#x0201C;Attention is all you need,&#x0201D;</article-title> in <source>Neural Information Processing Systems</source>.</mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Volkova</surname> <given-names>K.</given-names></name> <name><surname>Lebedev</surname> <given-names>M. A.</given-names></name> <name><surname>Kaplan</surname> <given-names>A.</given-names></name> <name><surname>Ossadtchi</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>Decoding movement from electrocorticographic activity: a review</article-title>. <source>Front. Neuroinform</source>. <volume>13</volume>:<fpage>74</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fninf.2019.00074</pub-id><pub-id pub-id-type="pmid">31849632</pub-id></mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wolpaw</surname> <given-names>J. R.</given-names></name> <name><surname>Birbaumer</surname> <given-names>N.</given-names></name> <name><surname>McFarland</surname> <given-names>D. J.</given-names></name> <name><surname>Pfurtscheller</surname> <given-names>G.</given-names></name> <name><surname>Vaughan</surname> <given-names>T. M.</given-names></name></person-group> (<year>2002</year>). <article-title>Brain-computer interfaces for communication and control</article-title>. <source>Clin. Neurophysiol</source>. <volume>113</volume>, <fpage>767</fpage>&#x02013;<lpage>791</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S1388-2457(02)00057-3</pub-id><pub-id pub-id-type="pmid">21984822</pub-id></mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Woo</surname> <given-names>S.</given-names></name> <name><surname>Park</surname> <given-names>J.</given-names></name> <name><surname>Lee</surname> <given-names>J. Y.</given-names></name> <name><surname>Kweon</surname> <given-names>I. S.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;CBAM: convolutional block attention module,&#x0201D;</article-title> in <source>European Conference on Computer Vision</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>.</mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>X.</given-names></name> <name><surname>Metcalfe</surname> <given-names>B.</given-names></name> <name><surname>He</surname> <given-names>S.</given-names></name> <name><surname>Tan</surname> <given-names>H.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>A review of motor brain-computer interfaces using intracranial electroencephalography based on surface electrodes and depth electrodes</article-title>. <source>IEEE Trans. Neural Syst. Rehabil. Eng</source>. <volume>32</volume>, <fpage>2408</fpage>&#x02013;<lpage>2431</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TNSRE.2024.3421551</pub-id><pub-id pub-id-type="pmid">38949928</pub-id></mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xiao</surname> <given-names>R.</given-names></name> <name><surname>Ding</surname> <given-names>L.</given-names></name></person-group> (<year>2013</year>). <article-title>Evaluation of EEG features in decoding individual finger movements from one hand</article-title>. <source>Comp. Mathem. Methods Med</source>. <volume>2013</volume>:<fpage>243257</fpage>. doi: <pub-id pub-id-type="doi">10.1155/2013/243257</pub-id><pub-id pub-id-type="pmid">23710250</pub-id></mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yao</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>T.</given-names></name> <name><surname>Ding</surname> <given-names>P.</given-names></name> <name><surname>Wang</surname> <given-names>F.</given-names></name> <name><surname>Zhao</surname> <given-names>L.</given-names></name> <name><surname>Gong</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Emotion classification based on transformer and CNN for EEG spatial-temporal feature learning</article-title>. <source>Brain Sci</source>. <volume>14</volume>:<fpage>268</fpage>. doi: <pub-id pub-id-type="doi">10.3390/brainsci14030268</pub-id><pub-id pub-id-type="pmid">38539656</pub-id></mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1512649/overview">Qingshan She</ext-link>, Hangzhou Dianzi University, China</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/8360/overview">Murat Okatan</ext-link>, Istanbul Technical University, T&#x000FC;rkiye</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/767870/overview">Ke Liu</ext-link>, Chongqing University of Posts and Telecommunications, China</p>
</fn>
</fn-group>
</back>
</article>