<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Neurosci.</journal-id>
<journal-title>Frontiers in Computational Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5188</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fncom.2024.1516440</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Multimodal consumer choice prediction using EEG signals and eye tracking</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Usman</surname> <given-names>Syed Muhammad</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2877644/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Khalid</surname> <given-names>Shehzad</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2935361/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Tanveer</surname> <given-names>Aimen</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2877719/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Imran</surname> <given-names>Ali Shariq</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1471132/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zubair</surname> <given-names>Muhammad</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2877652/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Computer Science, Bahria School of Engineering and Applied Science, Bahria University</institution>, <addr-line>Islamabad</addr-line>, <country>Pakistan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Computer Engineering, Bahria School of Engineering and Applied Science, Bahria University</institution>, <addr-line>Islamabad</addr-line>, <country>Pakistan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Creative Technologies, Air University</institution>, <addr-line>Islamabad</addr-line>, <country>Pakistan</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Computer Science, Norwegian University of Science and Technology</institution>, <addr-line>Gj&#x000F8;vik</addr-line>, <country>Norway</country></aff>
<aff id="aff5"><sup>5</sup><institution>Interdisciplinary Research Center for Finance and Digital Economy, King Fahd University of Petroleum and Minerals</institution>, <addr-line>Dhahran</addr-line>, <country>Saudi Arabia</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Saad Arif, King Faisal University, Saudi Arabia</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Ijaz Ul Haq, Universitat de Lleida, Spain</p>
<p>Syed Qasim Abbas, CSIRO e-Health Research Centre, Australia</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Ali Shariq Imran <email>ali.imran&#x00040;ntnu.no</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>08</day>
<month>01</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>18</volume>
<elocation-id>1516440</elocation-id>
<history>
<date date-type="received">
<day>24</day>
<month>10</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>19</day>
<month>12</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2025 Usman, Khalid, Tanveer, Imran and Zubair.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Usman, Khalid, Tanveer, Imran and Zubair</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Marketing plays a vital role in the success of a business, driving customer engagement, brand recognition, and revenue growth. Neuromarketing adds depth to this by employing insights into consumer behavior through brain activity and emotional responses to create more effective marketing strategies. Electroencephalogram (EEG) has typically been utilized by researchers for neuromarketing, whereas Eye Tracking (ET) has remained unexplored. To address this gap, we propose a novel multimodal approach to predict consumer choices by integrating EEG and ET data. Noise from EEG signals is mitigated using a bandpass filter, Artifact Subspace Reconstruction (ASR), and Fast Orthogonal Regression for Classification and Estimation (FORCE). Class imbalance is handled by employing the Synthetic Minority Over-sampling Technique (SMOTE). Handcrafted features, including statistical and wavelet features, and automated features from Convolutional Neural Network and Long Short-Term Memory (CNN-LSTM), have been extracted and concatenated to generate a feature space representation. For ET data, preprocessing involved interpolation, gaze plots, and SMOTE, followed by feature extraction using LeNet-5 and handcrafted features like fixations and saccades. Multimodal feature space representation was generated by performing feature-level fusion for EEG and ET, which was later fed into a meta-learner-based ensemble classifier with three base classifiers, including Random Forest, Extended Gradient Boosting, and Gradient Boosting, and Random Forest as the meta-classifier, to perform classification between buy vs. not buy. The performance of the proposed approach is evaluated using a variety of performance metrics, including accuracy, precision, recall, and F1 score. Our model demonstrated superior performance compared to competitors by achieving 84.01% accuracy in predicting consumer choices and 83% precision in identifying positive consumer preferences.</p></abstract>
<kwd-group>
<kwd>EEG</kwd>
<kwd>eye tracking</kwd>
<kwd>neuromarketing</kwd>
<kwd>CNN-LSTM</kwd>
<kwd>multimodal</kwd>
</kwd-group>
<counts>
<fig-count count="11"/>
<table-count count="7"/>
<equation-count count="23"/>
<ref-count count="39"/>
<page-count count="15"/>
<word-count count="9411"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1 Introduction</title>
<p>Neuromarketing, a dynamic fusion of neuroscience and marketing, has emerged through the innovative use of non-invasive Brain&#x02013;Computer Interface (BCI) technology, revolutionizing the concept of marketing. Marketing is a connection between production and consumers. A good product can fail to target its desired audience without effective marketing (Assel, <xref ref-type="bibr" rid="B6">1995</xref>). To create products and services with the highest profit potential, it is crucial to thoroughly understand consumer behavior and develop a corresponding advertising strategy. This requires a comprehensive understanding of the buyer&#x00027;s decision-making process, which typically includes need recognition, information search, evaluation, purchase decision, and post-purchase behavior (Armstrong et al., <xref ref-type="bibr" rid="B5">2014</xref>; Peter et al., <xref ref-type="bibr" rid="B29">1999</xref>; Vecchiato et al., <xref ref-type="bibr" rid="B36">2011</xref>). Researchers have employed Electroencephalography (EEG) and Eye Tracking (ET) to analyze the brain activity and gaze outcomes when exposed to different stimuli for several decades.</p>
<p>EEG is a technique used to assess the electrical activity within a person&#x00027;s cranial structure. This involves placing numerous electrodes on the scalp, a method known as scalp EEG. It is particularly preferred for recording brain waves because it is simple and does not involve any invasive procedure, while other methods are preferable because they are efficient in monitoring brain activity (Teplan, <xref ref-type="bibr" rid="B34">2002</xref>; Fisch, <xref ref-type="bibr" rid="B12">1999</xref>). It records changes in electrical activity and oscillations within the brain. The amplitude of the signals are proportional to the type of mental activity experienced when exposed to stimuli (Homan et al., <xref ref-type="bibr" rid="B20">1987</xref>). Eye tracking, on the other hand, involves gathering information on visual attention through the capturing of eye movements. The eye tracking revealed where and for how long a person looked at the different elements, whereas EEG can uncover the emotional and cognitive response elicited by these stimuli.</p>
<p>Neuromarketing, a multidisciplinary field at the intersection of neuroscience, psychology, and economics, explores the complex dynamics of how advertisements can significantly impact product sales. Unlike traditional marketing research methods such as interviews, reviews, and questionnaires, neuromarketing seeks to surpass the limitations inherent in these approaches. These conventional methods often fall short of fully revealing consumers&#x00027; insights toward products, as individuals may encounter challenges in conveying their preferences or may be hesitant to express them comprehensively. Moreover, the chances of data manipulation add a layer of complexity to the reliability of findings.</p>
<p>Human behavior is influenced by processes operating beneath the conscious threshold. In response to these challenges, neuromarketing offers a revolutionary shift, going beyond direct questions about products and exploring the deeper subconscious areas of consumers&#x00027; minds. The essence is to get insights in a non-invasive manner, extracting authentic preferences and choices that may outstand conventional probing techniques. It offers a deeper and more precise insight into consumer behavior. This leads to the development of innovative and successful marketing tactics, ultimately driving increased sales. In the expansive and intricate landscape of the advertising industry, where expenditures vary based on geographical location, industry sector, and individual company strategies. The main contributions of this research study are as follows:</p>
<list list-type="bullet">
<list-item><p>A novel multimodal framework has been proposed, integrating EEG signals and eye-tracking data to enhance consumer preference prediction. This approach combines the strengths of both modalities, addressing the lack of sufficient multimodal research in the domain.</p></list-item>
<list-item><p>A robust feature extraction pipeline has been designed, combining handcrafted features and automated features derived through deep learning. This hybrid approach provides a more comprehensive representation of the data, bridging an identified gap in the existing literature.</p></list-item>
<list-item><p>Ensemble classification techniques have been proposed to address the challenges of class imbalance and improve prediction accuracy. By utilizing multiple classifiers and optimizing their integration, significant improvements in performance metrics were achieved compared to traditional methods.</p></list-item>
</list></sec>
<sec id="s2">
<title>2 Literature review</title>
<p>Many individuals are often reserved in expressing their complete thoughts and preferences during product evaluation, creating a challenge in comprehending the complexities of consumer decision-making. The emergence of neuroimaging tools provides a quick and convenient method to understand a customer&#x00027;s brain activity when evaluating and choosing different products. Consumer choice recognition typically involves three pivotal stages. The initial step encompasses preprocessing, wherein unwanted noise is eliminated from both EEG and ET signals. Following this, relevant features are extracted, and subsequently, EEG and ET signals are classified based on consumer preferences. In neuromarketing studies, the recording of both EEG and ET data equips researchers to get into the complex interplay of factors that influence how the human psyche makes choices among different products.</p>
<sec>
<title>2.1 Predictive approaches for consumer preference based on EEG signals</title>
<p>Researchers have proposed multiple methods for classification between like vs. dislike for neuromarketing in recent years. A typical method consists of preprocessing the EEG signals and extracting the features followed by the classification. Researchers have used various preprocessing techniques employed in predicting consumer preferences. Bandpass filtering, widely utilized for EEG signal noise reduction in numerous studies (Murugappan et al., <xref ref-type="bibr" rid="B28">2014</xref>; Alimardani and Kaba, <xref ref-type="bibr" rid="B3">2021</xref>; Aldayel et al., <xref ref-type="bibr" rid="B2">2021</xref>; Georgiadis et al., <xref ref-type="bibr" rid="B15">2022</xref>, <xref ref-type="bibr" rid="B16">2023a</xref>), serves as a prominent technique. Independent Component Analysis (ICA) has been adopted by researchers to eliminate noise in their proposed methods (Aldayel et al., <xref ref-type="bibr" rid="B2">2021</xref>; Georgiadis et al., <xref ref-type="bibr" rid="B15">2022</xref>; Telpaz et al., <xref ref-type="bibr" rid="B33">2015</xref>; Hakim et al., <xref ref-type="bibr" rid="B18">2021</xref>). Telpaz et al. (<xref ref-type="bibr" rid="B33">2015</xref>) and Hakim et al. (<xref ref-type="bibr" rid="B18">2021</xref>) have also applied the Notch Filter for preprocessing. Downsampling, an effective method employed by several researchers like (Aldayel et al., <xref ref-type="bibr" rid="B2">2021</xref>), proves valuable for reducing the sampling rate of EEG data. Moreover, the Savitzky&#x02013;Golay filter was utilized to effectively remove artifacts (Aldayel et al., <xref ref-type="bibr" rid="B2">2021</xref>; Yadava et al., <xref ref-type="bibr" rid="B39">2017</xref>; Shah et al., <xref ref-type="bibr" rid="B31">2022</xref>). Murugappan et al. (<xref ref-type="bibr" rid="B28">2014</xref>) applied the Surface Laplacian Filter, and Kumar et al. (<xref ref-type="bibr" rid="B24">2019</xref>) used high and low pass filters for the purpose of preprocessing EEG signals.</p>
<p>After the preprocessing of EEG signals, the extraction of features is pivotal for classifying likes and dislikes. Many approaches are employed for feature extraction like LSTM (Shah et al., <xref ref-type="bibr" rid="B31">2022</xref>). Telpaz et al. (<xref ref-type="bibr" rid="B33">2015</xref>) have leveraged N200, or N2, is an event-related potential (ERP) component. The Power Spectrum Density (PSD) provides the distribution of power across diverse frequencies in the signal (Murugappan et al., <xref ref-type="bibr" rid="B28">2014</xref>; Alimardani and Kaba, <xref ref-type="bibr" rid="B3">2021</xref>; Shah et al., <xref ref-type="bibr" rid="B31">2022</xref>). Similarly, Discrete Wavelet Transform (DWT) (Arif et al., <xref ref-type="bibr" rid="B4">2023</xref>) introduces a process of iteratively breaking down the signal into approximation and detail coefficients across multiple scales, a technique adeptly utilized by researchers for feature extraction (Aldayel et al., <xref ref-type="bibr" rid="B2">2021</xref>; Yadava et al., <xref ref-type="bibr" rid="B39">2017</xref>; Shah et al., <xref ref-type="bibr" rid="B31">2022</xref>; Kumar et al., <xref ref-type="bibr" rid="B24">2019</xref>). Aldayel et al. (<xref ref-type="bibr" rid="B2">2021</xref>) have contributed by employing Welch Method. This metric, corresponding to the spatial standard deviation, offers insights into the amount of activity at each time point in the potential field. EEG signals are represented as Sample Covariance Matrices (SCMs) that are measured entities scattered over a particular Riemannian manifold by Georgiadis et al. (<xref ref-type="bibr" rid="B15">2022</xref>, <xref ref-type="bibr" rid="B16">2023a</xref>). One of the most commonly used method is to analyze EEG data is to break the signal into functionally distinct frequency bands. Telpaz et al. (<xref ref-type="bibr" rid="B33">2015</xref>) and Hakim et al. (<xref ref-type="bibr" rid="B18">2021</xref>) extracted frequency bands to extract features from EEG signals. These features provide high interclass variance which is useful in accurate classification. The details of these various features are briefly described in the following table understanding what kind of preprocessing techniques and feature extraction methods were used in this research, as shown in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Comparison of existing consumer preference prediction methods using EEG signals.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>References</bold></th>
<th valign="top" align="left"><bold>Data</bold></th>
<th valign="top" align="left"><bold>Year</bold></th>
<th valign="top" align="left"><bold>Preprocessing</bold></th>
<th valign="top" align="left"><bold>Feature extraction</bold></th>
<th valign="top" align="left"><bold>Classifier</bold></th>
<th valign="top" align="left"><bold>Accuracy (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Murugappan et al. (<xref ref-type="bibr" rid="B28">2014</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2014</td>
<td valign="top" align="left">Bandpass filter Surface Laplacian filter</td>
<td valign="top" align="left">PSD SE SC</td>
<td valign="top" align="left">kNN PNN</td>
<td valign="top" align="left">96.62</td>
</tr> <tr>
<td valign="top" align="left">Telpaz et al. (<xref ref-type="bibr" rid="B33">2015</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2015</td>
<td valign="top" align="left">Notch filter</td>
</tr>
<tr>
<td valign="top" align="left">ICA</td>
<td valign="top" align="left">ERSP N200 (ERP)</td>
<td valign="top" align="left">Random</td>
<td valign="top" align="left">59 65</td>
</tr> <tr>
<td valign="top" align="left">Yadava et al. (<xref ref-type="bibr" rid="B39">2017</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2017</td>
<td valign="top" align="left">Savitzky-Golay</td>
<td valign="top" align="left">DWT</td>
<td valign="top" align="left">HMM</td>
<td valign="top" align="left">70.33</td>
</tr> <tr>
<td valign="top" align="left">Aldayel et al. (<xref ref-type="bibr" rid="B2">2021</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2021</td>
<td valign="top" align="left">Downsampling Bandpass filter ICA Savitzky-Golay</td>
<td valign="top" align="left">DWT Welch method</td>
<td valign="top" align="left">DNN SVM kNN RF</td>
<td valign="top" align="left">83 81 73 87</td>
</tr> <tr>
<td valign="top" align="left">Alimardani and Kaba (<xref ref-type="bibr" rid="B3">2021</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2021</td>
<td valign="top" align="left">Bandpass filter</td>
<td valign="top" align="left">PSD</td>
<td valign="top" align="left">CNN EC (SVM RF, LOG)</td>
<td valign="top" align="left">74.57 63.5</td>
</tr> <tr>
<td valign="top" align="left">Hakim et al. (<xref ref-type="bibr" rid="B18">2021</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2021</td>
<td valign="top" align="left">Notch filter ICA</td>
<td valign="top" align="left">FBP Hemispheric symmetry</td>
<td valign="top" align="left">SVM LOG kNN DT</td>
<td valign="top" align="left">68.51</td>
</tr> <tr>
<td valign="top" align="left">Shah et al. (<xref ref-type="bibr" rid="B31">2022</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2022</td>
<td valign="top" align="left">Savitzky-Golay FFT SMOTE</td>
<td valign="top" align="left">DWT PSD LSTM</td>
<td valign="top" align="left">EC (SVM, DT, DNN)</td>
<td valign="top" align="left">96.89</td>
</tr> <tr>
<td valign="top" align="left">Georgiadis et al. (<xref ref-type="bibr" rid="B15">2022</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2022</td>
<td valign="top" align="left">Bandpass filter ICA</td>
<td valign="top" align="left">SCM</td>
<td valign="top" align="left">SVM Ensemble</td>
<td valign="top" align="left">73.11</td>
</tr> <tr>
<td valign="top" align="left">Georgiadis et al. (<xref ref-type="bibr" rid="B16">2023a</xref>)</td>
<td valign="top" align="left">EEG</td>
<td valign="top" align="left">2023</td>
<td valign="top" align="left">Bandpass Filter</td>
<td valign="top" align="left">SCM</td>
<td valign="top" align="left">SPDNet</td>
<td valign="top" align="left">72.18</td>
</tr></tbody>
</table>
</table-wrap>
<p>There are simple features such as the frequency distribution of words to parametric and non-parametric features, etc. for classification between the &#x0201C;like&#x0201D; and &#x0201C;dislike&#x0201D; classes. Statistical features in the time domain include the mean average, variance/ standard deviation, skewness, and kurtosis. Also, frequency domain features such as moments of spectrum like spectral centroid, variational coefficients, and even skewness in the spectrum can be incorporated. In addition, other techniques of dimensionality reduction such as Principal Component Analysis (PCA) have also been used in this study by the researchers to extract features and to reduce dimensionality. <xref ref-type="table" rid="T1">Table 1</xref> can give a brief idea of these various features and present an outline of the most important preprocessing proposals and the feature extraction applied in the present research.</p>
<p>Deep Neural Network (DNN), Support Vector Machine (SVM), Random Forest (RF), and <italic>k</italic>-Nearest Neighbors (kNN) resulted a maximum accuracy of 87%. Alimardani and Kaba (<xref ref-type="bibr" rid="B3">2021</xref>) proposed an ensemble classifier based on SVM, RF, Logistic Regression(LOG) and Convolution Neural Network(CNN). Murugappan et al. (<xref ref-type="bibr" rid="B28">2014</xref>) applied for <italic>k</italic>NN and Probabilistic Neural Network (PNN) for classification of EEG signals. Hakim et al. (<xref ref-type="bibr" rid="B18">2021</xref>) conducted a comprehensive study utilizing EEG, focusing solely on Machine Learning algorithms and acheived an accuracy of 68.51%. Shah et al. (<xref ref-type="bibr" rid="B31">2022</xref>) predicted users&#x00027; preferences for advertisements using an ensemble classifier [SVM, Decision Tree (DT), DNN] achieving an impressive accuracy of 96.89%. Yadava et al. (<xref ref-type="bibr" rid="B39">2017</xref>) presented the first dataset of neuromarketing. This dataset featured stimuli in the form of images of commercial products, labeled as either &#x0201C;like&#x0201D; or &#x0201C;dislike,&#x0201D; and Hidden Markov Model (HMM) was employed for the classification of EEG signals based on likes and dislikes. Georgiadis et al. (<xref ref-type="bibr" rid="B15">2022</xref>) applied a SVM Ensemble including three SVM classifiers, while their research (Georgiadis et al., <xref ref-type="bibr" rid="B16">2023a</xref>) used architecture of SPDNet. It is a deep learning architecture designed for processing data that lie on Symmetric Positive Definite (SPD) matrices.</p></sec>
<sec>
<title>2.2 Predictive approaches for consumer preference based on EEG signals and ET data</title>
<p>Researchers have employed various techniques to preprocess EEG signals and ET data for understanding consumer preferences. Khushaba et al. (<xref ref-type="bibr" rid="B22">2013</xref>) utilized a combination of ICA and DWT for EEG signal preprocessing. Matukin et al. (<xref ref-type="bibr" rid="B27">2016</xref>) and Samsuri et al. (<xref ref-type="bibr" rid="B30">2016</xref>) incorporated band-pass filtering in their methodologies. Christoforou et al. (<xref ref-type="bibr" rid="B10">2017</xref>) downsampled EEG data and applied a Notch filter to mitigate DC drifts. For processing pupil dilation signals, Slanzi et al. (<xref ref-type="bibr" rid="B32">2017</xref>) employed linear interpolation followed by band-pass filtering. Mashrur et al. (<xref ref-type="bibr" rid="B26">2024</xref>) adopted the Automatic Subspace reconstruction functionality from EEGLAB for noise reduction, subsequently applying a notch filter at 50 Hz to suppress power line artifacts.</p>
<p>Matukin et al. (<xref ref-type="bibr" rid="B27">2016</xref>) applied Fast Fourier Transform (FFT) to derive features from EEG signals. Samsuri et al. (<xref ref-type="bibr" rid="B30">2016</xref>) utilized P300 and N100 components for EEG signal analysis, while employing Pupil Dilation features for eye-tracking data. Christoforou et al. (<xref ref-type="bibr" rid="B10">2017</xref>) introduced the Attentional-asynchrony metric based on the Eye-Gaze Divergence Index and used epoched EEG measurements to formulate a Cognitive-congruency aggregate metric. Slanzi et al. (<xref ref-type="bibr" rid="B32">2017</xref>) employed Principal Component Analysis (PCA) to extract features from EEG signals. Garc&#x00301;&#x00131;a-Madariaga et al. (<xref ref-type="bibr" rid="B14">2019</xref>) focused on Alpha Band Oscillations for EEG signals and Area of Interest (AOI) for eye-tracking data. Mashrur et al. (<xref ref-type="bibr" rid="B26">2024</xref>) categorized features into three domains: time domain (TD), frequency domain (FD), and time-frequency domain (TFD), subsequently employing a classifier for optimal feature selection.</p>
<p><xref ref-type="table" rid="T2">Table 2</xref> provides comparative analysis of existing methods of neuromarketing based on EEG and ET. Khushaba et al. (<xref ref-type="bibr" rid="B22">2013</xref>) used mutual information analysis that indicated important factors affecting the buying decision. Samsuri et al. (<xref ref-type="bibr" rid="B30">2016</xref>) measured the attention levels of users when observing an advertisement through the use of EEG and ET signals. In the study by Christoforou et al. (<xref ref-type="bibr" rid="B10">2017</xref>), the R2 metric was employed to assess the predictive capability of the suggested neural and eye-tracking metrics on the box office success of films. Slanzi et al. (<xref ref-type="bibr" rid="B32">2017</xref>), aimed to determine the sections of a webpage that were most probable to attract clicks through the application of Logistic Regression. Mashrur et al. (<xref ref-type="bibr" rid="B26">2024</xref>) used the SVM classifier is used with RBF kernel for classifying strong and weak preference EEG signals attaining an accuracy of 97%.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Comparison of existing consumer preference prediction methods using EEG and ET data.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Author</bold></th>
<th valign="top" align="left"><bold>Data</bold></th>
<th valign="top" align="left"><bold>Year</bold></th>
<th valign="top" align="left"><bold>Preprocessing</bold></th>
<th valign="top" align="left"><bold>Feature extraction</bold></th>
<th valign="top" align="left"><bold>Classification</bold></th>
<th valign="top" align="left"><bold>Results</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Khushaba et al. (<xref ref-type="bibr" rid="B22">2013</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2013</td>
<td valign="top" align="left">ICA DWT</td>
<td valign="top" align="left">FFT</td>
<td valign="top" align="left">Mutual Information</td>
<td valign="top" align="left">Preference</td>
</tr> <tr>
<td valign="top" align="left">Matukin et al. (<xref ref-type="bibr" rid="B27">2016</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2016</td>
<td valign="top" align="left">Bandpass filter</td>
<td valign="top" align="left">FFT</td>
<td valign="top" align="left">Not specified</td>
<td valign="top" align="left">Improvement in Ads</td>
</tr> <tr>
<td valign="top" align="left">Samsuri et al. (<xref ref-type="bibr" rid="B30">2016</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2016</td>
<td valign="top" align="left">Bandpass filter</td>
<td valign="top" align="left">P300 ERP N100 Pupil dilation</td>
<td valign="top" align="left">Statistics</td>
<td valign="top" align="left">ERP and the ET results were inconsistent</td>
</tr> <tr>
<td valign="top" align="left">Christoforou et al. (<xref ref-type="bibr" rid="B10">2017</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2017</td>
<td valign="top" align="left">Downsampling Notch filter</td>
<td valign="top" align="left">Attent.Asynchrony Cogn. Congruency</td>
<td valign="top" align="left">Regression R2</td>
<td valign="top" align="left">72% accuracy</td>
</tr> <tr>
<td valign="top" align="left">Slanzi et al. (<xref ref-type="bibr" rid="B32">2017</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2017</td>
<td valign="top" align="left">Interpolation BandPass filter</td>
<td valign="top" align="left">PCA</td>
<td valign="top" align="left">Logistic regression</td>
<td valign="top" align="left">71.09% accuracy</td>
</tr> <tr>
<td valign="top" align="left">Garc&#x00301;&#x00131;a-Madariaga et al. (<xref ref-type="bibr" rid="B14">2019</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2019</td>
<td valign="top" align="left">Not specified</td>
<td valign="top" align="left">Alpha-Band Oscillation AOI</td>
<td valign="top" align="left">Not specified</td>
<td valign="top" align="left">Eye movements could predict packaging preference.</td>
</tr> <tr>
<td valign="top" align="left">Mashrur et al. (<xref ref-type="bibr" rid="B26">2024</xref>)</td>
<td valign="top" align="left">EEG ET</td>
<td valign="top" align="left">2023</td>
<td valign="top" align="left">ASR Notch filter</td>
<td valign="top" align="left">TD FD TFD</td>
<td valign="top" align="left">SVM-RBF</td>
<td valign="top" align="left">96.97% accuracy</td>
</tr></tbody>
</table>
</table-wrap>
<p>Following research gaps have been identified after a comprehensive literature review of both EEG and ET consumer preference prediction methods:</p>
<list list-type="bullet">
<list-item><p>There is a lack of sufficient multimodal research investigating the combined effectiveness of EEG and eye-tracking.</p></list-item>
<list-item><p>The issue of class imbalance remains a significant challenge in this field.</p></list-item>
<list-item><p>The integration of handcrafted and automated features in a combined feature set has not been much explored.</p></list-item>
<list-item><p>The limited use of ensemble learning methods represents a notable research gap.</p></list-item>
</list></sec></sec>
<sec id="s3">
<title>3 Dataset</title>
<p>The NeuMa dataset (Georgiadis et al., <xref ref-type="bibr" rid="B17">2023b</xref>) has been used for this particular research work and this comprises of 42 participants who were all Greek speakers; 23 males and 19 females. The dataset is made up of 144 supermarket products and this is presented in six brochure pages whereby each brochure is made up of 24 products. Targets were highlighted by users with a left-click of the mouse on products of interest. Consequently, each of the subjects has two files for every subject, which contains the records of their interactions with the products.</p>
<p><xref ref-type="table" rid="T3">Table 3</xref> provides brief description of the NeuMa dataset. Subjects were positioned at an arm&#x00027;s length or 50 cm away from the screen which is a 28 inch LCD monitor, and navigation on the digital brochure page and choice of products with the left click of the mouse. For the page navigation arrow keys of the keyboard were used. Every subject&#x00027;s data set involved EEG and eye-tracking information and mouse clicks and positions. There are EEG signals and eye movements, mouse clicks, and cursor movements collected in the given dataset. Among these data streams, currently only EEG and ET type data streams are being used.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Summary of NeuMa dataset.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Attribute</bold></th>
<th valign="top" align="left"><bold>Details</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Number of subjects</td>
<td valign="top" align="left">42 (23 males, 19 females)</td>
</tr> <tr>
<td valign="top" align="left">Number of products</td>
<td valign="top" align="left">144</td>
</tr> <tr>
<td valign="top" align="left">Number of pages</td>
<td valign="top" align="left">6 (24 products per page)</td>
</tr> <tr>
<td valign="top" align="left">Average selections</td>
<td valign="top" align="left">18 products per participant</td>
</tr> <tr>
<td valign="top" align="left">Data files per subject</td>
<td valign="top" align="left">2 (S01.xdf, S01.xls)</td>
</tr> <tr>
<td valign="top" align="left">EEG device</td>
<td valign="top" align="left">Wearable Sensing DSI24</td>
</tr> <tr>
<td valign="top" align="left">EEG sampling frequency</td>
<td valign="top" align="left">300 Hz</td>
</tr> <tr>
<td valign="top" align="left">EEG sensors</td>
<td valign="top" align="left">21 dry sensors</td>
</tr> <tr>
<td valign="top" align="left">ET device</td>
<td valign="top" align="left">Tobii pro fusion</td>
</tr> <tr>
<td valign="top" align="left">ET sampling frequency</td>
<td valign="top" align="left">120 Hz</td>
</tr></tbody>
</table>
</table-wrap>
<p>After the experiment, participants filled in a questionnaire containing demographic details about the individuals, profiling details about the participants as well as about the products provided to them like personality profile, tendency to indulge in impulse buying and about the products given to them like reasons for selection of product, familiarity with the product and frequency for buying the product. EEG data was recorded by DSI 24 system with the sampling rate of 300 Hz from 21 electrodes. This eye-tracking data was at a sampling rate of 120Hz and the Tobii Pro Fusion eye-tracker was used to collect the data. <xref ref-type="fig" rid="F1">Figures 1</xref>&#x02013;<xref ref-type="fig" rid="F3">3</xref> show the plots of EEG, ET, and Pupil dilation data, respectively (Tobii, <xref ref-type="bibr" rid="B35">2024</xref>; Georgiadis et al., <xref ref-type="bibr" rid="B17">2023b</xref>).</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Customer response (Subject S01): EEG data stream for product (NeuMa dataset: EEG data capturing the cerebral activity of a subject for a product).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0001.tif"/>
</fig>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Customer response (Subject S01): eye tracking coordinates (<italic>X, Y</italic>) for a product (NeuMa dataset: ET data revealing the gaze pattern linked to a product).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0002.tif"/>
</fig>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Customer response (Subject S01): pupil dilation stream for a product (NeuMa dataset: ET data revealing pupil dilation patterns linked to a product).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0003.tif"/>
</fig>
</sec>
<sec sec-type="methods" id="s4">
<title>4 Methodology</title>
<p>Proposed method consists of three steps: EEG and ET signal preprocessing, feature extraction and classification. The pre-processing of the EEG signals is done with the help of Bandpass Butterworth filter (0.5&#x02013;45Hz), Artifact Subspace Reconstruction (ASR) and the Fast Orthogonal Regression for Classification and Estimation (FORCE). Signals are then split in segments overlapping each other since the data amount is at a manageable size. In the same manner, preprocessing of Eye Tracking (ET) data; missing values are eliminated/taken care of using a linear interpolation and the data is segmented using overlapping window techniques. Non-technique based features are derived from the EEG and ET signals using statistical and frequency domain analysis The technique incorporated is CNN-LSTM for EEG and LeNet5 for ET data. First, for each input modality, feature-level fusion is used to combine these extracted features, and second, improvements are made to classification using both manually defined and learned features are used. <xref ref-type="fig" rid="F4">Figure 4</xref> displays the flow diagram of the proposed methodology.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Flow diagram of proposed methodology.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0004.tif"/>
</fig>
<sec>
<title>4.1 Preprocessing of EEG signals</title>
<p>Electroencephalogram (EEG) signals are often contaminated with various types of noise, including muscle activity, eye movements, and electrical interference from other devices. To analyze EEG data effectively, preprocessing steps such as filtering are crucial. Electroencephalography (EEG) signals require preprocessing to remove noise and isolate frequencies of interest. One method is applying a bandpass filter with band range 0.5&#x02013;45 Hz.</p>
<p>In order to filter out Signals with artifact in the EEG data a band pass filter was employed together with Artifact Subspace Reconstruction (ASR). ASR also helps in eradicating interferences like shrugs and blinks and leaves the signal&#x00027;s quality intact for analysis. This technique is very important in neuromarketing research as it offers clean signal filtration yet preserves the original signal. Fast method for Orthogonal Regression for Classification and Estimation EEG sounds are done using orthogonal basis vector and FORCE for the preprocessing of the data. It is applied to remove noise while improving the quality of the signal, making it suitable for the situations that require fast and accurate artifacts detection. Specifically, the signals from the EEG signals were band-pass filtered and then analyzed by ASR and FORCe to obtain the best results.</p>
<p>The overlapping window technique again divides the filtered signals to get more detailed data and make the signals continuous. The division of the continuous EEG signals make it more manageable and this was achieved by gaining small samples of 300 Hz with the window size being one second with 300 data entries. The overlapping of the windows has the advantage of achieving greater density of information and continuity of the signal.</p></sec>
<sec>
<title>4.2 Preprocessing of eye tracking data</title>
<p>Linear interpolation is a technique of curve fitting in which a straight line is drawn between two points to give the estimated point. It handles missing data if eye-tracking signals due to long blinking are missing using what is known as the straight-line interpolation method. Due to the ability to replace a missing value with approximated data samples that occur before and after the gap, a continuous signal is achieved. This step is necessary for preserving the quality of eye-tracking signal and further analysis of study subjects&#x00027; attention and eye movement behavior.</p>
<p>As for removing missing values in the eye-tracking dataset, linear interpolation has been applied, the next step of the data preprocessing is the data segmentation based on the overlapping window. This ensures that the maximum amount of information is collected as the windows have a 50% overlap in which every two consecutive windows have 50% of the same data points. This rises the density of data and contributes to non-fragmentation of signal which helps in maintaining coherency. The splitting of records further improves difference detection or comparison which is made possible by the window size of one second and a sampling rate of 120 Hz; this means that each segment&#x00027;s data set has 120 data points.</p>
<p>Gaze plots are basically eye movement data obtained through eye tracking displayed graphically as data points. They are developed by placing fixation areas on a graph of the observed stimulus. For the movements of the eyes, the <italic>X</italic> and <italic>Y</italic> coordinates are transformed into the 64 &#x000D7; 64 canvas where the black background implies no gaze while the white point marks a gaze. These points are the coordinates of the location on the canvas, and the original gaze plot images are saved, converted to the NumPy array and then to grayscale for analysis. <xref ref-type="fig" rid="F5">Figure 5</xref> presents various Gaze plots.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>A few gaze plots of ET data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0005.tif"/>
</fig>
<p>Class imbalance refers to situations where one class (the minority class) is significantly underrepresented compared to another class (the majority class). This class imbalance can lead to biased models that perform poorly on the minority class. To address the issue of class imbalance within the dataset, we used Synthetic Minority Over-sampling Technique (SMOTE) (Chawla et al., <xref ref-type="bibr" rid="B8">2002</xref>). SMOTE works by generating synthetic examples of the minority class to balance the class distribution. The process involves creating new instances of minority class samples by interpolating between existing minority class samples SMOTE first identifies the minority class samples in the dataset. For each minority class sample of EEG and ET data, SMOTE selects its k nearest neighbors in the feature space. The value of k we chose is 3, as it was giving the best results. For each minority class sample, SMOTE generates synthetic samples along the line segments connecting it to its k nearest neighbors. The number of synthetic samples created for each minority class sample is determined by a specified oversampling ratio. By creating synthetic samples, SMOTE increases the representation of the minority class in the dataset, balancing the class distribution. <xref ref-type="fig" rid="F6">Figure 6</xref> displays the class distribution before and after application of SMOTE.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Class distribution before and after SMOTE.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0006.tif"/>
</fig></sec>
<sec>
<title>4.3 Feature extraction of EEG signals</title>
<p>After preprocessing the EEG data, which typically involves filtering out noise and artifacts, the next step is to extract meaningful features from the cleaned data. Feature extraction transforms the raw EEG signals into a set of representative features that can be used for further analysis, such as classification. We used a few common statistical features include Mean, Variance, Skewness, and Kurtosis.</p>
<p>The mean of the EEG signal provides a measure of the central tendency of the signal. It indicates the average value of the signal over a specified period.</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>&#x003BC;</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>N</italic> is the number of data points and <italic>x</italic><sub><italic>i</italic></sub> represents the EEG signal values.</p>
<p>The co-efficient of variation is used to determine the spread of the signal value in relation to the mean of the EEG signals. That reveal information about the fluctuation in the activity of the brain.</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Skewness indicates the extent of probability distribution of the EEG signals asymmetrical nature. The absolute value of skewness is &#x0003E;1, &#x0003C;1 or zero if the distribution is highly skewed to the right, left, or symmetric respectively.</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Kurtosis quantifies the degree of the two at both the center and the tails of the probability density function of the EEG signal. It also implies that the data contains some outliers.</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>N</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msup><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>3</mml:mn><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Welch&#x00027;s Method (Welch, <xref ref-type="bibr" rid="B37">1967</xref>) is one of the robust and standard method to estimate power spectral density (PSD) of a signal. Even if it splits the signal into overlapping sections, then they apply a function known as windowing on sections, calculate the periodogram of each segment, and finally the averages these periodograms. In this feature, extraction was performed for all the EEG channels, considering the average power of the given signal within all possible frequency bands. This feature quantifies the amplitude deviations of the power from the energy of the signal at various frequency bands. The wavelet transform is the process by which a signal is broken down in different parts that are localized both temporally and in the frequency domain. Since, mean of DWT coefficients gives the average value of the coefficients, we obtained the mean of this parameter. This feature calculates the extent of fluctuations valued in the domain of wavelet coefficients.</p>
<p>Finally, Statistical features and frequency domain features and wavelet transform features are then combined to construct an information vector for each sample that will serve as the input to the model. The technical advantage implemented in the feature set uses time-frequency characteristic as well as multi-resolution analysis.</p>
<p>After removing noise from the EEG signals, features were extracted for the &#x0201C;Buy&#x0201D; and &#x0201C;No Buy&#x0201D; classes using two common approaches: handcrafted feature extraction and automated feature extraction via deep learning techniques. In the handcrafted approach, features are extracted without considering the class of the EEG signals. In contrast, automated feature extraction leverages deep learning models like Convolutional Neural Networks (CNNs) (LeCun et al., <xref ref-type="bibr" rid="B25">1998</xref>) and Long Short-Term Memory (LSTM) networks (Hochreiter and Schmidhuber, <xref ref-type="bibr" rid="B19">1997</xref>), which consider the class of the EEG signals during feature extraction. This method can lead the class of the EEG signals during feature extraction. This method can lead to improved classification performance due to lower intraclass variance and higher interclass variance.</p>
<p>LSTMs are a type of recurrent neural network (RNN) that are capable of learning long-term dependencies (Shah et al., <xref ref-type="bibr" rid="B31">2022</xref>). They have a chain-like structure with repeating modules. The core of the LSTM module consists of a cell state, and three gates to regulate the flow of information: the input gate, forget gate, and output gate. For feature extraction from EEG signals, we implement a convolutional neural network LSTM architecture. The CNN takes the segmented time-domain signals it has: the CNN&#x00027;s inputs are the number of EEG channels and temporal segments of signals. Convolutional layers perform spatial features extraction using filters of particular sizes which are succeeded by the max-pooling layers in an attempt to decrease the dimension and hence increasing the efficiency of the training process. The features from the CNN layers are flattened and reshaped so as to be fed into LSTM layer that takes into consideration temporality of the data. CNN and LSTM are combined because the former analyses the spatial information of the signals while the latter analyses the temporal information of the signals making it appropriate to classify the EEG signals. <xref ref-type="table" rid="T4">Table 4</xref> provides a summary of our proposed CNN-LSTM model.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Summary of proposed CNN-LSTM model.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Layer</bold></th>
<th valign="top" align="left"><bold>Output shape</bold></th>
<th valign="top" align="center"><bold>Parameters</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Input layer</td>
<td valign="top" align="left">(None, 19, 300, 1)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Conv2D</td>
<td valign="top" align="left">(None, 17, 298, 32)</td>
<td valign="top" align="center">320</td>
</tr> <tr>
<td valign="top" align="left">Max pooling</td>
<td valign="top" align="center">(None, 8, 149, 32)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Conv2D</td>
<td valign="top" align="left">(None, 6, 147, 64)</td>
<td valign="top" align="center">18,496</td>
</tr> <tr>
<td valign="top" align="left">MaxPooling</td>
<td valign="top" align="left">(None, 3, 73, 64)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Flatten layer</td>
<td valign="top" align="left">(None, 14,016)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Reshape layer</td>
<td valign="top" align="left">(None, 14,016, 1)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">LSTM</td>
<td valign="top" align="left">(None, 64)</td>
<td valign="top" align="center">16,832</td>
</tr> <tr>
<td valign="top" align="left">Dense</td>
<td valign="top" align="left">(None, 128)</td>
<td valign="top" align="center">8,320</td>
</tr> <tr>
<td valign="top" align="left">Dense</td>
<td valign="top" align="left">(None, 64)</td>
<td valign="top" align="center">8,256</td>
</tr></tbody>
</table>
</table-wrap></sec>
<sec>
<title>4.4 Feature extraction of ET data</title>
<p>After preprocessing ET data and handling the class imbalance issue features are extracted from it. Similar to feature extraction from EEG data, statistical features can be employed to quantify various aspects of these movements.</p>
<sec>
<title>4.4.1 Fixation duration</title>
<p>Fixation duration represents the average time a user spends fixating on a specific Area of Interest (AOI) and is analogous to the mean in EEG analysis. It provides insight into the level of attention paid to that area.</p></sec>
<sec>
<title>4.4.2 Saccade amplitude</title>
<p>Saccade amplitude is just like variance in EEG, its calculates the distance between one fixation to another fixation. Large value of saccade amplitude represents jump from one fixation to other.</p>
<p>We applied the LeNet-5 (LeCun et al., <xref ref-type="bibr" rid="B25">1998</xref>) model, a foundational CNN architecture developed by Yann LeCun, originally designed for image recognition tasks like classifying handwritten digits. The model processes input images through convolutional layers with filters to extract features, followed by max-pooling layers to reduce dimensionality. After multiple convolution and pooling layers, the feature maps are flattened into a vector for the classification layers. This structure effectively captures spatial features in the data, making it suitable for image recognition tasks. Here&#x00027;s an explanation for the LeNet-5 model summarized in <xref ref-type="table" rid="T5">Table 5</xref>.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Summary of LeNet-5 model.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Layer</bold></th>
<th valign="top" align="left"><bold>Output shape</bold></th>
<th valign="top" align="center"><bold>Parameters</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Input layer</td>
<td valign="top" align="left">(None, 64, 64, 1)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Conv2D</td>
<td valign="top" align="left">(None, 60, 60, 6)</td>
<td valign="top" align="center">456</td>
</tr> <tr>
<td valign="top" align="left">Max pooling</td>
<td valign="top" align="left">(None, 30, 30, 6)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Conv2D</td>
<td valign="top" align="left">(None, 26, 26, 16)</td>
<td valign="top" align="center">2,416</td>
</tr> <tr>
<td valign="top" align="left">MaxPooling</td>
<td valign="top" align="left">(None, 13, 13, 16)</td>
<td valign="top" align="center">0</td>
</tr> <tr>
<td valign="top" align="left">Flatten layer</td>
<td valign="top" align="left">(None, 2,704)</td>
<td valign="top" align="center">0</td>
</tr></tbody>
</table>
</table-wrap>
<p>Features extracted from EEG signals and ET data are concatenated to form a combined feature vector with a size of 16,720, which is then fed into an ensemble classifier. Optimizer used is Adam and loss function used is Mean Squared Error. Adam is a gradient-based optimization algorithm. Its update rules are given by:</p>
<p><bold>Update Rules:</bold></p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E7"><label>(7)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E8"><label>(8)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E9"><label>(9)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>&#x003B7;</mml:mi><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003F5;</mml:mi></mml:mrow></mml:mfrac><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where:</p>
<list list-type="bullet">
<list-item><p><italic>m</italic><sub><italic>t</italic></sub>: First moment (mean of gradients),</p></list-item>
<list-item><p><italic>v</italic><sub><italic>t</italic></sub>: Second moment (uncentered variance of gradients),</p></list-item>
<list-item><p><italic>g</italic><sub><italic>t</italic></sub>: Gradient at time step <italic>t</italic>,</p></list-item>
<list-item><p>&#x003B2;<sub>1</sub>, &#x003B2;<sub>2</sub>: Exponential decay rates for the moment estimates,</p></list-item>
<list-item><p>&#x003B7;: Learning rate,</p></list-item>
<list-item><p>&#x003F5;: Small constant to prevent division by zero,</p></list-item>
<list-item><p>&#x003B8;<sub><italic>t</italic></sub>: Parameters at time step <italic>t</italic>.</p></list-item>
</list>
<p>The Mean Squared Error loss function is given by:</p>
<disp-formula id="E10"><label>(10)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">MSE</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x00177;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where:</p>
<list list-type="bullet">
<list-item><p><italic>n</italic>: Number of data points,</p></list-item>
<list-item><p><italic>y</italic><sub><italic>i</italic></sub>: True value for the <italic>i</italic>-th data point,</p></list-item>
<list-item><p>&#x00177;<sub><italic>i</italic></sub>: Predicted value for the <italic>i</italic>-th data point.</p></list-item>
</list></sec></sec>
<sec>
<title>4.5 Ensemble classifier</title>
<p>After pre-processing and feature extraction, the final step is classification which is performed to categorize the sample as Buy vs. Non-buy. We have used three stacking ensemble classification approach in which features are first passed to three different classifiers including Random Forest, Gradient Boosting and XGBoost. Prediction obtained from these three classifiers is then stacked to get the final classification. Random Forest has beeen used as meta model in the stacking ensemble (Wolpert, <xref ref-type="bibr" rid="B38">1992</xref>).</p>
<sec>
<title>4.5.1 Base classifiers</title>
<p>Random Forest (RF): Random Forest grows a whole forest during training. It is a bagging technique in which every tree makes a prediction and then make a final prediction. This classifier is good to handle high dimensional data which is often the case with EEG and ET features (Breiman, <xref ref-type="bibr" rid="B7">2001</xref>)</p>
<disp-formula id="E11"><label>(11)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">Class label predicted by the&#x000A0;</mml:mtext><mml:mi>k</mml:mi><mml:mtext class="textrm" mathvariant="normal">-th tree.</mml:mtext></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E12"><label>(12)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">Majority Vote</mml:mtext><mml:mrow><mml:mo stretchy="false">{</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Gradient Boosting (GB): This is a strong method that constructs decision trees iteratively where each stage used in identifying the mistakes committed by the prior trees. Friedman (<xref ref-type="bibr" rid="B13">2001</xref>). The final prediction is:</p>
<disp-formula id="E13"><label>(13)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where:</p>
<list list-type="bullet">
<list-item><p><italic>h</italic><sub><italic>m</italic></sub>(<italic>x</italic>): the <italic>m</italic>-th weak learner,</p></list-item>
<list-item><p>&#x003B1;<sub><italic>m</italic></sub>: the weight of the <italic>m</italic>-th learner.</p></list-item>
</list>
<p>XGBoost (XGB): XGBoost is an optimized version of Gradient Boosting (Chen and Guestrin, <xref ref-type="bibr" rid="B9">2016</xref>). The prediction for XGBoost is:</p>
<disp-formula id="E14"><label>(14)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x003A9;</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where:</p>
<list list-type="bullet">
<list-item><p>&#x003B7;<sub><italic>m</italic></sub>: learning rate,</p></list-item>
<list-item><p>&#x003A9;(<italic>h</italic><sub><italic>m</italic></sub>): regularization term.</p></list-item>
</list></sec></sec>
<sec>
<title>4.6 Meta-classifier</title>
<p>The meta-classifier takes the outputs of the base classifiers as input. In this case, a Random Forest is used as the meta-classifier.</p>
<p>In the first step, predictions are collected from the base classifiers for the training dataset:</p>
<disp-formula id="E15"><label>(15)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E16"><label>(16)</label><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>Z</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x022EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x022EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x022EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In the final step, meta classifiers is trained to get the final classification result on <italic>Z</italic>. We have used Random Forest as meta classifier.</p>
<disp-formula id="E17"><label>(17)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Meta</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">Meta-RF</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>For unseen data <italic>x</italic>, the stacking ensemble works as follows: Each base model makes a prediction:</p>
<disp-formula id="E18"><label>(18)</label><mml:math id="M18"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>These predictions form a new feature vector for <italic>x</italic>:</p>
<disp-formula id="E19"><label>(19)</label><mml:math id="M19"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>Z</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mtable style="text-align:axis;" equalrows="false" columnlines="none none none none none none none none none" equalcolumns="false" class="array"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">RF</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">GB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">XGB</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The meta-classifier uses <italic>Z</italic><sub><italic>x</italic></sub> to make the final prediction:</p>
<disp-formula id="E20"><label>(20)</label><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Final</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Meta</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Base model outputs are as follows:</p>
<disp-formula id="E21"><mml:math id="M21"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mtext>RF</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mtext>Random&#x000A0;Forest&#x000A0;prediction</mml:mtext><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mtext>GB</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mtext>Gradient&#x000A0;Boosting&#x000A0;prediction</mml:mtext><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mtext>XGB</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mtext>XGBoost&#x000A0;prediction</mml:mtext><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Meta-model (Random Forest) output is as follows:</p>
<disp-formula id="E22"><label>(21)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Meta</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">Majority Vote</mml:mtext><mml:mrow><mml:mo stretchy="false">{</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>Z</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">}</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:mi>K</mml:mi><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Final stacking ensemble prediction:</p>
<disp-formula id="E23"><label>(22)</label><mml:math id="M23"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Final</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Meta</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></sec>
<sec>
<title>4.7 Hyperparameters optimization</title>
<p>For the machine learning models, Random Forest Classifier was set with 265 estimators for the Optuna (Akiba et al., <xref ref-type="bibr" rid="B1">2019</xref>) tuned model and 100 for the Stacking Classifier final estimator. The Gradient Boosting Classifier uses 89 estimators and the XGB Classifier is set with 300 estimators. These three estimators are surrounded by the Stacking Classifier such that the Random Forest classifier is used as the final estimator. Further, the imbalance of the data is tackled using SMOTE with the specified random state of 42. To split the dataset into cross-validation, the keyword Stratified K-Fold is used with the parameter setting of the number of folds as 10, shuffle as True, and random state as 42.</p>
<p>In the case of the deep learning models used in automatic feature extraction, the CNN connected with the LSTM is applied for the feature extraction of the EEG data. The structure of the model consists of an LSTM layer with 64 neurons and dense layers with 128 and 64 neurons, optimizer used is Adam and loss function used is Mean Squared Error. For the eye-tracking data the LeNet-5 is employed; it consists of two dense layers with 120 and 84 units and a sigmoid layer is used at the output for binary classification. This model is trained with the Adam optimizer with binary cross-entropy as the loss function and accuracy as the parameter over 50 epochs. The data splitting involves a train test split of 80&#x02013;20 and further division of the remaining data in equal proportions to validate and test the model.</p>
<sec>
<title>4.7.1 Stratified cross-validation</title>
<p>Once we were set up, with the ensemble classification pipeline formulated, the next logical step was to assess its utility. To this end, the strategy used was a rigorous method known as the stratified 10-fold cross-validation (Kohavi, <xref ref-type="bibr" rid="B23">1995</xref>). However, stratified cross-validation goes one step further than this as it guarantees the resultant folds as having the same proportion of classes as those of the original data-set.</p></sec></sec></sec>
<sec id="s5">
<title>5 Results and discussion</title>
<p>The efficiency of classification models is evaluated in terms of the metrics that measure the ability of the ML algorithm to classify the objects appropriately. Selecting the appropriate metrics is essential for achieving an accurate and objective assessment and measuring performance in such problems with skewed classes or different costs associated with an error. Accuracy for the most basic performance indicator that show the number of instances out of all the data that belong to the correct class. Precision also known as positive predictive value, measures the proportion of true positives among all predicted positives. It reflects how often the model correctly identifies a positive case.</p>
<p>Specificity test evaluates the proportion of actual negatives which are correctly identified by the model as negative, while, recall or sensitivity evaluates the proportion of actual positives which are correctly identified by the model as positive. They indicate how well the model captures all the positive instances in relation to the available training examples. F1 score is an average of recall and precision that yields proportional insights into both these measures. It&#x00027;s particularly useful when both false positives and false negatives are equally undesirable. It&#x00027;s particularly useful when both false positives and false negatives are equally undesirable.</p>
<p><xref ref-type="table" rid="T6">Table 6</xref> represents the quantitative comparison of the employed methods, namely accuracy, precision rate, recall, and F1 score. The proposed method achieves the highest accuracy of 0.84, significantly outperforming the other methods. The improvement in accuracy can be attributed to the effective integration of ML and DL features along with the stacking ensemble technique. The precision of the proposed method (0.83) indicates its superior ability to correctly identify positive instances compared to other methods. This is particularly important in reducing false positives, which is critical in applications where the cost of false positives is high. For recall the proposed method gives 0. 84 which shows that the proposed method is also good in the recall sense it captures most of the true positive instances. Large recall component means that the model is going to include many more positives into the result set at the cost of possibly including negative instances, which is particularly important where false negatives are undesirable. The proposed method claims to achieve an F1 score of 0.83 which balances the precision and recall rates of identifying fishes with equal importance. This is because F1 score is a harmonic mean of precision and recall and a high F1 score indicates that the propose method is both precise and accurate, studied and tested on different measure standards and tables.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Evaluation metrics for different methods.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
<th valign="top" align="center"><bold>F1 score</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1- EEG (not preprocessed, ML features, SVM)</td>
<td valign="top" align="center">0.62</td>
<td valign="top" align="center">0.60</td>
<td valign="top" align="center">0.59</td>
<td valign="top" align="center">0.59</td>
</tr> <tr>
<td valign="top" align="left">2- EEG (preprocessed, DL features, SVM &#x0002B; RF)</td>
<td valign="top" align="center">0.65</td>
<td valign="top" align="center">0.61</td>
<td valign="top" align="center">0.64</td>
<td valign="top" align="center">0.62</td>
</tr> <tr>
<td valign="top" align="left">3- EEG (preprocessed, ML &#x0002B; DL features, SVM &#x0002B; RF &#x0002B; DT)</td>
<td valign="top" align="center">0.74</td>
<td valign="top" align="center">0.70</td>
<td valign="top" align="center">0.75</td>
<td valign="top" align="center">0.72</td>
</tr> <tr>
<td valign="top" align="left">4- ET (not preprocessed, ML features, RF)</td>
<td valign="top" align="center">0.60</td>
<td valign="top" align="center">0.58</td>
<td valign="top" align="center">0.59</td>
<td valign="top" align="center">0.59</td>
</tr> <tr>
<td valign="top" align="left">5- ET (preprocessed, DL features, DT)</td>
<td valign="top" align="center">0.62</td>
<td valign="top" align="center">0.55</td>
<td valign="top" align="center">0.56</td>
<td valign="top" align="center">0.55</td>
</tr> <tr>
<td valign="top" align="left">6- ET (preprocessed, ML &#x0002B; DL features, XGB &#x0002B; RF &#x0002B; DT)</td>
<td valign="top" align="center">0.65</td>
<td valign="top" align="center">0.64</td>
<td valign="top" align="center">0.61</td>
<td valign="top" align="center">0.62</td>
</tr> <tr>
<td valign="top" align="left">7- EEG &#x00026; ET (not preprocessed, ML features, SVM)</td>
<td valign="top" align="center">0.72</td>
<td valign="top" align="center">0.71</td>
<td valign="top" align="center">0.67</td>
<td valign="top" align="center">0.68</td>
</tr> <tr>
<td valign="top" align="left">8- EEG &#x00026; ET (preprocessed, DL features, RF)</td>
<td valign="top" align="center">0.75</td>
<td valign="top" align="center">0.74</td>
<td valign="top" align="center">0.72</td>
<td valign="top" align="center">0.73</td>
</tr> <tr>
<td valign="top" align="left">9- EEG &#x00026; ET (preprocessed, ML &#x0002B; DL features, SVM &#x0002B; RF &#x0002B; XGB)</td>
<td valign="top" align="center">0.80</td>
<td valign="top" align="center">0.78</td>
<td valign="top" align="center">0.79</td>
<td valign="top" align="center">0.78</td>
</tr> <tr>
<td valign="top" align="left"><bold>Proposed- EEG &#x00026; ET (preprocessed, ML &#x0002B; DL features, stacking ensemble)</bold></td>
<td valign="top" align="center"><bold>0.84</bold></td>
<td valign="top" align="center"><bold>0.83</bold></td>
<td valign="top" align="center"><bold>0.84</bold></td>
<td valign="top" align="center"><bold>0.83</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Ablation study has been performed and bold values show the results obtained from final methodology.</p>
</table-wrap-foot>
</table-wrap>
<p><xref ref-type="table" rid="T6">Table 6</xref> represents the quantitative comparison of the employed methods, namely accuracy, precision rate, recall, and F1 score. <xref ref-type="fig" rid="F7">Figures 7</xref>&#x02013;<xref ref-type="fig" rid="F9">9</xref> show the evaluation score of the method, area under ROC curve, and Confusion Matrix of the proposed method. The ROC curve is a graphical approach that indicates a model&#x00027;s performance at different classification hurdles. It maps True Positive Rate or Sensitivity on the y-axis, against False Positive Rate or Fall out on the x-axis. An ideal ROC curve looks like a graph that plots the data close to the upper left-hand corner of the axes, which means that the performance of the model was satisfactory and it could distinguish between the classes accurately. The AUC gives overall performance of the ROC curve, from this the probability that the model ranks positive instance higher to a randomly chosen negative instance can be determined. Higher AUC shows that the tester has better ability in classifying. AUC-ROC of 0.89 has been achieved as shown in the <xref ref-type="fig" rid="F8">Figure 8</xref>, whereas, confusion matrix is presented in <xref ref-type="fig" rid="F9">Figure 9</xref> which further proves that it is highly effective when it comes to discriminating between the positive and the negative classes. This score can be classified within the &#x0201C;good&#x0201D; region; hence it can be deduced that the method purposed is good in segregating the two classes of interest.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Comparison of results obtained from proposed method with existing methods.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0007.tif"/>
</fig>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>ROC curve of proposed method.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0008.tif"/>
</fig>
<fig id="F9" position="float">
<label>Figure 9</label>
<caption><p>Confusion matrix of proposed method.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0009.tif"/>
</fig>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> compares the results obtained from proposed method with the existing state of the art methods. <xref ref-type="table" rid="T6">Table 6</xref> describes the evaluation criteria to different methods. As can be seen, the proposed method that uses the given preprocessing for EEG and ET data and incorporates the features of ML as well as DL within stacking ensemble provides the highest results on all of the listed measures.</p>
<p>At the highest accuracy, Method 1 employing raw EEG data with ML features and SVM yielded an accuracy 0.62. At the same time, the proposed method is much more effective with accuracy 0.84. It can be seen that this improvement is universality for precision, recall, and F1 score, more manifesting the advantages of data preprocessing and more successful attempt of the stacking ensemble method combining the ML &#x0002B; DL features. If we compare the methods in which preprocessing was used (e.g., Method 1) with those for which preprocessing was not used (e.g., Method 2), one can see that, in many cases, preprocessing has a positive effect on the performance. For instance, in method number 2, the EEG data is preprocessed and the DL features generates higher percentages of accuracy and recall than in method 1.</p>
<p>Methods which simultaneously utilize both EEG and ET data are superior to the methods based on only one type of data. For example, the feature that incorporates preprocessed EEG and ET data with conventional ML features and DL results in Method 9 has an accuracy of 0.80. This shows when there is an integration of the EEG and ET data it is able to provide better results for the model. The proposed method incorporates stacking ensemble, which also improves the performance of classifiers due to features adopted by this method. This leads to the highest values on all accounts, hence promoting a resilient and efficient model.</p>
<p>Classification results for EEG data when analyzed with the help of ML and DL incorporated with SVM, RF, and DT were seen to be quite satisfactory. By using the approach, the objective was attained with a 0.74 accuracy, and the precision, recall, and the F1 score equal to 0.70%, 0.75%, and 0.72% respectively. Furthermore, the have relatively high AUC, mean of 0.79 as shown in <xref ref-type="fig" rid="F10">Figure 10</xref> therefore support the reliability and discriminant capacity of the developed model, in the classification of consumers&#x00027; preferences from EEG signals. Altogether, these metrics can be discussed as demonstrating the efficiency of the proposed approach of applying the traditional ML algorithms alongside with the DL features.</p>
<fig id="F10" position="float">
<label>Figure 10</label>
<caption><p>ROC curve of proposed method (EEG only).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0010.tif"/>
</fig>
<p>On the other hand, the study that the referenced paper dealt with proposed a new deep learning decoder based on Riemannian Geometry and SPDNet structure (Georgiadis et al., <xref ref-type="bibr" rid="B16">2023a</xref>) for analyzing the signals of the NeuMa dataset of EEG. From the research, the investigators obtained a mean accuracy of 72%. <xref ref-type="table" rid="T7">Table 7</xref> shows a comparison between our proposed method(EEG Only) with a state-of-art method of Georgiadis et al. (<xref ref-type="bibr" rid="B16">2023a</xref>). The comparison of accuracies is shown in <xref ref-type="fig" rid="F11">Figure 11</xref>. Although this research infuses ML and DL with regular classifiers, the paper&#x00027;s presentation of domain&#x00027;s Riemannian Geometry and SPDNet demonstrates higher accuracy than conventional EEG- based approaches like Tangent Space SVM (Kalaganis et al., <xref ref-type="bibr" rid="B21">2019</xref>), EEG-Fusion (Hakim et al., <xref ref-type="bibr" rid="B18">2021</xref>) and R-kNN (Congedo et al., <xref ref-type="bibr" rid="B11">2017</xref>). The statistical significance thus obtained particularly with reference to the results achieved by Tangent Space SVM which was 67.72%, EEG-Fusion 52.75% and R-kNN was 51.96%.</p>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>Comparison with Georgiadis et al. (<xref ref-type="bibr" rid="B16">2023a</xref>)</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Aspect</bold></th>
<th valign="top" align="center"><bold>Georgiadis et al. (<xref ref-type="bibr" rid="B16">2023a</xref>)</bold></th>
<th valign="top" align="center"><bold>Proposed method (only EEG)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Dataset</td>
<td valign="top" align="center">NeuMa</td>
<td valign="top" align="center">NeuMa</td>
</tr> <tr>
<td valign="top" align="left">Accuracy</td>
<td valign="top" align="center">0.72</td>
<td valign="top" align="center">0.74</td>
</tr> <tr>
<td valign="top" align="left">Precision</td>
<td valign="top" align="center">Not mentioned</td>
<td valign="top" align="center">0.70</td>
</tr> <tr>
<td valign="top" align="left">Recall</td>
<td valign="top" align="center">Not mentioned</td>
<td valign="top" align="center">0.75</td>
</tr> <tr>
<td valign="top" align="left">F1-score</td>
<td valign="top" align="center">Not mentioned</td>
<td valign="top" align="center">0.72</td>
</tr> <tr>
<td valign="top" align="left">AUC score</td>
<td valign="top" align="center">Not mentioned</td>
<td valign="top" align="center">0.79</td>
</tr></tbody>
</table>
</table-wrap>
<fig id="F11" position="float">
<label>Figure 11</label>
<caption><p>Comparison of accuracy with other methods.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fncom-18-1516440-g0011.tif"/>
</fig>
<p>There are some limitations to the study that need to be noted. First off, although 42 participants is a small sample size, it might not be enough to extrapolate the results to a broader population. Furthermore, the findings are predicated on a particular dataset, which can restrict their generalizability to other product categories, markets, and cultural settings. The accuracy of the data acquired may be affected by the sampling rate and precision constraints of the EEG and eye-tracking sensors, despite their effectiveness. Furthermore, even though the used feature extraction strategies which combined manually created and automatically generated features proved successful, more research into different approaches or sophisticated deep learning architectures may enhance model performance. Finally, the integration of EEG and ET data adds complexity to the analysis, and potential synchronization challenges may have influenced the overall accuracy of the model.</p></sec>
<sec sec-type="conclusions" id="s6">
<title>6 Conclusions</title>
<p>Prediction of consumer preferences that we suggest is based on the machine learning and deep neural network methodology characterized by a high degree of accuracy and precision. These results could have been achieved because of correct preprocessing of images, use of the right features, and the high accuracy classifier. In preprocessing, we have increased the signal-to-noise ratio of EEG signals and ET data by removing noise and balanced the number of samples for classes, specifically the Buy class, by creating more through SMOTE. From the EEG and ET dataset, we created manual features by using the same method as before. Similarly, we used CNN-LSTM for the feature extraction of the selected EEG signals and LeNet-5 for the ET data. In classification, a most dependable stacking classifier was used for classification with a high level of accuracy.</p>
<p>The proposed method demonstrates stable results in the context of consumer preference prediction, though there are opportunities for future studies. For the current extraction feature, we could definitely do better in terms of advanced methodologies such as deep learning architectures or location of brain sources. Classification methods could be enhanced by considering other subject-dependent models or by developing the concept of a prediction. Generalizability is critical, which makes cross-validation mandatory across larger and more diverse data sets. Furthermore, it is necessary to discuss the similarities and differences of the proposed approach with other neuromarketing methods, as well as consider issues of the user&#x00027;s consent and data privacy. In addition, extending this method for uses outside of e-commerce, such as physical store promotion or measuring ad campaign effectiveness, provides more arenas for possible research and practical implementation.</p></sec>
</body>
<back>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>SU: Conceptualization, Methodology, Supervision, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. SK: Methodology, Resources, Supervision, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. AT: Data curation, Formal analysis, Methodology, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. AI: Supervision, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. MZ: Formal analysis, Validation, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing.</p>
</sec>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The author(s) declare that no Gen AI was used in the creation of this manuscript.</p></sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Akiba</surname> <given-names>T.</given-names></name> <name><surname>Sano</surname> <given-names>S.</given-names></name> <name><surname>Yanase</surname> <given-names>T.</given-names></name> <name><surname>Ohta</surname> <given-names>T.</given-names></name> <name><surname>Koyama</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Optuna: a next-generation hyperparameter optimization framework,&#x0201D;</article-title> in <source>Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery</source> &#x00026; <italic>Data Mining</italic> (New York, NY: ACM), <fpage>2623</fpage>&#x02013;<lpage>2631</lpage>. <pub-id pub-id-type="doi">10.1145/3292500.3330701</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aldayel</surname> <given-names>M.</given-names></name> <name><surname>Ykhlef</surname> <given-names>M.</given-names></name> <name><surname>Al-Nafjan</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>Recognition of consumer preference by analysis and classification EEG signals</article-title>. <source>Front. Hum. Neurosci</source>. <volume>14</volume>:<fpage>604639</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2020.604639</pub-id><pub-id pub-id-type="pmid">33519402</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Alimardani</surname> <given-names>M.</given-names></name> <name><surname>Kaba</surname> <given-names>M.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Deep learning for neuromarketing; classification of user preference using EEG signals,&#x0201D;</article-title> in <source>12th Augmented Human International Conference</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1145/3460881.3460930</pub-id><pub-id pub-id-type="pmid">35634118</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Arif</surname> <given-names>S.</given-names></name> <name><surname>Munawar</surname> <given-names>S.</given-names></name> <name><surname>Marie</surname> <given-names>R. R.</given-names></name> <name><surname>Shah</surname> <given-names>S. A.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Leveraging wavelets and deep CNN for sleep pattern recognition in road safety: an EEG study,&#x0201D;</article-title> in <source>International Conference on Recent Trends in Image Processing and Pattern Recognition</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>227</fpage>&#x02013;<lpage>241</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-53082-1_19</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Armstrong</surname> <given-names>G.</given-names></name> <name><surname>Adam</surname> <given-names>S.</given-names></name> <name><surname>Denize</surname> <given-names>S.</given-names></name> <name><surname>Kotler</surname> <given-names>P.</given-names></name></person-group> (<year>2014</year>). <source>Principles of Marketing</source>. <publisher-loc>Melbourne, VIC</publisher-loc>: <publisher-name>Pearson Australia</publisher-name>.</citation>
</ref>
<ref id="B6">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Assel</surname> <given-names>H.</given-names></name></person-group> (<year>1995</year>). <source>Consumer behavior and marketing action fifth edition.</source> <publisher-loc>Ohio</publisher-loc>: <publisher-name>South</publisher-name>.</citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Breiman</surname> <given-names>L.</given-names></name></person-group> (<year>2001</year>). <article-title>Random forests</article-title>. <source>Mach. Learn</source>. <volume>45</volume>, <fpage>5</fpage>&#x02013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chawla</surname> <given-names>N. V.</given-names></name> <name><surname>Bowyer</surname> <given-names>K. W.</given-names></name> <name><surname>Hall</surname> <given-names>L. O.</given-names></name> <name><surname>Kegelmeyer</surname> <given-names>W. P.</given-names></name></person-group> (<year>2002</year>). <article-title>Smote: synthetic minority over-sampling technique</article-title>. <source>J. Artif. Intell. Res</source>. <volume>16</volume>, <fpage>321</fpage>&#x02013;<lpage>357</lpage>. <pub-id pub-id-type="doi">10.1613/jair.953</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>T.</given-names></name> <name><surname>Guestrin</surname> <given-names>C.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Xgboost: a scalable tree boosting system,&#x0201D;</article-title> in <source>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>785</fpage>&#x02013;<lpage>794</lpage>. <pub-id pub-id-type="doi">10.1145/2939672.2939785</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Christoforou</surname> <given-names>C.</given-names></name> <name><surname>Papadopoulos</surname> <given-names>T. C.</given-names></name> <name><surname>Constantinidou</surname> <given-names>F.</given-names></name> <name><surname>Theodorou</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Your brain on the movies: a computational approach for predicting box-office performance from viewer&#x00027;s brain responses to movie trailers</article-title>. <source>Front. Neuroinform</source>. <volume>11</volume>:<fpage>72</fpage>. <pub-id pub-id-type="doi">10.3389/fninf.2017.00072</pub-id><pub-id pub-id-type="pmid">29311885</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Congedo</surname> <given-names>M.</given-names></name> <name><surname>Barachant</surname> <given-names>A.</given-names></name> <name><surname>Bhatia</surname> <given-names>R.</given-names></name></person-group> (<year>2017</year>). <article-title>Riemannian geometry for eeg-based brain-computer interfaces; a primer and a review</article-title>. <source>Brain-Comput. Interfaces</source> <volume>4</volume>, <fpage>155</fpage>&#x02013;<lpage>174</lpage>. <pub-id pub-id-type="doi">10.1080/2326263X.2017.1297192</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Fisch</surname> <given-names>B. J.</given-names></name></person-group> (<year>1999</year>). <source>Fisch and Spehlmann&#x00027;s EEG Primer: Basic Principles of Digital and Analog EEG</source>. <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Elsevier</publisher-name>.</citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friedman</surname> <given-names>J. H.</given-names></name></person-group> (<year>2001</year>). <article-title>Greedy function approximation: a gradient boosting machine</article-title>. <source>Ann. Stat</source>. <volume>29</volume>, <fpage>1189</fpage>&#x02013;<lpage>1232</lpage>. <pub-id pub-id-type="doi">10.1214/aos/1013203451</pub-id><pub-id pub-id-type="pmid">38281721</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a-Madariaga</surname> <given-names>J.</given-names></name> <name><surname>L&#x000F3;pez</surname> <given-names>M.-F. B.</given-names></name> <name><surname>Burgos</surname> <given-names>I. M.</given-names></name> <name><surname>Virto</surname> <given-names>N. R.</given-names></name></person-group> (<year>2019</year>). <article-title>Do isolated packaging variables influence consumers&#x00027; attention and preferences?</article-title> <source>Physiol. Behav</source>. <volume>200</volume>, <fpage>96</fpage>&#x02013;<lpage>103</lpage>. <pub-id pub-id-type="doi">10.1016/j.physbeh.2018.04.030</pub-id><pub-id pub-id-type="pmid">29702121</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Georgiadis</surname> <given-names>K.</given-names></name> <name><surname>Kalaganis</surname> <given-names>F. P.</given-names></name> <name><surname>Oikonomou</surname> <given-names>V. P.</given-names></name> <name><surname>Nikolopoulos</surname> <given-names>S.</given-names></name> <name><surname>Laskaris</surname> <given-names>N. A.</given-names></name> <name><surname>Kompatsiaris</surname> <given-names>I.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Rneumark: a riemannian EEG analysis framework for neuromarketing</article-title>. <source>Brain Inf</source>. <volume>9</volume>:<fpage>22</fpage>. <pub-id pub-id-type="doi">10.1186/s40708-022-00171-7</pub-id><pub-id pub-id-type="pmid">36112235</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Georgiadis</surname> <given-names>K.</given-names></name> <name><surname>Kalaganis</surname> <given-names>F. P.</given-names></name> <name><surname>Oikonomou</surname> <given-names>V. P.</given-names></name> <name><surname>Nikolopoulos</surname> <given-names>S.</given-names></name> <name><surname>Laskaris</surname> <given-names>N. A.</given-names></name> <name><surname>Kompatsiaris</surname> <given-names>I.</given-names></name> <etal/></person-group>. (<year>2023a</year>). <article-title>&#x0201C;Harneshing the potential of EEG in neuromarketing with deep learning and riemannian geometry,&#x0201D;</article-title> in <source>International Conference on Brain Informatics</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>21</fpage>&#x02013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-43075-6_3</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Georgiadis</surname> <given-names>K.</given-names></name> <name><surname>Kalaganis</surname> <given-names>F. P.</given-names></name> <name><surname>Riskos</surname> <given-names>K.</given-names></name> <name><surname>Matta</surname> <given-names>E.</given-names></name> <name><surname>Oikonomou</surname> <given-names>V. P.</given-names></name> <name><surname>Yfantidou</surname> <given-names>I.</given-names></name> <etal/></person-group>. (<year>2023b</year>). <article-title>Neuma-the absolute neuromarketing dataset en route to an holistic understanding of consumer behaviour</article-title>. <source>Sci. Data</source> <volume>10</volume>:<fpage>508</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-023-02392-9</pub-id><pub-id pub-id-type="pmid">37537187</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hakim</surname> <given-names>A.</given-names></name> <name><surname>Klorfeld</surname> <given-names>S.</given-names></name> <name><surname>Sela</surname> <given-names>T.</given-names></name> <name><surname>Friedman</surname> <given-names>D.</given-names></name> <name><surname>Shabat-Simon</surname> <given-names>M.</given-names></name> <name><surname>Levy</surname> <given-names>D. J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Machines learn neuromarketing: improving preference prediction from self-reports using multiple EEG measures and machine learning</article-title>. <source>Int. J. Res. Mark</source>. <volume>38</volume>, <fpage>770</fpage>&#x02013;<lpage>791</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijresmar.2020.10.005</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hochreiter</surname> <given-names>S.</given-names></name> <name><surname>Schmidhuber</surname> <given-names>J.</given-names></name></person-group> (<year>1997</year>). <article-title>Long short-term memory</article-title>. <source>Neural Comput</source>. <volume>9</volume>, <fpage>1735</fpage>&#x02013;<lpage>1780</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id><pub-id pub-id-type="pmid">9377276</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Homan</surname> <given-names>R.</given-names></name> <name><surname>Herman</surname> <given-names>J.</given-names></name> <name><surname>Purdy</surname> <given-names>P.</given-names></name></person-group> (<year>1987</year>). <article-title>Cerebral location of international 10-20 system electrode placement</article-title>. <source>Electroencephalogr. Clin. Neurophysiol</source>. <volume>66</volume>, <fpage>376</fpage>&#x02013;<lpage>382</lpage>. <pub-id pub-id-type="doi">10.1016/0013-4694(87)90206-9</pub-id><pub-id pub-id-type="pmid">2435517</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kalaganis</surname> <given-names>F. P.</given-names></name> <name><surname>Laskaris</surname> <given-names>N. A.</given-names></name> <name><surname>Chatzilari</surname> <given-names>E.</given-names></name> <name><surname>Nikolopoulos</surname> <given-names>S.</given-names></name> <name><surname>Kompatsiaris</surname> <given-names>I.</given-names></name></person-group> (<year>2019</year>). <article-title>A Riemannian geometry approach to reduced and discriminative covariance estimation in brain computer interfaces</article-title>. <source>IEEE Trans. Biomed. Eng</source>. <volume>67</volume>, <fpage>245</fpage>&#x02013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2019.2912066</pub-id><pub-id pub-id-type="pmid">30998456</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Khushaba</surname> <given-names>R. N.</given-names></name> <name><surname>Wise</surname> <given-names>C.</given-names></name> <name><surname>Kodagoda</surname> <given-names>S.</given-names></name> <name><surname>Louviere</surname> <given-names>J.</given-names></name> <name><surname>Kahn</surname> <given-names>B. E.</given-names></name> <name><surname>Townsend</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Consumer neuroscience: assessing the brain response to marketing stimuli using electroencephalogram (EEG) and eye tracking</article-title>. <source>Expert Syst. Appl</source>. <volume>40</volume>, <fpage>3803</fpage>&#x02013;<lpage>3812</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2012.12.095</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kohavi</surname> <given-names>R.</given-names></name></person-group> (<year>1995</year>). <article-title>&#x0201C;A study of cross-validation and bootstrap for accuracy estimation and model selection,&#x0201D;</article-title> in <source>Proceedings of the 14th international joint conference on Artificial intelligence, Volume 2</source> (<publisher-loc>Princeton, NJ</publisher-loc>: <publisher-name>Citeseer</publisher-name>), <fpage>1137</fpage>&#x02013;<lpage>1143</lpage>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kumar</surname> <given-names>S.</given-names></name> <name><surname>Yadava</surname> <given-names>M.</given-names></name> <name><surname>Roy</surname> <given-names>P. P.</given-names></name></person-group> (<year>2019</year>). <article-title>Fusion of EEG response and sentiment analysis of products review to predict customer satisfaction</article-title>. <source>Inf. Fusion</source> <volume>52</volume>, <fpage>41</fpage>&#x02013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2018.11.001</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>LeCun</surname> <given-names>Y.</given-names></name> <name><surname>Bottou</surname> <given-names>L.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Haffner</surname> <given-names>P.</given-names></name></person-group> (<year>1998</year>). <article-title>Gradient-based learning applied to document recognition</article-title>. <source>Proc. IEEE</source> <volume>86</volume>, <fpage>2278</fpage>&#x02013;<lpage>2324</lpage>. <pub-id pub-id-type="doi">10.1109/5.726791</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mashrur</surname> <given-names>F. R.</given-names></name> <name><surname>Rahman</surname> <given-names>K. M.</given-names></name> <name><surname>Miya</surname> <given-names>M. T. I.</given-names></name> <name><surname>Vaidyanathan</surname> <given-names>R.</given-names></name> <name><surname>Anwar</surname> <given-names>S. F.</given-names></name> <name><surname>Sarker</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Intelligent neuromarketing framework for consumers&#x00027; preference prediction from electroencephalography signals and eye tracking</article-title>. <source>J. Consum. Behav</source>. <volume>23</volume>, <fpage>1146</fpage>&#x02013;<lpage>1157</lpage>. <pub-id pub-id-type="doi">10.1002/cb.2253</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matukin</surname> <given-names>M.</given-names></name> <name><surname>Ohme</surname> <given-names>R.</given-names></name> <name><surname>Boshoff</surname> <given-names>C.</given-names></name></person-group> (<year>2016</year>). <article-title>Toward a better understanding of advertising stimuli processing: exploring the link between consumers&#x00027; eye fixation and their subconscious responses</article-title>. <source>J. Advert. Res</source>. <volume>56</volume>, <fpage>205</fpage>&#x02013;<lpage>216</lpage>. <pub-id pub-id-type="doi">10.2501/JAR-2016-017</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Murugappan</surname> <given-names>M.</given-names></name> <name><surname>Murugappan</surname> <given-names>S.</given-names></name> <name><surname>Gerard</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>&#x0201C;Wireless EEG signals based neuromarketing system using Fast Fourier transform (FFT),&#x0201D;</article-title> in <source>2014 IEEE 10th international colloquium on signal processing and its applications</source> (<publisher-loc>Kuala Lumpur</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>25</fpage>&#x02013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1109/CSPA.2014.6805714</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Peter</surname> <given-names>J. P.</given-names></name> <name><surname>Olson</surname> <given-names>J. C.</given-names></name> <name><surname>Grunert</surname> <given-names>K. G.</given-names></name></person-group> (<year>1999</year>). <source>Consumer Behaviour and Marketing Strategy. Edisi IV</source>. <publisher-loc>Jakarta</publisher-loc>: <publisher-name>PT Erlangga</publisher-name>.</citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Samsuri</surname> <given-names>N.</given-names></name> <name><surname>Begum</surname> <given-names>T.</given-names></name> <name><surname>Reza</surname> <given-names>F.</given-names></name> <name><surname>Omar</surname> <given-names>H.</given-names></name> <name><surname>Yusoff</surname> <given-names>N.</given-names></name> <name><surname>Idris</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Left lateral singular (LLS) view enhances attention at stimulus discrimination during automobile display setting revealed by EEG/ERP and eye tracking technique</article-title>. <source>Int. J. Enhanc. Res. Sci. Technol. Eng</source>. <volume>5</volume>, <fpage>1</fpage>&#x02013;<lpage>9</lpage>.</citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shah</surname> <given-names>S. M. A.</given-names></name> <name><surname>Usman</surname> <given-names>S. M.</given-names></name> <name><surname>Khalid</surname> <given-names>S.</given-names></name> <name><surname>Rehman</surname> <given-names>I. U.</given-names></name> <name><surname>Anwar</surname> <given-names>A.</given-names></name> <name><surname>Hussain</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>An ensemble model for consumer emotion prediction using EEG signals for neuromarketing applications</article-title>. <source>Sensors</source> <volume>22</volume>:<fpage>9744</fpage>. <pub-id pub-id-type="doi">10.3390/s22249744</pub-id><pub-id pub-id-type="pmid">36560113</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Slanzi</surname> <given-names>G.</given-names></name> <name><surname>Balazs</surname> <given-names>J. A.</given-names></name> <name><surname>Vel&#x000E1;squez</surname> <given-names>J. D.</given-names></name></person-group> (<year>2017</year>). <article-title>Combining eye tracking, pupil dilation and EEG analysis for predicting web users click intention</article-title>. <source>Inf. Fusion</source> <volume>35</volume>, <fpage>51</fpage>&#x02013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2016.09.003</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Telpaz</surname> <given-names>A.</given-names></name> <name><surname>Webb</surname> <given-names>R.</given-names></name> <name><surname>Levy</surname> <given-names>D. J.</given-names></name></person-group> (<year>2015</year>). <article-title>Using EEG to predict consumers&#x00027; future choices</article-title>. <source>J. Mark. Res</source>. <volume>52</volume>, <fpage>511</fpage>&#x02013;<lpage>529</lpage>. <pub-id pub-id-type="doi">10.1509/jmr.13.0564</pub-id><pub-id pub-id-type="pmid">11670861</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Teplan</surname> <given-names>M.</given-names></name></person-group> (<year>2002</year>). <article-title>Fundamentals of EEG measurement</article-title>. <source>Meas. Sci. Rev</source>. <volume>2</volume>, <fpage>1</fpage>&#x02013;<lpage>11</lpage>.</citation>
</ref>
<ref id="B35">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Tobii</surname> <given-names>A. B.</given-names></name></person-group> (<year>2024</year>). <source>What is eye tracking</source>? Available at: <ext-link ext-link-type="uri" xlink:href="https://www.tobii.com/learn-and-support/get-started/what-is-eye-tracking">https://www.tobii.com/learn-and-support/get-started/what-is-eye-tracking</ext-link> (accessed December 30, 2024).</citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vecchiato</surname> <given-names>G.</given-names></name> <name><surname>Astolfi</surname> <given-names>L.</given-names></name> <name><surname>Fallani</surname> <given-names>F. D. V.</given-names></name> <name><surname>Toppi</surname> <given-names>J.</given-names></name> <name><surname>Aloise</surname> <given-names>F.</given-names></name> <name><surname>Bez</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2011</year>). <article-title>On the use of EEG or MEG brain imaging tools in neuromarketing research</article-title>. <source>Comput. Intell. Neurosci</source>. <volume>2011</volume>:<fpage>643489</fpage>. <pub-id pub-id-type="doi">10.1155/2011/643489</pub-id><pub-id pub-id-type="pmid">21960996</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Welch</surname> <given-names>P.</given-names></name></person-group> (<year>1967</year>). <article-title>The use of Fast Fourier Transform for the estimation of power spectra: a method based on time averaging over short, modified periodograms</article-title>. <source>IEEE Trans. Audio Electroacoust</source>. <volume>15</volume>, <fpage>70</fpage>&#x02013;<lpage>73</lpage>. <pub-id pub-id-type="doi">10.1109/TAU.1967.1161901</pub-id></citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wolpert</surname> <given-names>D. H.</given-names></name></person-group> (<year>1992</year>). <article-title>Stacked generalization</article-title>. <source>Neural. Netw</source>. <volume>5</volume>, <fpage>241</fpage>&#x02013;<lpage>259</lpage>. <pub-id pub-id-type="doi">10.1016/S0893-6080(05)80023-1</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yadava</surname> <given-names>M.</given-names></name> <name><surname>Kumar</surname> <given-names>P.</given-names></name> <name><surname>Saini</surname> <given-names>R.</given-names></name> <name><surname>Roy</surname> <given-names>P. P.</given-names></name> <name><surname>Prosad Dogra</surname> <given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>Analysis of EEG signals and its application to neuromarketing</article-title>. <source>Multimed. Tools Appl</source>. <volume>76</volume>, <fpage>19087</fpage>&#x02013;<lpage>19111</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-017-4580-6</pub-id></citation>
</ref>
</ref-list>
</back>
</article>
