<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Virtual Real.</journal-id>
<journal-title>Frontiers in Virtual Reality</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Virtual Real.</abbrev-journal-title>
<issn pub-type="epub">2673-4192</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1616442</article-id>
<article-id pub-id-type="doi">10.3389/frvir.2025.1616442</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Virtual Reality</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Neuroadaptive haptics: a proof-of-concept comparing reinforcement learning from explicit ratings and neural signals for adaptive XR systems</article-title>
<alt-title alt-title-type="left-running-head">Gehrke et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frvir.2025.1616442">10.3389/frvir.2025.1616442</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Gehrke</surname>
<given-names>Lukas</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/659859/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Koselevs</surname>
<given-names>Aleksandrs</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3067598/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Klug</surname>
<given-names>Marius</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1089715/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Gramann</surname>
<given-names>Klaus</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/5473/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Department of Biopsychology and Neuroergonomics</institution>, <institution>Technische Universit&#xe4;t Berlin</institution>, <addr-line>Berlin</addr-line>, <country>Germany</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Young Investigator Group &#x2013; Intuitive XR</institution>, <institution>Brandenburg University of Technology Cottbus-Senftenberg</institution>, <addr-line>Cottbus</addr-line>, <country>Germany</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/628824/overview">Christos Mousas</ext-link>, Purdue University, United States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3048413/overview">Nikitha Donekal Chandrashekar</ext-link>, Virginia Tech, United States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3049625/overview">Zhanxun Dong</ext-link>, Shanghai Jiao Tong University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3049801/overview">Apostolos Vrontos</ext-link>, RWTH Aachen University, Germany</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Lukas Gehrke, <email>lukas.gehrke@tu-berlin.de</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>11</day>
<month>08</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>6</volume>
<elocation-id>1616442</elocation-id>
<history>
<date date-type="received">
<day>22</day>
<month>04</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>24</day>
<month>07</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Gehrke, Koselevs, Klug and Gramann.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Gehrke, Koselevs, Klug and Gramann</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Neuroadaptive technology provides a promising path to enhancing immersive extended reality (XR) experiences by dynamically tuning multisensory feedback to user preferences. This study introduces a novel system employing reinforcement learning (RL) to adapt haptic rendering in XR environments based on user feedback derived either explicitly from user ratings or implicitly from neural signals measured via Electroencephalography (EEG).</p>
</sec>
<sec>
<title>Methods</title>
<p>Participants interacted with virtual objects in a VR environment and rated their experience using a traditional questionnaire while their EEG data were recorded. Then, in two RL conditions, an RL agent tried to tune the haptics to the user &#x2014; learning either on the rewards from explicit ratings, or on implicit neural signals decoded from EEG.</p>
</sec>
<sec>
<title>Results</title>
<p>The neural decoder achieved a mean F1 score of 0.8, supporting informative yet noisy classification. Exploratory analyses revealed instability in the RL agent&#x2019;s behavior in both explicit and implicit feedback conditions.</p>
</sec>
<sec>
<title>Discussion</title>
<p>A limited number of interaction steps likely constrained exploration and contributed to convergence instability. Revisiting the interaction design to support more frequent sampling may improve robustness to EEG noise and mitigate drifts in subjective experience. By demonstrating RL&#x2010;based adaptation from implicit neural signals, our proof-of-concept is a step towards seamless, low-friction personalization in XR.</p>
</sec>
</abstract>
<kwd-group>
<kwd>human-computer interaction</kwd>
<kwd>reinforcement learning</kwd>
<kwd>RLHF</kwd>
<kwd>brain-computer interface</kwd>
<kwd>EEG</kwd>
<kwd>error detection</kwd>
<kwd>neuroadaptive technology</kwd>
</kwd-group>
<contract-num rid="cn001">GR 2627/13-1</contract-num>
<contract-sponsor id="cn001">Deutsche Forschungsgemeinschaft<named-content content-type="fundref-id">10.13039/501100001659</named-content>
</contract-sponsor>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Technologies for VR</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Extended Reality (XR) has the potential to create profoundly immersive and awe-inspiring experiences. However, achieving an optimal experience requires fine-tuning various settings, from brightness and field of view to haptic feedback (<xref ref-type="bibr" rid="B57">Ramsamy et al., 2006</xref>) and spatial audio (<xref ref-type="bibr" rid="B54">Potter et al., 2022</xref>). Currently, users manually adjust these parameters to their liking through conventional menu interfaces that closely resemble traditional desktop environments.</p>
<p>Unfortunately, this introduces significant friction. Frequent interruptions, particularly during initial setup, can break immersion, reduce excitement, and potentially lower long-term adoption rates. Additionally, conventional settings menus may carry a higher cost than just disrupting the immediate experience; they reposition the user into a known, age-old computing paradigm that is entirely disconnected from the immersive nature of XR. This disconnect makes personalization feel like a chore rather than a seamless and intuitive part of the experience.</p>
<p>Given these challenges, we set out to develop a method that effectively personalizes XR experiences while minimizing manual configuration and preserving immersion. One promising approach is to leverage Reinforcement Learning (RL), empowering an autonomous system to learn user preferences over time (<xref ref-type="bibr" rid="B25">Kaufmann et al., 2023</xref>). However, this presents its own obstacles, such as the need for human-provided labels and the challenge of balancing automation with user control.</p>
<p>One solution lies in obtaining implicit feedback from the user through neural and physiological data (<xref ref-type="bibr" rid="B70">Zander et al., 2016</xref>). Instead of relying on explicit user input, neural signals can serve as a real-time indicator of a user&#x2019;s preferences, engagement, and immersion. Here, the term &#x2018;neuroadaptive technology&#x2019; is used to mark the shift from &#x2018;direct control&#x2019; brain-computer interfaces (BCI) to implicit adaptation (<xref ref-type="bibr" rid="B70">Zander et al., 2016</xref>; <xref ref-type="bibr" rid="B33">Krol and Zander, 2022</xref>).</p>
<p>We introduce &#x2018;Neuroadaptive Haptics&#x2019;, a novel neuroadaptive system for a tailor-made, multisensory XR experience. &#x2018;Neuroadaptive Haptics&#x2019; is an interactive system that integrates real-time neural and physiological data to dynamically modify haptics in XR environments. Our system leverages the output from a BCI as a reward signal for RL. In this paper, we applied our system to tune haptic parameters of a virtual reality (VR) system. We tested whether the system was able to dynamically adjust the VR settings to optimize users&#x2019; haptic experience over time without requiring manual interventions.</p>
</sec>
<sec id="s2">
<title>2 Related work</title>
<p>Our research draws inspiration from neuroscience and engineering work on BCIs, specifically neuroadaptive technology. In order to situate our research, we provide a background on haptic experiences in XR.</p>
<sec id="s2-1">
<title>2.1 Haptic experiences in VR</title>
<p>Haptic feedback in VR has been shown to be a key component in creating a realistic user experience. In fact, for a long time now, researchers have argued that attaining haptic realism is the next grand challenge in virtual reality (<xref ref-type="bibr" rid="B7">Brooks, 1999</xref>). In most use cases, the goal of haptic devices is to render (i.e., generate artificial touch sensations through devices) realistic sensations that mimic the sensory experience a user would normally expect when interacting with the real world. For instance, multisensory haptic renderings can combine vibrotactile feedback with force feedback, rendered via exoskeletons or electrical muscle stimulation (EMS), to simulate not only the sensation of touch but also the resistance and rigidity of objects (<xref ref-type="bibr" rid="B37">Lopes et al., 2015</xref>). Additionally, other sensory cues&#x2013;such as carefully synchronized auditory feedback or subliminal EMS signals&#x2013;are increasingly employed to enhance haptic illusions (<xref ref-type="bibr" rid="B9">Cho et al., 2024</xref>; <xref ref-type="bibr" rid="B63">Takahashi et al., 2024</xref>). These methods can trick the brain into perceiving properties like texture, weight, or even the subtleties of material composition, by engaging multiple sensory pathways simultaneously. However, complex haptic interactions are still error-prone. The synchrony of sensory information relies on the quality of motion tracking and the accuracy of feedback presentation, and incongruous temporal feedback during object interaction may occur due to technical reasons.</p>
<p>At its core, this relies on predictive coding mechanisms underlying sensory integration&#x2013;a framework in which our brains leverage foundational models (<xref ref-type="bibr" rid="B55">Pouget et al., 2013</xref>), originally established through interactions with the physical world, to interpret sensory information in both real and virtual environments (see next section). However, as evidenced by the significant technological leaps in each new hardware release, next-generation VR technology still exhibits frequent glitches and sensory mismatches, especially in key moments of multisensory integration.</p>
</sec>
<sec id="s2-2">
<title>2.2 Sampling the predictive coding mechanisms in the brain</title>
<p>The brain is frequently conceptualized as a predictive system that continuously generates foundational models of the environment to infer the causes of sensory input (<xref ref-type="bibr" rid="B58">Rao and Ballard, 1999</xref>; <xref ref-type="bibr" rid="B14">Friston, 2010</xref>); <xref ref-type="bibr" rid="B10">Clark (2013)</xref>. In this framework, perception emerges from an iterative process in which predictions are compared with incoming sensory data, and discrepancies&#x2013;known as prediction errors&#x2013;drive model updates. These processes have been widely studied in sensory perception (<xref ref-type="bibr" rid="B4">Bastos et al., 2012</xref>; <xref ref-type="bibr" rid="B26">Keller and Mrsic-Flogel, 2018</xref>; <xref ref-type="bibr" rid="B28">Knill and Pouget, 2004</xref>), with research showing that the brain dynamically adjusts its internal representations to minimize these errors.</p>
<p>Prediction errors are particularly crucial in interactive and multimodal contexts, where sensorimotor integration plays a key role. Previous studies have demonstrated that the brain detects visuo-haptic mismatches in real-time, measurable as Prediction Error Negativity (PEN) in Electroencephalogram (EEG) responses. This neural signature has been targeted as a marker of error processing in VR scenarios, where prediction errors appear to correlate with disruptions in user experience and physical immersion (<xref ref-type="bibr" rid="B16">Gehrke et al., 2019</xref>; <xref ref-type="bibr" rid="B61">Singh et al., 2018</xref>; <xref ref-type="bibr" rid="B60">Si-mohammed et al., 2020</xref>; <xref ref-type="bibr" rid="B17">Gehrke et al., 2022</xref>; <xref ref-type="bibr" rid="B18">2024</xref>). These findings suggest that predictive processing extends beyond passive perception and is deeply embedded in embodied cognition, where action and perception are considered to be tightly coupled.</p>
</sec>
<sec id="s2-3">
<title>2.3 Neuroadaptive technologies</title>
<p>Neuroadaptive systems are interactive technologies that adapt their behavior based on real-time neural or physiological activity (<xref ref-type="bibr" rid="B20">Hettinger et al., 2003</xref>). These systems aim to respond to internal user states by dynamically modifying the interface or the environment. A system can be considered neuroadaptive when it includes a closed-loop in which neural or physiological signals are used not just for passive monitoring, but for actively shaping the user experience. Some examples include adaptive interfaces that change based on mental workload (<xref ref-type="bibr" rid="B11">Dehais et al., 2020</xref>) BCI-driven cursor control (<xref ref-type="bibr" rid="B70">Zander et al., 2016</xref>), as well as neurofeedback systems in learning and rehabilitation contexts (<xref ref-type="bibr" rid="B40">Mahmoudi et al., 2025</xref>).</p>
<p>Most applications to date have focused on desktop-based scenarios, where the challenges related to signal stability, real-time processing, and interface control are more manageable. Fewer studies have explored neuroadaptive approaches in XR, largely due to the difficulty of integrating physiological sensing in dynamic, multisensory environments. Still, early XR applications have demonstrated promise in areas like meditation support (<xref ref-type="bibr" rid="B30">Kosunen et al., 2016</xref>), exposure therapy for phobias (<xref ref-type="bibr" rid="B66">Weber et al., 2024</xref>), and adaptive training systems (<xref ref-type="bibr" rid="B43">Mark et al., 2022</xref>). However, these implementations generally rely on predefined scenarios and offer only limited autonomy to the computer, which typically cannot decide when or how to seek additional information from the user. Moving toward truly integrated neuroadaptive XR requires empowering the system to autonomously probe the user when necessary&#x2013;sampling new data points to improve adaptation (<xref ref-type="bibr" rid="B32">Krol et al., 2020</xref>).</p>
<p>Looking beyond the scope of EEG-based measurements, other physiological data can be used for adaptive XR systems as well. For example, eye tracking signals have been explored as a measure for cognitive load in VR industrial training (<xref ref-type="bibr" rid="B48">Nasri et al., 2024</xref>), functional near-infrared spectroscopy (fNIRS) has been used to modulate scene intensity in a real-time VR horror experience (<xref ref-type="bibr" rid="B5">Berger et al., 2024</xref>), cardiac measures have guided relaxation content in immersive nature environments (<xref ref-type="bibr" rid="B56">Pratviel et al., 2024</xref>), and adaptive surface-EMG gesture decoders during VR object manipulation improved task success and reduced workload compared to static decoders (<xref ref-type="bibr" rid="B15">Gagn&#xe9; et al., 2025</xref>). Peripheral modalities (measuring the eyes, heart, or muscles) have strong potential, but their disadvantage is clear: They cannot access central nervous system information, i.e., the neural basis of cognition and stimulus processing and evaluation. Functional NIRS can do this, but has a temporal disadvantage compared to EEG, as it relies on the hemodynamic response, which lags several seconds behind the actual processing (<xref ref-type="bibr" rid="B44">Mehta and Parasuraman, 2013</xref>). Since we are interested in the direct neural responses of predictive processing, we relied on EEG in our study.</p>
<p>In this work, we configured an RL agent to autonomously sample the human-in-the-loop, aiming to find a haptic feedback configuration that the user experiences as the most consistent with their expectations. This approach moves beyond traditional neuroadaptive applications by combining passive physiological sensing with active decision-making in one interactive prototype, paving the way for more intelligent and responsive XR experiences. To tune the interaction over time, we used feedback, or labels, given directly by the human-in-the-loop, a special form of RL from human feedback (RLHF).</p>
</sec>
<sec id="s2-4">
<title>2.4 Reinforcement learning from human feedback</title>
<p>RLHF is a paradigm that enhances traditional reinforcement learning (RL) by incorporating human evaluative signals into the learning process (<xref ref-type="bibr" rid="B25">Kaufmann et al., 2023</xref>; <xref ref-type="bibr" rid="B29">Knox, 2011</xref>; <xref ref-type="bibr" rid="B36">Li et al., 2019</xref>. Instead of relying solely on a predefined reward function, RLHF allows human feedback&#x2013;either explicit, i.e., numerical ratings, scores, and rankings, or implicit, i.e., physiological signals such as EEG-based brain activity (<xref ref-type="bibr" rid="B39">Luo et al., 2018</xref>; <xref ref-type="bibr" rid="B68">Xavier Fid&#xea;ncio et al., 2022</xref>)&#x2013;to shape an agent&#x2019;s behavior dynamically. This approach is particularly useful in domains where reward functions are difficult to specify, such as in robotic interaction, as well as user experience optimization as it applies in adaptive XR environments.</p>
<p>Traditional RL systems require extensive exploration to learn optimal behaviors, which can be time-consuming and inefficient. RLHF mitigates this by enabling systems to leverage human expertise, reducing the sample complexity of learning tasks. For instance, prior work has demonstrated the effectiveness of human preference-based RL for training AI assistants, robotic control, and interactive game agents (<xref ref-type="bibr" rid="B36">Li et al., 2019</xref>).</p>
<p>A key benefit of computing rewards based on EEG-based brain activity in this context is that it enables a seamless and non-disruptive interaction: the system can adapt to the user without requiring them to stop and provide explicit input. In contrast, while explicit labels might offer more reliable signals, they introduce cognitive load and interrupt the immersive flow of the experience. EEG-based approaches address this limitation by allowing the system to adapt in the background, minimizing disruptions while still responding to changes in the user&#x2019;s internal states. Studies on EEG-based RLHF have shown that BCIs can provide real-time feedback signals that improve learning efficiency while reducing human effort (<xref ref-type="bibr" rid="B69">Xu et al., 2021</xref>; <xref ref-type="bibr" rid="B68">Xavier Fid&#xea;ncio et al., 2022</xref>).</p>
</sec>
</sec>
<sec id="s3">
<title>3 User study and methods</title>
<p>We set out to answer three questions: First, can we tune haptic rendering to participants&#x2019; preferences using an RL agent based on human feedback? Second, is this possible through implicit labels obtained through a neural decoder? And third, are there disadvantages when relying on implicit instead of explicit labels?</p>
<p>To investigate these questions, we designed a user study where participants performed a pick-and-place task in VR: they had to pick up virtual objects and move them to designated locations. During object pick-up, visual, auditory, and haptic feedback were systematically varied to create different combinations of sensory cues. In an initial recording session, participants provided labels about the interaction through answering a question. This labeled data was then used to train a neural decoder of expected haptic sensations. Next, participants completed two blocks in which an RL agent tried to predict the haptic feedback participants deemed to best match their expectation. In one block, the RL agent operated on participants&#x2019; explicit scores on the questionnaire, and in the other, it operated on the (implicit) output of the neural decoder.</p>
<sec id="s3-1">
<title>3.1 Participants</title>
<p>14 participants (M &#x3d; 29 years, SD &#x3d; 5.2) were recruited from our local institution and through the online participant pool of the institute. Nine participants self-identified as women, 5 as men. All were right-handed (self-identification, no test). Participants were compensated with course credit or 12 &#x20ac; per hour of study participation. Before participating, they were informed of the nature of the experiment, recording, and anonymization procedures, and signed a consent form. The project was approved by the local ethics committee of the Department of Psychology and Ergonomics at the TU Berlin (Ethics protocol approval code: BPN_GEH_220421).</p>
<p>The first five participants only completed the <italic>training</italic> part of the experiment, which means that they did not complete the two additional blocks with the RL agent. One participant in the group who completed the blocks with the RL agent had to be excluded from any analyses concerning the agent, since they did not complete at least one of the experimental blocks due to technical issues with the EEG recording hardware. Hence, statistics for all analyses about the agent were computed for eight participants and all other analyses were computed for the full sample of 14 participants.</p>
</sec>
<sec id="s3-2">
<title>3.2 Apparatus</title>
<p>The experimental setup, depicted in <xref ref-type="fig" rid="F1">Figure 1</xref>, comprised: (1) a VR headset with a built-in eye tracker, (2) a haptic glove with an attached motion tracker, (3) a 64-channel EEG system, and a VR capable computer (CPU: AMD Ryzen 5 5600X, GPU: AMD RADEON RX 6600 XT 8GB). To assist readers in replicating our experiment, we provide the necessary technical details, the complete source code for the VR experiment, the collected data, and the analysis scripts<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Experimental setup featuring a participant equipped with EEG (BrainProducts), Head-mounted display (VIVE Pro Eye), haptic feedback (SenseGlove Nova), motion capture (VIVE Tracker), and audio output (earcup speaker).</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g001.tif">
<alt-text content-type="machine-generated">A person is wearing a headset with EEG sensors labeled &#x22;BrainProducts,&#x22; a VR device labeled &#x22;VIVE Pro Eye,&#x22; and headphones with a &#x22;Earcup Speaker.&#x22; They are using a haptic glove labeled &#x22;SenseGlove Nova&#x22; with a &#x22;VIVE Tracker&#x22; for motion capture.</alt-text>
</graphic>
</fig>
<p>(1) VR. We used an HTC Vive Pro Eye 1 headset (HTC Corporation, Taiwan) to display the scene and sampled eye-tracking using SRanipal (Tobii AB, version Core SW 2.16.4.67). We replaced the stock strap of the headset with the Vive Deluxe Audio Strap to ensure a good fit and reduce discomfort potentially caused by the EEG cap.</p>
<p>(2) Haptic Glove. The SenseGlove Nova V1.0 (SenseGlove, Netherlands) was used to sample finger movements and render vibro-tactile sensations. To track hand movements, an HTC Vive tracker was attached to the glove, as recommended by the manufacturer.</p>
<p>(3) EEG Setup. EEG data was recorded from 64 actively amplified wet electrodes using BrainAmp DC amplifiers (BrainProducts, Germany) with a high-pass filter set at 0.016Hz. Electrodes were placed according to the 10-system (<xref ref-type="bibr" rid="B8">Chatrian et al., 1985</xref>). One electrode was placed under the right eye to provide additional information about eye movements (vEOG). After fitting the cap, all electrodes were filled with conductive gel to ensure proper conductivity, and electrode impedance was brought below 10k<inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a9;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> where possible. EEG data was recorded with a sampling rate of 250&#xa0;Hz. The reference electrode was &#x2018;FCz&#x2019; and the ground electrode &#x2018;AFz&#x2019;.</p>
<p>Motion data of the hand and eye-gaze was streamed using custom scripts of &#x2018;labstreaminglayer&#x2019; (LSL) (<xref ref-type="bibr" rid="B31">Kothe et al., 2024</xref>). Additionally, EEG data and an experiment marker stream that marked sections of the study procedure were streamed using LSL. LSL&#x2019;s LabRecorder was used to collect all data streams with timestamps.</p>
</sec>
<sec id="s3-3">
<title>3.3 Task</title>
<p>In the task, participants were instructed to pick up an object and then place it in a target location. They were instructed to be accurate while maintaining a steady pace. The object was placed on a table in front of them and for grabbing they used the grab functionality of the haptic glove. Successful grabbing required them to use all fingers, ensuring that both the thumb and at least one other finger securely held the object. After picking up the object, the participants moved it to the center of a semi-transparent sphere that visually indicated the location of the goal. Once the object was released, participants received feedback about their placement accuracy, displayed as a numerical value in centimeters on the table, indicating the distance between the object&#x2019;s placement and the center of the goal sphere.</p>
<p>Depending on the trial condition (see procedure below), participants were then asked to rate their experience concerning the prompt &#x201c;My experience in the virtual environment seemed consistent with my experiences in the real world.&#x201d; (translated) which was chosen from the <italic>Multimodal Presence Scale</italic> (<xref ref-type="bibr" rid="B42">Makransky et al., 2017</xref>). The anchors of the prompt were &#x201c;completely disagree&#x201d; and &#x201c;strongly agree&#x201d;. To give their answer, participants could move a (continuous) slider handle by grabbing it in the same way as they were grabbing the object in the task; see <xref ref-type="fig" rid="F2">Figure 2</xref>. For every trial the slider handle was reset to the center, i.e., a score of 0.5.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Task and data flow in <italic>baseline</italic> (top), <italic>explicit</italic> (top) and <italic>implicit</italic> (bottom) experimental conditions.</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g002.tif">
<alt-text content-type="machine-generated">Diagram showing a process flow where an RL Agent receives feedback from two pathways. The first, labeled &#x22;baseline, explicit,&#x22; involves scoring with visible results. The second, labeled &#x22;implicit,&#x22; involves LDA transform and normalization. Feedback includes stages: start, pick, place, represented by hand motions interacting with spheres.</alt-text>
</graphic>
</fig>
<sec id="s3-3-1">
<title>3.3.1 Interface conditions</title>
<p>The pick-and-place interaction was designed to simulate a multimodal interaction with visual, auditory, and haptic feedback. To this end, the following sensations were rendered:</p>
<p>(1) Visual Baseline. The object changed its color from white to red when it was grabbed.</p>
<p>(2) Visual with Sound. Together with the color change from white to red, a sound was played at 50% volume through the Vive&#x2019;s earcup speakers. We used a simple &#x2018;plop&#x2019; like sound with a duration of 200&#xa0;m.</p>
<p>(3) Visual with Vibrotactile. With the color change, vibrotactile sensations were rendered at the tip of the thumb, index&#x2013;and middle finger, and the back of the hand.</p>
<p>(4) Visual with Sound and Vibrotactile. Color change, sound, and vibrotactile feedback were rendered together.</p>
</sec>
<sec id="s3-3-2">
<title>3.3.2 Procedure</title>
<p>In total, participants completed three experimental blocks. In the first block, 140 trials had to be completed, with each interface condition being experienced 35 times. The order of the interface conditions was randomized. After every pick-and-place, the questionnaire was presented and participants had to give their score of the preceding pick-and-place interaction. These labeled data were later used to train and assess the neural decoder.</p>
<p>After this first block, the features for the decoder were extracted from the EEG data, and the decoder was trained. This took about 5&#x2013;10&#xa0;min, during which participants could rest. Next, the two experimental conditions of interest were conducted. The order in which the conditions were tested was counterbalanced across participants.</p>
<p>In the <italic>explicit</italic> condition, participants rated every interaction using the slider. The slider values were extracted as 0&#x2013;1 and fed forward to the RL agent, see 3.4 for technical details on the leveraged RL implementation. The agent then selected the interface condition of the next trial. In the <italic>implicit</italic> condition, the question was omitted and instead the output from the decoder, after normalization to the 0&#x2013;1 range, was fed to the agent. As in <italic>explicit</italic>, the agent then selected the interface condition of the following trial.</p>
<p>For both experimental conditions, we set the agent to stop after picking the same interface condition 5 times in a row, i.e., the <italic>convergence</italic> criterion&#x2013;a threshold selected based on a series of pilot experiments. With this criterion we hoped to approximate a practical level of confidence that the agent will stabilize on a preferred policy. Furthermore, this threshold limited the overall interaction time, kept the experiment feasible for participants and prevented fatigue while at the same time provided some robustness against noisy or inconsistent feedback, ensuring that convergence was not triggered by brief fluctuations or outlier responses in the reward signal.</p>
</sec>
</sec>
<sec id="s3-4">
<title>3.4 Reinforcement learning agent</title>
<p>Our RL agent was designed to select the interface condition that best matched the participant&#x2019;s expectation and, given that our experimental setting involved a single state, the problem reduced to a multi-armed bandit with four possible actions&#x2013;essentially, a scenario where the agent repeatedly chooses from four options to maximize its reward over time. Given that our experimental setting involved a single state, the problem reduced to a multi-armed bandit with four possible actions, i.e., a scenario where the agent repeatedly chooses from four options to maximize its reward over time. In this respect, our environment was similar to that of (<xref ref-type="bibr" rid="B53">Porssut et al., 2022</xref>), and we adopted their validated implementation, while acknowledging that other features of the environment might not be equivalent. By following (<xref ref-type="bibr" rid="B53">Porssut et al., 2022</xref>), our agent also employed a combined strategy consisting of an epsilon-greedy approach and the Upper Confidence Bound (UCB) method (<xref ref-type="bibr" rid="B2">Auer et al., 2002</xref>), thereby balancing exploration and exploitation. This combined strategy promotes the selection of actions with high estimated Q-values while still encouraging the exploration of less-visited actions.</p>
<p>For our setting with <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> actions, let <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mn>1,2,3,4</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> denote the set of available interface conditions. We defined the UCB value as:<disp-formula id="equ1">
<mml:math id="m4">
<mml:mrow>
<mml:mtext>UCB</mml:mtext>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>Q</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>c</mml:mi>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>log</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>where action <inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> was one of the four discrete interface conditions, <inline-formula id="inf5">
<mml:math id="m6">
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> was the current Q-value estimate for action <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> (initially set to 1), <inline-formula id="inf7">
<mml:math id="m8">
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> was the exploration parameter (set to 0.25) and <inline-formula id="inf8">
<mml:math id="m9">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> was the number of times action <inline-formula id="inf9">
<mml:math id="m10">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> had been selected up to time <inline-formula id="inf10">
<mml:math id="m11">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. For actions that had not yet been chosen (i.e., <inline-formula id="inf11">
<mml:math id="m12">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>), the corresponding UCB value was set to <inline-formula id="inf12">
<mml:math id="m13">
<mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> to ensure they were explored. The constant <inline-formula id="inf13">
<mml:math id="m14">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.25</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> was chosen in a small suite of pilot simulations using the pure UCB algorithm under our human-feedback noise model. We swept <inline-formula id="inf14">
<mml:math id="m15">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mn>0.25</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and found that 0.25 offered the most robust convergence behavior. In addition to time constraints, we carried this same <inline-formula id="inf15">
<mml:math id="m16">
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> value forward into the hybrid model for data collection to preserve consistency and stability when layering <inline-formula id="inf16">
<mml:math id="m17">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-greedy exploration on top of UCB. This also helped minimize unexpected agent behavior during the real-time user-facing experiment.</p>
<p>At each time step <inline-formula id="inf17">
<mml:math id="m18">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the agent selected an action <inline-formula id="inf18">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> according to the following policy:<disp-formula id="equ2">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mtable class="array">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mtext>a&#x2009;random&#x2009;action&#x2009;chosen</mml:mtext>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mtext>uniformly&#x2009;from</mml:mtext>
<mml:mspace width="0.3333em"/>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mtext>with&#x2009;probability&#x2009;</mml:mtext>
<mml:mi>&#x3b5;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:munder>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">argmax</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mtext>UCB</mml:mtext>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="2.2em"/>
<mml:mtext>with&#x2009;probability&#x2009;</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b5;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>After executing the chosen action and observing a reward <inline-formula id="inf19">
<mml:math id="m21">
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the agent updated its Q-value for that action using:<disp-formula id="equ3">
<mml:math id="m22">
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2190;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>Q</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b3;</mml:mi>
<mml:munder>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
</mml:munder>
<mml:mi>Q</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>Any action not selected on a given trial simply retained its Q-value from the previous time step. Additionally, this update rule differs from the traditional Q-learning rule, but in our tests it converged more quickly and reliably. We suspect that, due to the inherent noise in both human and neural feedback, individual Q-value estimates may fluctuate. By using <inline-formula id="inf20">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mi>Q</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> as a reference, the update was anchored to the best observed performance across all actions, which minimized the impact of these fluctuations and mitigated the risk of unreliable updates when an action&#x2019;s Q-value temporarily deviated due to noise.</p>
<p>Unlike the reference implementation (<xref ref-type="bibr" rid="B53">Porssut et al., 2022</xref>), which used an adjusted reward defined as the mode of the reward history for action <inline-formula id="inf21">
<mml:math id="m24">
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, we chose to use the actual, non-adjusted reward <inline-formula id="inf22">
<mml:math id="m25">
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> since this yielded better performance in our pre-tests.</p>
<p>The initial learning rate <inline-formula id="inf23">
<mml:math id="m26">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> was set to 0.5, the exploration rate <inline-formula id="inf24">
<mml:math id="m27">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> was initially set to 1, and the discount factor was set to <inline-formula id="inf25">
<mml:math id="m28">
<mml:mrow>
<mml:mi>&#x3b3;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.95</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. Both <inline-formula id="inf26">
<mml:math id="m29">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf27">
<mml:math id="m30">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> decayed over time as follows:<disp-formula id="equ4">
<mml:math id="m31">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>max</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>min</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>log</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>40</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
<mml:mtext>with&#x2009;</mml:mtext>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>min</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.001</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="equ5">
<mml:math id="m32">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>max</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>min</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b5;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>log</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
<mml:mtext>with&#x2009;</mml:mtext>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>min</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.01</mml:mn>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>This decay mechanism follows the reference implementation in (<xref ref-type="bibr" rid="B53">Porssut et al., 2022</xref>) and allowed the agent to gradually shift from exploration to exploitation as it gathered more information about the user&#x2019;s preferences. All parameters, with the exception of the UCB exploration constant <inline-formula id="inf28">
<mml:math id="m33">
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, were initialized to the same values as in the reference implementation (<xref ref-type="bibr" rid="B53">Porssut et al., 2022</xref>), as our simulations confirmed that these settings were suitable for our task. <xref ref-type="fig" rid="F3">Figure 3</xref> illustrates the agent performance for one episode.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Performance of a Q-learning agent over time. <bold>(a)</bold> Evolution of Q-values for different actions, with each line representing a specific action. Line segments between successive markers are for illustration only; Q-values are updated only at discrete trial steps, and each segment simply connects the Q-value at step <inline-formula id="inf29">
<mml:math id="m34">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> to that at step <inline-formula id="inf30">
<mml:math id="m35">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> <bold>(b)</bold> Reward, alpha, and epsilon metrics.</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g003.tif">
<alt-text content-type="machine-generated">Graphical illustration showing two plots. Plot (a) displays Q-value against steps, with four lines representing different conditions: baseline, sound, vibration, and vibration plus sound. Baseline and sound stabilize around Q-value of 1, while vibration and vibration plus sound decrease and stabilize lower. Plot (b) shows metric in arbitrary units against steps, illustrating three lines: reward fluctuates, epsilon decreases steadily, and alpha remains low and constant.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-5">
<title>3.5 Neural signal decoder</title>
<p>For loading, synchronizing, and pre-processing the EEG data from the 140 trials of labeled training data, we utilized the EEGLAB (<xref ref-type="bibr" rid="B12">Delorme and Makeig, 2004</xref>) toolbox with wrapper functions from (BeMoBIL-pipeline <xref ref-type="bibr" rid="B27">Klug et al., 2022</xref>) running in a MATLAB 2023b environment (The MathWorks Inc., USA).</p>
<p>Our goal was to design a general decoder of the expected &#x2018;haptic&#x2019; sensations in VR. Therefore, we presumed the most salient neural data to be present directly following the &#x2018;haptic&#x2019; event of picking up the object in our pick-and-place task. Hence, we extracted 1&#xa0;s long data segments following this event and trained a binary model to score expected&#x2013;unexpected sensations.</p>
<p>In the first step to prepare the data for decoder training, noisy (extremely large amplitude fluctuations) trials were rejected. To this end, the EEGLAB function &#x2018;autorej&#x2019; was used on the EEG data, keeping the default parameters: a voltage threshold of 1,000 <inline-formula id="inf31">
<mml:math id="m36">
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>V, a standard deviation cutoff of 5, a maximum rejection rate of 5% per iteration, and all channels included. This function iteratively rejects epochs with improbable data based on standard deviation thresholds, adaptively increasing the threshold when too many epochs exceed the limit.</p>
<p>Next, further trials were flagged and removed by detecting extreme outliers in participants&#x2019; behavior. We used Tukey&#x2019;s method <xref ref-type="bibr" rid="B64">Tukey (1949)</xref>, excluding values exceeding <inline-formula id="inf32">
<mml:math id="m37">
<mml:mrow>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>1,5</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>I</mml:mi>
<mml:mi>Q</mml:mi>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> (Interquartile Range) to determine trials where participants&#x2019; behavior deviated significantly with respect to the placement accuracy of the target object. This resulted in the exclusion of an average 16.1 (SD &#x3d; 2.9) per participant.</p>
<p>In line with earlier work, the EEG data was band-pass filtered to retain frequencies from 0.1 to 15&#xa0;Hz via Fast Fourier Transform (FFT) prior to feature extraction (<xref ref-type="bibr" rid="B17">Gehrke et al., 2022</xref>; <xref ref-type="bibr" rid="B70">Zander et al., 2016</xref>).</p>
<sec id="s3-5-1">
<title>3.5.1 Features</title>
<p>To obtain a robust decoder with many samples for training, we reduced the classification problem to a binary situation. Using a median split on the questionnaire scores, we grouped all trials below the median into a <italic>mismatching expectation</italic> class and all trials above into <italic>matching expectation</italic>, resulting in a balanced dataset with two classes of, at most, 70 trials per class (minus the rejected trials).</p>
<p>The filtered data of all channels was then segmented into 12 epochs of interest of 50&#xa0;m size from 0 to 600&#xa0;m following the grab event. The samples in each 50&#xa0;m segments were then aggregated using the mean, resulting in a matrix of 64 channels X 12 aggregated time windows. Next, baseline correction was performed by subtracting the first time window, i.e. 0&#x2013;50&#xa0;m after the grab. the event was used to subtract the physical noise of the vibrotactile stimulation from the data. This post-event window was chosen to capture and remove early physical artifacts from the vibrotactile stimulation. The first two time windows were discarded, resulting in a 64 X 10 feature matrix retained for each trial.</p>
<p>For each participant, a paired t-test was performed for each feature (channel and time window) to identify the most discriminative features between the <italic>mismatching</italic> and <italic>matching</italic> conditions. The absolute value t-statistics were then sorted and stored. The top 100 were then used in a grid-search on the number of features to use for decoding. To this end, the decoder&#x2019;s accuracy was assessed at 10 to 100 features, increasing in steps of 5. The number of features resulting in a model with the highest accuracy was saved for real-time application. These models were always fit on 80% of the data (80&#x2013;20 train-test split using a 5-fold cross-validation scheme). To assess the decoder&#x2019;s performance, the following performance metrics were calculated: Accuracy, F1 score, and ROC.</p>
<p>We used a linear discriminant analysis (LDA) with shrinkage regularization (automatic shrinkage using the Ledoit-Wolf lemma (<xref ref-type="bibr" rid="B34">Ledoit and Wolf, 2004</xref>)) using the implementations from scikit-learn (<xref ref-type="bibr" rid="B51">Pedregosa et al., 2012</xref>). To normalize LDA scores during real-time application to the 0&#x2013;1 range, we extracted the LDA scores of the 20% test data held out during model fitting. Then, for min-max normalization, the 5th and 95th percentiles were taken from that distribution and stored for normalization of single trials during real-time application. After normalization, any value that exceeded either 0 or 1, was then set to 0 or 1, respectively. This allowed us to retrieve 0&#x2013;1 values from single trials from the binary representation of matching vs mismatching feedback expectations.</p>
</sec>
<sec id="s3-5-2">
<title>3.5.2 Real-time application</title>
<p>During real-time application, the EEG data was buffered for one second following a grab. The data was band-pass filtered analogously to the training data from 0.1 to 15&#xa0;Hz. Next, the features from the best performing model were extracted and, using that model, transformed to an LDA score. Using the min-max anchors, the score was normalized to the 0&#x2013;1 range and sent to an LSL stream in order to be fetched by the RL agent.</p>
</sec>
</sec>
<sec id="s3-6">
<title>3.6 Hypotheses and statistical testing</title>
<p>To answer whether haptic rendering can be tuned using an RL agent, we inspected whether the agent arrived at the <italic>true</italic> label. The true label was operationalized by selecting the interface condition with the highest mean score in block 1, i.e., the training data.</p>
<p>We hypothesized that the RL agent would require the same number of steps until convergence in both implicit and explicit feedback conditions. Hence, the system would perform identically, irrespective of the origin of the feedback. As a reminder, the agent finished picking when it chose the same label 5 times in a row. To test the hypothesis, two one-sided <italic>ttest</italic> (TOST) were conducted. We decided to set the equivalence bounds to five steps, meaning that within <inline-formula id="inf33">
<mml:math id="m38">
<mml:mrow>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> steps difference, we determined the systems to perform equally.</p>
<p>To test whether our experimental manipulation was effective a linear mixed effects model was fit to explain the single-trial questionnaire scores. The interface conditions were entered as a fixed effect and a random intercept was added for each participant. Hence, the model was specified as &#x2018;score <inline-formula id="inf34">
<mml:math id="m39">
<mml:mrow>
<mml:mo>&#x223c;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> interface-condition &#x2b; (1&#x2013; participantID)&#x2019; and fit using the <italic>pymer4</italic> package <xref ref-type="bibr" rid="B22">Jolly (2018)</xref>. A test statistic was obtained by calculating likelihood-ratio tests comparing the full model as specified above against the null model&#x2019;s score <inline-formula id="inf35">
<mml:math id="m40">
<mml:mrow>
<mml:mo>&#x223c;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 1 &#x2b; (1&#x2013; participantID)&#x2019;. All parameters were estimated by maximum likelihood estimation <xref ref-type="bibr" rid="B52">Pinheiro and Bates (2006)</xref>. We computed post-hoc pair-wise tests for &#x2018;interface-condition&#x2019; corrected for multiple comparisons (Tukey method) using the <italic>emmeans</italic> package (<xref ref-type="bibr" rid="B35">Lenth et al., 2020</xref>).</p>
<p>In this work, we did not formulate any <italic>a priori</italic> hypotheses regarding the cortical sources or specific electrode sites contributing to the EEG-based classification. Our approach was exploratory, with all channels initially considered. Sensor-level importance emerged through a data-driven feature selection process during grid search, rather than through predefined regions of interest (ROIs). However, we present ERP waveforms at electrode Cz for visual inspection, given its central location and frequent use in prior literature.</p>
<sec id="s3-6-1">
<title>3.6.1 Post-hoc analysis: Participants&#x2019; scoring consistency over time</title>
<p>In this study, we were generally interested in how an RL agent can handle the noise inherent any reward provided through a neural interface. On top of that, we noticed that participants scoring behavior also exhibited some noise (over time). To address this source of noise in subjective scoring behavior a correlation analysis between time on task and subjective scores was conducted. The Pearson correlation coefficient (<xref ref-type="bibr" rid="B50">Pearson, 1895</xref>) between trial number and corresponding score was computed as a summary statistic per participant. Next, we tested whether the coefficients differed from 0 using <italic>ttest</italic> (<xref ref-type="bibr" rid="B62">Student, 1908</xref>) on the group level.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s4">
<title>4 Results</title>
<p>The contingency table in <xref ref-type="fig" rid="F4">Figure 4a</xref> summarizes the cases of convergence when either implicit or explicit labels were used. We observed that three out of eight times, the agent converged on the &#x2018;correct&#x2019; feedback when using explicit user scores as rewards. On the other hand, the agent converged correctly for 2/8 participants when using implicit rewards. For two participants, the agent converged correctly using explicit rewards, but incorrectly using implicit rewards. For four participants, neither reward origin resulted in the agent converging.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>
<bold>(a)</bold> Contingency table of implicit and explicit convergence states, with color intensity representing frequency, <bold>(b)</bold> Box plot of steps until stoppage for implicit and explicit feedback conditions. Overlaid line depicts the mean trend and grey squares represent individual participants.</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g004.tif">
<alt-text content-type="machine-generated">Diagram consisting of two parts: a) a 2x2 matrix showing implicit vs. explicit convergence, with counts in each cell (top row: true; bottom row: false). b) a box plot comparing steps required for implicit and explicit feedback, with explicit feedback generally taking fewer steps.</alt-text>
</graphic>
</fig>
<p>b. However, the equivalence bounds were not met (lower bound: <inline-formula id="inf36">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>7</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.22</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, upper bound: <inline-formula id="inf37">
<mml:math id="m42">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>7</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.22</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). Hence, we observed evidence that there actually is a difference in convergence time. We noticed that for either reward source, several runs converged after about 25&#x2013;30 picks by the agent.</p>
<p>The mean difference in the number of steps to convergence between implicit and explicit reward sources was 0 (SD &#x3d; 11.6), see <xref ref-type="fig" rid="F4">Figure 4b</xref>. However, the TOST procedure did not yield significance (lower bound: <inline-formula id="inf38">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>7</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.22</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf39">
<mml:math id="m44">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>; upper bound: <inline-formula id="inf40">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>7</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.22</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf41">
<mml:math id="m46">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>), meaning that statistical equivalence within the predefined bounds could not be established.</p>
<sec id="s4-1">
<title>4.1 Task validation</title>
<p>The four interface conditions significantly influenced participants&#x2019; ratings of how consistent the virtual experience felt compared to their real-world expectations (<inline-formula id="inf42">
<mml:math id="m47">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>&#x3c7;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>9.6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf43">
<mml:math id="m48">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>02</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). Among the conditions, vibration feedback was rated as most consistent with real-world experience, while sound feedback received the lowest ratings. A follow-up comparison revealed that vibration was rated significantly higher than sound (<inline-formula id="inf44">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>45.2</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2.8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf45">
<mml:math id="m50">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>037</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>); see <xref ref-type="fig" rid="F5">Figure 5a</xref>.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>
<bold>(a)</bold> Box plot of scores in the training block across haptic profiles. <bold>(b)</bold> Box plot of Pearson&#x2019;s correlation coefficient between trial number and scores in the training block across haptic profiles. In both plots, the overlaid line represents the mean trend.</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g005.tif">
<alt-text content-type="machine-generated">Box plots in two panels labeled a and b. Panel a shows scores across four conditions: baseline, sound, vibration, and vibration plus sound. Panel b presents Pearson correlation coefficients for the same conditions. Both plots display variations and individual data points, with asterisks indicating significant differences.</alt-text>
</graphic>
</fig>
<p>We found that in both interface conditions &#x2018;baseline&#x2019; <inline-formula id="inf46">
<mml:math id="m51">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>14</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2.41</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>031</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and &#x2018;vibration&#x2019; <inline-formula id="inf47">
<mml:math id="m52">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>14</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2.28</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>04</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> participants changed their scoring behavior over time, see <xref ref-type="fig" rid="F5">Figure 5B</xref>. In these conditions, their scoring became increasingly more positive with time on task (Pearson <inline-formula id="inf48">
<mml:math id="m53">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> for &#x2018;baseline&#x2019;: M &#x3d; 0.25, SD &#x3d; 0.38, and &#x2018;vibration&#x2019;: M &#x3d; 0.15, SD &#x3d; 0.25).</p>
</sec>
<sec id="s4-2">
<title>4.2 Decoder performance</title>
<p>Visual inspection of the amplitudes at electrode Cz revealed an increase in the difference between <italic>high</italic> and <italic>low</italic> scored trials towards the later stages of the 100&#x2013;500&#xa0;m window used for classification, see <xref ref-type="fig" rid="F6">Figure 6a</xref>. Indeed, the most frequently leveraged time windows as determined by the grid-search were the last three windows, i.e., 400&#x2013;450, 450&#x2013;500, and 500&#x2013;550&#xa0;m following the grab event (see <xref ref-type="fig" rid="F6">Figure 6b</xref>). In terms of sensors, the five most leveraged channels were TP10, AFz, AF8, T8, and C6, see <xref ref-type="fig" rid="F6">Figure 6c</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>
<bold>(a)</bold> ERP at CZ for the preference split dark Gy&#xa0;bar near the x-axis indicates time window considered for classification, light Gy&#xa0;bar indicates baseline; <bold>(b)</bold> Number of times time window was selected for classification; <bold>(c)</bold> Number of times channel was selected for classification; <bold>(d)</bold> ROC curves for all participants. The dashed diagonal line represents random classification, while the solid curves indicate model performance.</p>
</caption>
<graphic xlink:href="frvir-06-1616442-g006.tif">
<alt-text content-type="machine-generated">(a) Line chart showing amplitude in microvolts over time in milliseconds with high and low conditions. (b) Bar chart of counts within various time windows after an event. (c) Scalp map displaying electrode occurrence in a classifier using a red color gradient. (d) ROC curve graph showing true positive rate versus false positive rate with multiple curves and a diagonal reference line.</alt-text>
</graphic>
</fig>
<p>The automatic selection of the number of features used for classification for each participant resulted in an average of 40.7 (SD &#x3d; 25) features being picked by the procedure. The classifier cross-validation resulted in a mean accuracy of 0.7 (SD &#x3d; 0.06) and a mean F1 score of 0.8 (SD &#x3d; 0.06), see <xref ref-type="fig" rid="F6">Figure 6d</xref> for the mean, as well as individual participants&#x2019; ROC.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<title>5 Discussion</title>
<p>In this paper, we set out to answer three questions: (1) Can we tune haptic rendering to participants&#x2019; preferences using an RL agent based on human feedback? (2) Is this possible through implicit labels obtained through a neural decoder? And (3) are there disadvantages when relying on implicit instead of explicit labels?</p>
<p>We investigated these questions by building a novel, proof-of-concept, neuroadaptive system comprising an LDA-based EEG decoder and a UCB-based RL agent. The system was designed to automatically select the multisensory haptic experience for the human-in-the-loop in each following trial. We found the EEG decoder to operate at satisfactory levels (Mean F1 score of <inline-formula id="inf49">
<mml:math id="m54">
<mml:mrow>
<mml:mo>&#x223c;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). However, in our real-time application scenario, we observed considerably poor performance with the agent converging to the &#x2018;correct&#x2019; solution in around 25% of cases across the two different reward sources. Due to the small number of participants completing the full RL protocol <inline-formula id="inf50">
<mml:math id="m55">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and the limited number of converged runs, the statistical confidence and generalizability of our findings are constrained. Future work should aim to expand the sample size to improve reliability and enable more robust conclusions about system performance and user variability. Here, we discuss our exploratory findings and provide a roadmap for future work.</p>
<sec id="s5-1">
<title>5.1 RL agents learning from (noisy) human feedback</title>
<p>Regarding the first question&#x2013;whether we can tune haptic rendering to participants&#x2019; preferences using an RL agent based on human feedback&#x2013;we found the performance of our prototype to be hindered by noisy rewards, particularly those derived from the EEG decoder. In human-in-the-loop settings, where rewards can be both noisy and non-stationary, effective exploration becomes critical. This contrasts with conventional RL scenarios, where reward functions are typically stable and well-defined.</p>
<p>In our study, the RL agent took approximately 25&#x2013;30 steps per episode to converge on one of the four haptic feedback conditions. While this relatively short duration aligns with the simplicity of our VR pick-and-place task, it contrasts significantly with typical RL scenarios that often involve thousands or millions of actions (e.g., in game-playing AI or simulated robotics environments). Shorter episodes facilitate rapid convergence and manageable computational demands, making them practical for real-time neuroadaptive applications. Futhermore, they reduce cognitive and physiological fatigue, resulting in more stable neural rewards. Conversely, longer episodes, e.g., exceeding 10&#xa0;min, risk significant shifts in cognitive states.</p>
<p>However, limited episode length inherently restricts the agent&#x2019;s ability to explore the action space thoroughly. In our data, with just 25&#x2013;30 steps per episode, the RL agent might not have gathered sufficient experience to robustly model nuanced haptic expectations or adapt effectively to subtle reward variations over time. Indeed, the instability in convergence we observed indicates that limited step count posed a crucial bottleneck. The noise and variability in participant ratings also highlight the need for higher-frequency sampling. Future improvements in interaction design, such as brief, minimally intrusive feedback probes or passive, implicit feedback triggered by specific user behaviors, could enhance sampling frequency without disrupting user immersion.</p>
<p>Conversely, while longer episodes might provide richer exploration, they introduce challenges such as user fatigue and drift in subjective ratings, complicating the agent&#x2019;s policy learning. Future research could thus explore hybrid or hierarchical RL approaches, incorporating long-term memory or meta-learning frameworks designed to better handle limited step counts and dynamic user feedback. Dynamically adjusting episode lengths based on real-time measures of learning progress or reward consistency could further optimize the balance between exploration depth, computational efficiency, and user experience.</p>
<p>In our implementation, we chose to combine <inline-formula id="inf51">
<mml:math id="m56">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-greedy with UCB exploration but omitted additional noise-handling mechanisms. Both <inline-formula id="inf52">
<mml:math id="m57">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-greedy and UCB are mechanisms designed to ensure exploration, and thus one might consider using both simultaneously to be redundant. One could argue that to achieve a higher level of exploration&#x2013;one of the reasons given by <xref ref-type="bibr" rid="B53">Porssut et al. (2022)</xref> for this dual approach&#x2013;it would suffice to simply increase the UCB exploration constant <inline-formula id="inf53">
<mml:math id="m58">
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, thereby pushing the algorithm to explore more aggressively. However, in practice this is not equivalent. UCB&#x2019;s exploration bonus naturally decreases as actions are sampled, meaning that in noisy or non-stationary settings, the associated confidence intervals can shrink too quickly, which may lead the agent to prematurely settle on suboptimal actions. By contrast, a fixed <inline-formula id="inf54">
<mml:math id="m59">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> in an <inline-formula id="inf55">
<mml:math id="m60">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-greedy strategy guarantees that even well-sampled arms are occasionally revisited, an advantage that becomes especially important when human feedback is noisy and the environment is subject to change.</p>
<p>Our Q-learning update deviated from the traditional rule by anchoring each step directly to <inline-formula id="inf56">
<mml:math id="m61">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mi>Q</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Although this modification was introduced to accelerate convergence under noisy human and neural rewards, it also served as an implicit regularizer: by filtering out erratic spikes or dips in individual action values, it promotes smoother learning trajectories. However, this anchoring may bias the agent toward historically high-valued actions, which could reduce exploratory behavior in environments where reward contingencies shift over time. Future work could investigate adaptive anchoring strategies or hybrid update schemes that retain robustness to noise while preserving sufficient exploration, especially in multi-state or non-stationary settings.</p>
<p>In this work, we decided against using a perturbed rewards mechanism in the final solution, deviating from previous implementations <xref ref-type="bibr" rid="B53">Porssut et al. (2022)</xref>. The combination of UCB with <inline-formula id="inf57">
<mml:math id="m62">
<mml:mrow>
<mml:mi>&#x3b5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-greedy exploration already averages out random fluctuations in the reward signal over many trials; when the noise is moderate, the RL agent&#x2019;s inherent averaging means that extra corrections&#x2014;such as majority voting&#x2014;do not significantly change the outcome. Moreover, our empirical data showed that the RL agent converged to a stable threshold even without the perturbed rewards, suggesting that this additional mechanism was redundant since its intended effect of cleaning up noisy rewards was already achieved by the standard exploration&#x2013;exploitation dynamics.</p>
</sec>
<sec id="s5-2">
<title>5.2 BCI</title>
<p>Turning to the second question&#x2013;whether tuning is possible using implicit labels from a neural decoder&#x2013;we observed that the difference in the number of steps to convergence between explicit and implicit rewards was zero. However, the TOST procedure did not confirm statistical equivalence (i.e., performance was not significantly similar within the predefined bounds). This does not imply a meaningful difference between the two, but rather that the data were insufficient to establish equivalence. Across both reward sources, several runs converged after roughly 25&#x2013;30 selections, indicating overall similar but highly variable convergence behavior.</p>
<p>Our implementation used a simple, linear EEG classifier that was subject-specific. This choice was motivated by the implementation simplicity, the need for low-latency deployment (low camputational demands), as well as the near-optimal performance of linear models on EEG features with roughly homoscedastic class distributions (<xref ref-type="bibr" rid="B6">Blankertz et al., 2011</xref>; <xref ref-type="bibr" rid="B38">Lotte et al., 2018</xref>). Classifiers were trained on labeled data collected around 10&#xa0;minutes prior to deployment. While necessary for counterbalancing, this setup introduced a risk of temporal overfitting where the model may have adapted too specifically to neural states at training, and performed less well if participants&#x2019; cognitive states shifted during the session (<xref ref-type="bibr" rid="B21">Hosseini et al., 2020</xref>).</p>
<p>For a first look at interpretable signals, we visualized ERP waveforms at electrode Cz. However, this was a post-hoc decision rather than based on predefined ROIs. Future work should more systematically relate decoding features to known ERP components such as the P3 or Prediction Error Negativity (PEN). In particular, fronto-central sites like electrode FCz, which is often associated with activity originating in or near the anterior cingulate cortex (ACC), may be especially informative due to their established role in feedback evaluation. Our grid search for feature selection consistently prioritized late time windows (e.g., 400&#x2013;550&#xa0;m after the grab event), aligning with the temporal dynamics of evaluative ERP components like the P3 complex (<xref ref-type="bibr" rid="B46">Metcalfe, 2017</xref>), and suggesting that decoding relied more on later-stage cognitive appraisal processes than early sensory encoding. Understanding which brain regions contribute to decoding is especially interesting given our multisensory task, where participants evaluated haptic and multimodal stimuli, likely engaging a broad network spanning primary and secondary sensory cortices (via modality-specific thalamic inputs), posterior insula, and posterior parietal cortex (<xref ref-type="bibr" rid="B59">Sathian, 2016</xref>; <xref ref-type="bibr" rid="B1">Andersen and Buneo, 2002</xref>). The rating process itself probably involved valuation networks including the insula, dorsal anterior cingulate cortex, and prefrontal areas such as the orbitofrontal, dorsolateral, and ventromedial cortex (<xref ref-type="bibr" rid="B45">Menon and Uddin, 2010</xref>; <xref ref-type="bibr" rid="B3">Bartra et al., 2013</xref>). This distributed cortical activation may explain the absence of a single dominant spatial feature (see scalp map in <xref ref-type="fig" rid="F6">Figure 6c</xref>) and supports the view that classification relied on signals distributed across multiple EEG channels.</p>
<p>While our linear decoder served as a baseline, more expressive models (e.g., CNNs or RNNs) could be explored. Coupled with explainable AI techniques such as saliency maps or layer-wise relevance propagation (<xref ref-type="bibr" rid="B13">Farahat et al., 2019</xref>; <xref ref-type="bibr" rid="B47">Nam et al., 2023</xref>), these could help reveal which spectral and/or spatial EEG components drive feedback decoding. This could not only improve classification but also provide deeper insight into the neural underpinnings reflecting how user experience unfolds over time.</p>
<p>As EEG devices become wireless (<xref ref-type="bibr" rid="B49">Niso et al., 2023</xref>), more compact (<xref ref-type="bibr" rid="B24">Kaongoen et al., 2023</xref>), and increasingly integrated into XR hardware<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref>, questions of wearability and long-term comfort are moving into focus, an essential step toward real-world deployment of neuroadaptive systems. Still, key challenges remain. EEG is highly prone to artifacts from movement and muscle activity, especially in naturalistic settings (<xref ref-type="bibr" rid="B23">Jungnickel et al., 2019</xref>; <xref ref-type="bibr" rid="B19">Gramann et al., 2011</xref>; <xref ref-type="bibr" rid="B41">Makeig et al., 2009</xref>; <xref ref-type="bibr" rid="B27">Klug et al., 2022</xref>). Real-time deployment also requires low-latency pipelines, which limits the complexity of models and preprocessing. Moreover, EEG signals vary substantially within and across individuals, often requiring personalized calibration or adaptive learning approaches (<xref ref-type="bibr" rid="B65">Wan et al., 2021</xref>; <xref ref-type="bibr" rid="B67">Wu et al., 2022</xref>). Finally, data privacy is a critical concern: because neural signals are sensitive and potentially identifiable, processing EEG data locally on the XR device instead of streaming it to external servers, can help minimize privacy risks while supporting faster, more secure interaction.</p>
</sec>
<sec id="s5-3">
<title>5.3 Task and procedure</title>
<p>Finally, regarding the third question&#x2013;are there disadvantages when relying on implicit rather than explicit labels&#x2013;we observed several challenges. A primary difficulty was the variability of user-provided labels over time. As discussed above, RL algorithms typically rely on stable reward signals, yet participant ratings often fluctuated. Correlation analyses revealed gradual shifts in subjective scores in some haptic feedback conditions as the experiment progressed, specifically for the visual-only baseline and vibration. This suggests that repeated exposure influenced participant judgments and potentially introduced biases into the RL process.</p>
<p>A related methodological consideration was our convergence criterion: the RL agent stopped adapting after selecting the same feedback condition five consecutive times. Participants were not informed about this convergence criterion, potentially leading to unintended confusion or frustration if they perceived no clear pattern in system responses. Future studies might explicitly communicate adaptive goals or provide intermediate feedback to clarify the system&#x2019;s intent, helping participants form more consistent expectations.</p>
<p>We also observed substantial individual differences in rating distributions. Some participants showed near-binary preference structures, consistently rating one condition as highly consistent with real-world experience while rejecting others, whereas others exhibited more graded preferences, implying nuanced perception of sensory integration. This divergence complicates RL-based adaptation, as binary structures favor rapid convergence, while graded responses introduce greater noise. Note that in this study the slider was reset to the midpoint after each trial, potentially inducing a central-tendency bias. Such a bias might have dampened extreme ratings, thereby flattening the reward gradient and slowing adaptation. Consequently, our findings in fact may underestimate rather than overestimate the achievable performance of the RL agent. Furthermore, as participants repeatedly moved the slider using substantial arm movements, first to grab the handle and then to drag it, fatigue likely increased over time. This may have led to two effects: (1) reduced rating variability, amplifying the central-tendency bias, and (2) a directional bias favoring the physically more comfortable movement direction. Future iterations, leveraging a scale without a pre-placed handle, may eliminate this confound to some degree, thereby sharpen the reward signal and accelerate convergence.</p>
<p>Another potential confounding factor was an anchoring effect. Depending on the haptic condition participants first experienced, subsequent ratings might have been influenced by initial exposure. Ideally, pseudo-randomizing the initial condition to balance early experiences could mitigate this effect. However, in our study, the starting condition was fully randomized.</p>
<p>Together, these anchoring and scale biases may have further complicated training and interpreting the neural decoder. High heterogeneity in rating distributions, ranging from bimodal to unimodal response patterns, made it challenging to generate consistent labels for classifier training. Overall, while implicit neuroadaptive rewards offer an interesting alternative to explicit ratings, human perception and neural appraisal remain dynamic, context-sensitive, and influenced by both methodological and psychological factors. Future work should explore adaptive mechanisms accounting for evolving preferences, biases, and individual variability, ultimately supporting robust and scalable neuroadaptive XR systems.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s6">
<title>6 Conclusion</title>
<p>In this study, we presented a proof-of-concept neuroadaptive XR system using RL to adapt multisensory haptic feedback based on explicit user ratings and implicit EEG-derived feedback. Although our EEG decoder achieved satisfactory offline performance, real-time RL performance was hindered by noisy and non-stationary feedback signals, resulting in an RL agent with poor performance.</p>
<p>Several critical challenges emerged from our study. Short episodes allowed rapid convergence but limited the RL agent&#x2019;s exploration capacity, resulting in poor overall performance. Conversely, longer episodes risk cognitive fatigue and shifting user preferences. Future work should explore interaction designs that enable higher-frequency feedback sampling. Regarding EEG-based implicit rewards, potential temporal shifts in user cognitive states complicated decoder generalization. Periodic recalibration or transfer learning paradigms spanning multiple sessions or users could mitigate this. Our linear EEG decoder prioritized later ERP components. We think it to be important to systematically link EEG decoding features with established ERP (e.g., P300, PEN) and spectral components, to enhance interpretability and neuroscientific insights, a promising avenue for future research. Finally, we noted significant variability in explicit user ratings. These biases likely introduced additional noise, impeding RL adaptation.</p>
<p>While we believe implicit EEG-based rewards offer an interesting alternative with a high upside to using explicit labels, we encountered both approaches to be challenged by dynamic, noisy, and context-sensitive human feedback. Addressing these challenges will be key to advancing neuroadaptive XR systems that are more attuned to the nuances of human experience.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="ethics-statement" id="s8">
<title>Ethics statement</title>
<p>The studies involving humans were approved by the local ethics committee of the Department of Psychology and Ergonomics at the TU Berlin (Ethics protocol approval code: BPN_GEH_220421). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.</p>
</sec>
<sec sec-type="author-contributions" id="s9">
<title>Author contributions</title>
<p>LG: Formal Analysis, Conceptualization, Project administration, Methodology, Supervision, Data curation, Investigation, Software, Visualization, Writing &#x2013; original draft, Writing &#x2013; review and editing. AK: Methodology, Software, Visualization, Formal Analysis, Validation, Writing &#x2013; original draft, Writing &#x2013; review and editing. MK: Methodology, Supervision, Writing &#x2013; review and editing. KG: Resources, Funding acquisition, Conceptualization, Writing &#x2013; review and editing.</p>
</sec>
<sec sec-type="funding-information" id="s10">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. This research was conducted within the project Brain Dynamics in Cyber-Physical Systems as a Measure of User Presence, funded by Deutsche Forschungsgemeinschaft (DFG) - project number GR 2627/13-1. We acknowledge support by the Open Access Publication Fund of TU Berlin.</p>
</sec>
<ack>
<p>We thank Magdalena Biada&#x142;a for helping with data collection. ChatGPT (OpenAI, San Francisco, USA) was used to copy-edit author-generated content.</p>
</ack>
<sec sec-type="COI-statement" id="s11">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
<p>The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.</p>
</sec>
<sec sec-type="ai-statement" id="s12">
<title>Generative AI statement</title>
<p>The author(s) declare that Generative AI was used in the creation of this manuscript. To copy-edit author-generated text.</p>
</sec>
<sec sec-type="disclaimer" id="s13">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://lukasgehrke.github.io/neuroadaptive-xr">https://lukasgehrke.github.io/neuroadaptive-xr</ext-link>
</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://galea.co">https://galea.co</ext-link>
</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Andersen</surname>
<given-names>R. A.</given-names>
</name>
<name>
<surname>Buneo</surname>
<given-names>C. A.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>Intentional maps in posterior parietal cortex</article-title>. <source>Annu. Rev. Neurosci.</source> <volume>25</volume>, <fpage>189</fpage>&#x2013;<lpage>220</lpage>. <pub-id pub-id-type="doi">10.1146/annurev.neuro.25.112701.142922</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Auer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Cesa-Bianchi</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>Finite-time analysis of the multiarmed bandit problem</article-title>. <source>Mach. Learn.</source> <volume>47</volume>, <fpage>235</fpage>&#x2013;<lpage>256</lpage>. <pub-id pub-id-type="doi">10.1023/a:1013689704352</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bartra</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>McGuire</surname>
<given-names>J. T.</given-names>
</name>
<name>
<surname>Kable</surname>
<given-names>J. W.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>The valuation system: a coordinate-based meta-analysis of BOLD fMRI experiments examining neural correlates of subjective value</article-title>. <source>Neuroimage</source> <volume>76</volume>, <fpage>412</fpage>&#x2013;<lpage>427</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2013.02.063</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bastos</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Usrey</surname>
<given-names>W. M.</given-names>
</name>
<name>
<surname>Adams</surname>
<given-names>R. A.</given-names>
</name>
<name>
<surname>Mangun</surname>
<given-names>G. R.</given-names>
</name>
<name>
<surname>Fries</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Friston</surname>
<given-names>K. J.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Canonical microcircuits for predictive coding</article-title>. <source>Neuron</source> <volume>76</volume>, <fpage>695</fpage>&#x2013;<lpage>711</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2012.10.038</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Berger</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Tennent</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Spence</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Maior</surname>
<given-names>H. A.</given-names>
</name>
<name>
<surname>Ramchurn</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Wilson</surname>
<given-names>M. L.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Adapting virtual reality horror experiences using fNIRS</article-title>,&#x201d; in <source>Proceedings of the fNIRS biennial meeting 2024</source>.</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Blankertz</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Lemm</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Treder</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Haufe</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>M&#xfc;ller</surname>
<given-names>K.-R.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Single-trial analysis and classification of ERP components &#x2014; a tutorial</article-title>. <source>Neuroimage</source> <volume>56</volume>, <fpage>814</fpage>&#x2013;<lpage>825</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2010.06.048</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Brooks</surname>
<given-names>F. P.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>What&#x2019;s real about virtual reality?</article-title> <source>IEEE Comput. Graph. Appl.</source> <volume>19</volume>, <fpage>16</fpage>&#x2013;<lpage>27</lpage>. <pub-id pub-id-type="doi">10.1109/38.799723</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chatrian</surname>
<given-names>G. E.</given-names>
</name>
<name>
<surname>Lettich</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Nelson</surname>
<given-names>P. L.</given-names>
</name>
</person-group> (<year>1985</year>). <article-title>Ten percent electrode system for topographic studies of spontaneous and evoked EEG activities</article-title>. <source>Am. J. EEG Technol.</source> <volume>25</volume>, <fpage>83</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1080/00029238.1985.11080163</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cho</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Sendhilnathan</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Nebeling</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Padmanabhan</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Browder</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>SonoHaptics: an audio-haptic cursor for gaze-based object selection in XR</article-title>. <source>
<italic>Proc. 37th Annu. ACM Symposium User Interface Softw. Technol.</italic> (New York, N. Y. U. S. A ACM)</source> <volume>8</volume>, <fpage>1</fpage>&#x2013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1145/3654777.3676384</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Clark</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Whatever next? predictive brains, situated agents, and the future of cognitive science</article-title>. <source>Behav. Brain Sci.</source> <volume>36</volume>, <fpage>181</fpage>&#x2013;<lpage>204</lpage>. <pub-id pub-id-type="doi">10.1017/s0140525x12000477</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dehais</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Lafont</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Roy</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fairclough</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A neuroergonomics approach to mental workload, engagement and human performance</article-title>. <source>Front. Neurosci.</source> <volume>14</volume>, <fpage>268</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2020.00268</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Delorme</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Makeig</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis</article-title>. <source>J. Neurosci. Methods</source> <volume>134</volume>, <fpage>9</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneumeth.2003.10.009</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Farahat</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Reichert</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Sweeney-Reed</surname>
<given-names>C. M.</given-names>
</name>
<name>
<surname>Hinrichs</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Convolutional neural networks for decoding of covert attention focus and saliency maps for EEG feature visualization</article-title>. <source>J. Neural Eng.</source> <volume>16</volume>, <fpage>066010</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/ab3bb4</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Friston</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>The free-energy principle: a unified brain theory?</article-title> <source>Nat. Rev. Neurosci.</source> <volume>11</volume>, <fpage>127</fpage>&#x2013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2787</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gagn&#xe9;</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Azad</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Labb&#xe9;</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Campbell</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Isabel</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Scheme</surname>
<given-names>E.</given-names>
</name>
<etal/>
</person-group> (<year>2025</year>). <article-title>Context informed incremental learning improves myoelectric control performance in virtual reality object manipulation tasks</article-title>.</citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gehrke</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Akman</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lopes</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>A. K.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.-T.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). &#x201c;<article-title>Detecting visuo-haptic mismatches in virtual reality using the prediction error negativity of event-related brain potentials</article-title>,&#x201d;, <publisher-loc>New York, New York, USA</publisher-loc>: <publisher-name>ACM Press, CHI &#x2019;19</publisher-name>, <volume>427</volume> <fpage>1</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1145/3290605.3300657</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gehrke</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Lopes</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Klug</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Akman</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Neural sources of prediction errors detect unrealistic VR interactions</article-title>. <source>J. Neural Eng.</source> <volume>19</volume>, <fpage>036002</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/ac69bc</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gehrke</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Terfurth</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Akman</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Visuo-haptic prediction errors: a multimodal dataset (EEG, motion) in BIDS format indexing mismatches in haptic interaction</article-title>. <source>Front. Neuroergonomics</source> <volume>5</volume>, <fpage>1411305</fpage>. <pub-id pub-id-type="doi">10.3389/fnrgo.2024.1411305</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Gwin</surname>
<given-names>J. T.</given-names>
</name>
<name>
<surname>Ferris</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Oie</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Jung</surname>
<given-names>T.-P.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>C.-T.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>Cognition in action: imaging brain/body dynamics in Mobile humans</article-title>. <source>Rev. Neurosci.</source> <volume>22</volume>, <fpage>593</fpage>&#x2013;<lpage>608</lpage>. <pub-id pub-id-type="doi">10.1515/rns.2011.047</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hettinger</surname>
<given-names>L. J.</given-names>
</name>
<name>
<surname>Branco</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Encarnacao</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Bonato</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Neuroadaptive technologies: applying neuroergonomics to the design of advanced interfaces</article-title>. <source>Theor. Issues Ergon.</source> <volume>4</volume>, <fpage>220</fpage>&#x2013;<lpage>237</lpage>. <pub-id pub-id-type="doi">10.1080/1463922021000020918</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hosseini</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Powell</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Collins</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Callahan-Flintoft</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Bowman</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>I tried a bunch of things: the dangers of unexpected overfitting in classification of brain data</article-title>. <source>Neurosci. Biobehav. Rev.</source> <volume>119</volume>, <fpage>456</fpage>&#x2013;<lpage>467</lpage>. <pub-id pub-id-type="doi">10.1016/j.neubiorev.2020.09.036</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jolly</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Pymer4: connecting R and python for linear mixed modeling</article-title>. <source>J. Open Source Softw.</source> <volume>3</volume>, <fpage>862</fpage>. <pub-id pub-id-type="doi">10.21105/joss.00862</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Jungnickel</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Gehrke</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Klug</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Chapter 10 - mobi&#x2014;mobile brain/body imaging</article-title>,&#x201d; in <source>Neuroergonomics</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Ayaz</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Dehais</surname>
<given-names>F.</given-names>
</name>
</person-group> (<publisher-name>Academic Press</publisher-name>), <fpage>59</fpage>&#x2013;<lpage>63</lpage>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kaongoen</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Woo Choi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kwon</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Hwang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Hwang</surname>
<given-names>G.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>The future of wearable EEG: a review of ear-EEG technology and its applications</article-title>. <source>J. Neural Eng.</source> <volume>20</volume>, <fpage>051002</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/acfcda</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kaufmann</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Weng</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Bengs</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>H&#xfc;llermeier</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>A survey of reinforcement learning from human feedback</article-title>. <source>arXiv</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2312.14925</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Keller</surname>
<given-names>G. B.</given-names>
</name>
<name>
<surname>Mrsic-Flogel</surname>
<given-names>T. D.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Predictive processing: a canonical cortical computation</article-title>. <source>Neuron</source> <volume>100</volume>, <fpage>424</fpage>&#x2013;<lpage>435</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2018.10.003</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Klug</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jeung</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wunderlich</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gehrke</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Protzak</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Djebbara</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>The BeMoBIL pipeline for automated analyses of multimodal Mobile brain and body imaging data</article-title>. <source>bioRxiv</source> <volume>2022</volume>, <fpage>510051</fpage>. <pub-id pub-id-type="doi">10.1101/2022.09.29.510051</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Knill</surname>
<given-names>D. C.</given-names>
</name>
<name>
<surname>Pouget</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>The Bayesian brain: the role of uncertainty in neural coding and computation</article-title>. <source>Trends Neurosci.</source> <volume>27</volume>, <fpage>712</fpage>&#x2013;<lpage>719</lpage>. <pub-id pub-id-type="doi">10.1016/j.tins.2004.10.007</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Knox</surname>
<given-names>W. B.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Augmenting reinforcement learning with human feedback</article-title>
</citation>
</ref>
<ref id="B30">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kosunen</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Salminen</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>J&#xe4;rvel&#xe4;</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ruonala</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ravaja</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Jacucci</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>RelaWorld</article-title>,&#x201d; in <source>
<italic>Proceedings of the 21st international conference on intelligent user interfaces</italic> (new York, NY, USA: acm)</source>.</citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kothe</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shirazi</surname>
<given-names>S. Y.</given-names>
</name>
<name>
<surname>Stenner</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Medine</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Boulay</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Grivich</surname>
<given-names>M. I.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <source>The lab streaming layer for synchronized multimodal recording</source>. <publisher-name>bioRxivorg Hollow, New York: Cold Spring Harbor Laboratory</publisher-name>. <pub-id pub-id-type="doi">10.1101/2024.02.13.580071</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krol</surname>
<given-names>L. R.</given-names>
</name>
<name>
<surname>Haselager</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zander</surname>
<given-names>T. O.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Cognitive and affective probing: a tutorial and review of active learning for neuroadaptive technology</article-title>. <source>J. Neural Eng.</source> <volume>17</volume>, <fpage>012001</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/ab5bb5</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Krol</surname>
<given-names>L. R.</given-names>
</name>
<name>
<surname>Zander</surname>
<given-names>T. O.</given-names>
</name>
</person-group> (<year>2022</year>). &#x201c;<article-title>Defining neuroadaptive technology: the trouble with implicit human-computer interaction</article-title>,&#x201d; in <source>Current research in neuroadaptive technology</source> (<publisher-name>Elsevier</publisher-name>), <fpage>17</fpage>&#x2013;<lpage>42</lpage>.</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ledoit</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Wolf</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>A well-conditioned estimator for large-dimensional covariance matrices</article-title>. <source>J. Multivar. Anal.</source> <volume>88</volume>, <fpage>365</fpage>&#x2013;<lpage>411</lpage>. <pub-id pub-id-type="doi">10.1016/s0047-259x(03)00096-4</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lenth</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Singmann</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Love</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Buerkner</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Herve</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Package &#x2018;Emmeans</article-title>. <source>R. package version</source> <volume>34</volume>, <fpage>216</fpage>&#x2013;<lpage>221</lpage>. <pub-id pub-id-type="doi">10.32614/CRAN.package.emmeans</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Gomez</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Nakamura</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Human-centered reinforcement learning: a survey</article-title>. <source>IEEE Trans. Hum. Mach. Syst.</source> <volume>49</volume>, <fpage>337</fpage>&#x2013;<lpage>349</lpage>. <pub-id pub-id-type="doi">10.1109/thms.2019.2912447</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lopes</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Ion</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Baudisch</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Impacto: simulating physical impact by combining tactile stimulation with electrical muscle stimulation</article-title>,&#x201d; in <source>Proceedings of the 28th annual ACM symposium on user interface software and technology</source> (<publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>11</fpage>&#x2013;<lpage>19</lpage>.</citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lotte</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Bougrain</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Cichocki</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Clerc</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Congedo</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Rakotomamonjy</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>A review of classification algorithms for EEG-Based brain&#x2013;computer interfaces: a 10 year update</article-title>. <source>J. Neural Eng.</source> <volume>15</volume>, <fpage>031005</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/aab2f2</pub-id>
</citation>
</ref>
<ref id="B39">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Luo</surname>
<given-names>T.-J.</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>Y.-C.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>J.-T.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>C.-L.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Deep reinforcement learning from error-related potentials <italic>via</italic> an EEG-Based brain-computer interface</article-title>,&#x201d; in <source>
<italic>IEEE international conference on bioinformatics and biomedicine (BIBM)</italic> (IEEE)</source>.</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mahmoudi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Khosrotabar</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Rinderknecht</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Sharbafi</surname>
<given-names>M. A.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Using passive BCI for personalization of assistive wearable devices: a proof-of-concept study</article-title>. <source>IEEE Trans. Neural Syst. Rehabil. Eng.</source>, <fpage>1</fpage>. <pub-id pub-id-type="doi">10.1109/TNSRE.2025.3530154</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Makeig</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Jung</surname>
<given-names>T.-P.</given-names>
</name>
<name>
<surname>Sejnowski</surname>
<given-names>T. J.</given-names>
</name>
<name>
<surname>Poizner</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Linking brain, mind and behavior</article-title>. <source>Int. J. Psychophysiol.</source> <volume>73</volume>, <fpage>95</fpage>&#x2013;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijpsycho.2008.11.008</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Makransky</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Lilleholt</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Aaby</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Development and validation of the multimodal presence scale for virtual reality environments: a confirmatory factor analysis and item response theory approach</article-title>. <source>Comput. Hum. Behav.</source> <volume>72</volume>, <fpage>276</fpage>&#x2013;<lpage>285</lpage>. <pub-id pub-id-type="doi">10.1016/j.chb.2017.02.066</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mark</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Kraft</surname>
<given-names>A. E.</given-names>
</name>
<name>
<surname>Ziegler</surname>
<given-names>M. D.</given-names>
</name>
<name>
<surname>Ayaz</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Neuroadaptive training <italic>via</italic> fNIRS in flight simulators</article-title>. <source>Front. Neuroergonomics</source> <volume>3</volume>, <fpage>820523</fpage>. <pub-id pub-id-type="doi">10.3389/fnrgo.2022.820523</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mehta</surname>
<given-names>R. K.</given-names>
</name>
<name>
<surname>Parasuraman</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Neuroergonomics: a review of applications to physical and cognitive work</article-title>. <source>Front. Hum. Neurosci.</source> <volume>7</volume>, <fpage>889</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2013.00889</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Menon</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Uddin</surname>
<given-names>L. Q.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Saliency, switching, attention and control: a network model of insula function</article-title>. <source>Brain Struct. Funct.</source> <volume>214</volume>, <fpage>655</fpage>&#x2013;<lpage>667</lpage>. <pub-id pub-id-type="doi">10.1007/s00429-010-0262-0</pub-id>
</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Metcalfe</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Learning from errors</article-title>. <source>Annu. Rev. Psychol.</source> <volume>68</volume>, <fpage>465</fpage>&#x2013;<lpage>489</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-psych-010416-044022</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nam</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>J.-M.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Bak</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kam</surname>
<given-names>T.-E.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>The effects of layer-wise relevance propagation-based feature selection for EEG classification: a comparative study on multiple datasets</article-title>. <source>Front. Hum. Neurosci.</source> <volume>17</volume>, <fpage>1205881</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2023.1205881</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nasri</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kosa</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chukoskie</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Moghaddam</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Harteveld</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Exploring eye tracking to detect cognitive load in complex virtual reality training</article-title>,&#x201d; in <source>2024 IEEE international symposium on mixed and augmented reality adjunct (ISMAR-Adjunct)</source> (<publisher-name>IEEE</publisher-name>), <fpage>51</fpage>&#x2013;<lpage>54</lpage>.</citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Niso</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Romero</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Moreau</surname>
<given-names>J. T.</given-names>
</name>
<name>
<surname>Araujo</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Krol</surname>
<given-names>L. R.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Wireless EEG: a survey of systems and studies</article-title>. <source>Neuroimage</source> <volume>269</volume>, <fpage>119774</fpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2022.119774</pub-id>
</citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pearson</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>1895</year>). <article-title>VII. Note on regression and inheritance in the case of two parents</article-title>. <source>Proc. R. Soc. Lond.</source> <volume>58</volume>, <fpage>240</fpage>&#x2013;<lpage>242</lpage>. <pub-id pub-id-type="doi">10.1098/rspl.1895.0041</pub-id>
</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pedregosa</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Varoquaux</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Gramfort</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Michel</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Thirion</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Grisel</surname>
<given-names>O.</given-names>
</name>
<etal/>
</person-group> (<year>2012</year>). <article-title>Scikit-learn: machine learning in python</article-title>. <source>arXiv [cs.LG]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1201.0490</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Pinheiro</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Bates</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2006</year>). <source>
<italic>Mixed-Effects models in S and S-PLUS</italic>
</source>. <publisher-name>Springer Science and Business Media</publisher-name>.</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Porssut</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Blanke</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Herbelin</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Boulic</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Adapting virtual embodiment through reinforcement learning</article-title>. <source>IEEE Trans. Vis. Comput. Graph.</source> <volume>28</volume>, <fpage>3193</fpage>&#x2013;<lpage>3205</lpage>. <pub-id pub-id-type="doi">10.1109/tvcg.2021.3057797</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Potter</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Cvetkovi&#x107;</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>De Sena</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>On the relative importance of visual and spatial audio rendering on VR immersion</article-title>. <source>Front. Signal Process</source> <volume>2</volume>. <pub-id pub-id-type="doi">10.3389/frsip.2022.904866</pub-id>
</citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pouget</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Beck</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>W. J.</given-names>
</name>
<name>
<surname>Latham</surname>
<given-names>P. E.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Probabilistic brains: knowns and unknowns</article-title>. <source>Nat. Neurosci.</source> <volume>16</volume>, <fpage>1170</fpage>&#x2013;<lpage>1178</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3495</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pratviel</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bouny</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Deschodt-Arsac</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Immersion in a relaxing virtual reality environment is associated with similar effects on stress and anxiety as heart rate variability biofeedback</article-title>. <source>Front. Virtual Real</source> <volume>5</volume>. <pub-id pub-id-type="doi">10.3389/frvir.2024.1358981</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ramsamy</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Haffegee</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Jamieson</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Alexandrov</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2006</year>). &#x201c;<article-title>Using haptics to improve immersion in virtual environments</article-title>,&#x201d; in <source>Lecture notes in computer science</source> (<publisher-loc>Berlin, Heidelberg: Springer Berlin Heidelberg</publisher-loc>: <publisher-name>Lecture notes in computer science</publisher-name>), <fpage>603</fpage>&#x2013;<lpage>609</lpage>.</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rao</surname>
<given-names>R. P.</given-names>
</name>
<name>
<surname>Ballard</surname>
<given-names>D. H.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects</article-title>. <source>Nat. Neurosci.</source> <volume>2</volume>, <fpage>79</fpage>&#x2013;<lpage>87</lpage>. <pub-id pub-id-type="doi">10.1038/4580</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sathian</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Analysis of haptic information in the cerebral cortex</article-title>. <source>J. Neurophysiol.</source> <volume>116</volume>, <fpage>1795</fpage>&#x2013;<lpage>1806</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00546.2015</pub-id>
</citation>
</ref>
<ref id="B60">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Si-mohammed</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lopes-dias</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Duarte</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jeunet</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Scherer</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Detecting system errors in virtual reality using EEG through error-related potentials</source>, <fpage>653</fpage>&#x2013;<lpage>661</lpage>.</citation>
</ref>
<ref id="B61">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singh</surname>
<given-names>A. K.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.-T.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>Y.-F.</given-names>
</name>
<name>
<surname>King</surname>
<given-names>J.-T.</given-names>
</name>
<name>
<surname>Ko</surname>
<given-names>L.-W.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Visual appearance modulates prediction error in virtual reality</article-title>. <source>IEEE Access</source> <volume>6</volume>, <fpage>24617</fpage>&#x2013;<lpage>24624</lpage>. <pub-id pub-id-type="doi">10.1109/access.2018.2832089</pub-id>
</citation>
</ref>
<ref id="B62">
<citation citation-type="journal">
<collab>Student</collab> (<year>1908</year>). <article-title>The probable error of a mean</article-title>. <source>Biometrika</source> <volume>6</volume> (<issue>1</issue>), <fpage>1</fpage>. <pub-id pub-id-type="doi">10.2307/2331554</pub-id>
</citation>
</ref>
<ref id="B63">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Takahashi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Tanaka</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tamhane</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Teng</surname>
<given-names>S.-Y.</given-names>
</name>
<name>
<surname>Lopes</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Can a smartwatch move your fingers? Compact and practical electrical muscle stimulation in a smartwatch</article-title>,&#x201d; in <source>Proceedings of the 37th annual ACM symposium on user interface software and technology</source> (<publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1</fpage>&#x2013;<lpage>15</lpage>.</citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tukey</surname>
<given-names>J. W.</given-names>
</name>
</person-group> (<year>1949</year>). <article-title>Comparing individual means in the analysis of variance</article-title>. <source>Biometrics</source> <volume>5</volume>, <fpage>99</fpage>&#x2013;<lpage>114</lpage>. <pub-id pub-id-type="doi">10.2307/3001913</pub-id>
</citation>
</ref>
<ref id="B65">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A review on transfer learning in EEG signal analysis</article-title>. <source>Neurocomputing</source> <volume>421</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2020.09.017</pub-id>
</citation>
</ref>
<ref id="B66">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Weber</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Dash</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Wriessnegger</surname>
<given-names>S. C.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Design of a virtual reality-based neuroadaptive system for treatment of arachnophobia</article-title>,&#x201d; in <source>2024 IEEE international conference on metrology for eXtended reality, artificial intelligence and neural engineering (MetroXRAINE)</source> (<publisher-name>IEEE</publisher-name>), <fpage>255</fpage>&#x2013;<lpage>259</lpage>.</citation>
</ref>
<ref id="B67">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>B.-L.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Transfer learning for EEG-Based brain&#x2013;computer interfaces: a review of progress made since 2016</article-title>. <source>IEEE Trans. Cogn. Dev. Syst.</source> <volume>14</volume>, <fpage>4</fpage>&#x2013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.1109/tcds.2020.3007453</pub-id>
</citation>
</ref>
<ref id="B68">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xavier Fid&#xea;ncio</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Klaes</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Iossifidis</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Error-related potentials in reinforcement learning-based brain-machine interfaces</article-title>. <source>Front. Hum. Neurosci.</source> <volume>16</volume>, <fpage>806517</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2022.806517</pub-id>
</citation>
</ref>
<ref id="B69">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Agarwal</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Fekri</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Sivakumar</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Accelerating reinforcement learning using EEG-Based implicit human feedback</article-title>. <source>Neurocomputing</source> <volume>460</volume>, <fpage>139</fpage>&#x2013;<lpage>153</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2021.06.064</pub-id>
</citation>
</ref>
<ref id="B70">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zander</surname>
<given-names>T. O.</given-names>
</name>
<name>
<surname>Krol</surname>
<given-names>L. R.</given-names>
</name>
<name>
<surname>Birbaumer</surname>
<given-names>N. P.</given-names>
</name>
<name>
<surname>Gramann</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Neuroadaptive technology enables implicit cursor control based on medial prefrontal cortex activity</article-title>. <source>Proc. Natl. Acad. Sci. U. S. A.</source> <volume>113</volume>, <fpage>14898</fpage>&#x2013;<lpage>14903</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1605155114</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>