<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Virtual Real.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Virtual Reality</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Virtual Real.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2673-4192</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1655545</article-id>
<article-id pub-id-type="doi">10.3389/frvir.2025.1655545</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>The impact of audience avatar movements on presence in virtual live events</article-title>
<alt-title alt-title-type="left-running-head">Guang et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frvir.2025.1655545">10.3389/frvir.2025.1655545</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Guang</surname>
<given-names>Yang</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2907259"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sakurai</surname>
<given-names>Sho</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/860478"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Matsumura</surname>
<given-names>Kohei</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn1">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3200648"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Okafuji</surname>
<given-names>Yuki</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2141956"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<institution>Hirota Laboratory, University of Electro-Communications</institution>, <city>Chofu</city>, <state>Tokyo</state>, <country country="JP">Japan</country>
</aff>
<aff id="aff2">
<label>2</label>
<institution>Playful Laboratory, Ritsumeikan University</institution>, <city>Osaka</city>, <country country="JP">Japan</country>
</aff>
<author-notes>
<corresp id="c001">
<label>&#x2a;</label>Correspondence: Yang Guang, <email xlink:href="yangguang@vogue.is.uec.ac.jp">yangguang@vogue.is.uec.ac.jp</email>
</corresp>
<fn fn-type="other" id="fn1">
<label>
<sup>&#x2020;</sup>
</label>
<p>ORCID: Kohei Matsumura, <ext-link ext-link-type="uri" xlink:href="http://orcid.org/0000-0001-6397-7255">orcid.org/0000-0001-6397-7255</ext-link>
</p>
</fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-11-13">
<day>13</day>
<month>11</month>
<year>2025</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>6</volume>
<elocation-id>1655545</elocation-id>
<history>
<date date-type="received">
<day>28</day>
<month>06</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>21</day>
<month>10</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Guang, Sakurai, Matsumura and Okafuji.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Guang, Sakurai, Matsumura and Okafuji</copyright-holder>
<license>
<ali:license_ref start_date="2025-11-13">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Virtual live (VL) events allow audiences to participate as avatars in immersive environments, but uniform and repetitive audience avatar motions can limit emotional engagement and presence. This pilot study investigates whether varying proportions of user-controlled versus dummy (scripted) avatars affect perceived presence and synchrony.</p>
</sec>
<sec>
<title>Methods</title>
<p>Seven participants experienced a VR concert under three conditions (0%, 50%, 100% user-controlled avatars) using the Oculus Quest 2. Presence was measured using the Igroup Presence Questionnaire (IPQ), and avatar-movement synchrony was evaluated using autocorrelation and cross-correlation analyses. A Synchronization Index (SI) was defined as cross-correlation between each participant&#x2019;s motion and the group-averaged motion.</p>
</sec>
<sec>
<title>Results</title>
<p>Presence scores increased with a higher proportion of user-controlled avatars. SI also tended to increase in the 100% condition. Although correlations between SI and presence were moderate (r &#x3d; .42) and did not reach statistical significance due to the pilot sample size, both subjective and objective trends supported the hypothesized direction.</p>
</sec>
<sec>
<title>Discussion</title>
<p>User-controlled audience avatars enhanced presence compared to scripted avatars, suggesting movement diversity contributes to immersive VL experiences. This study demonstrates a feasible evaluation pipeline for avatar synchrony and highlights design implications for VR live events. Future work will expand sample size, incorporate additional behavior modalities, and refine synchrony metrics.</p>
</sec>
</abstract>
<kwd-group>
<kwd>virtual reality</kwd>
<kwd>virtual live events</kwd>
<kwd>avatars</kwd>
<kwd>user interaction</kwd>
<kwd>synchronization in VR</kwd>
<kwd>immersive presence</kwd>
<kwd>emotional engagement</kwd>
<kwd>cross-correlation</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declare that no financial support was received for the research and/or publication of this article. This research did not involve any commercial or financial relationships, and no competing interests exist.</funding-statement>
</funding-group>
<counts>
<fig-count count="11"/>
<table-count count="2"/>
<equation-count count="0"/>
<ref-count count="44"/>
<page-count count="15"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Virtual Reality and Human Behaviour</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>In virtual reality (VR) research, immersion is commonly defined following Slater and Wilbur&#x2019;s framework as the extent to which a system delivers an enveloping and coherent sensory environment that supports user involvement (<xref ref-type="bibr" rid="B32">Slater and Wilbur, 1997</xref>). Presence refers to the subjective sensation of &#x201c;being there&#x201d; in the virtual environment, as conceptualized by Lombard and Ditton and measured via instruments such as the Presence Questionnaire by Witmer and Singer (<xref ref-type="bibr" rid="B19">Lombard and Ditton, 1997</xref>; <xref ref-type="bibr" rid="B43">Witmer and Singer, 1998</xref>). In this study, we adopt these definitions to ensure consistency: immersion in the sense of environmental enveloping (<xref ref-type="bibr" rid="B32">Slater and Wilbur, 1997</xref>) and presence as the subjective experience measured by validated questionnaires (<xref ref-type="bibr" rid="B43">Witmer and Singer, 1998</xref>).</p>
<p>Virtual Live events (VL events) refer to music concerts, theater performances, virtual tours, and similar experiences that take place entirely within virtual environments rather than physical venues. Unlike traditional live streaming, VL events typically leverage VR avatars to enable deeper immersion and novel forms of user interaction unique to VR platforms. Key characteristics include participants&#x2019; interactive engagement via avatars, a heightened sense of self-projection in the virtual space, and the potential for synchronous, co-located experiences in distributed settings. During the COVID-19 pandemic, VL events emerged as an important alternative for entertainment under social distancing constraints; surveys and industry reports indicate a surge in participation and development of VR-based performances (<xref ref-type="bibr" rid="B33">Swarbrick et al., 2021</xref>; <xref ref-type="bibr" rid="B44">XR Association, 2023</xref>). Consequently, VL performances have drawn research interest as platforms for immersive user experiences.</p>
<p>Building on this context, our work examines one of the critical challenges reported by users&#x2014;namely, the perceived lack of presence in audience roles&#x2014;and investigates how avatar movement diversity may mitigate this issue.</p>
<p>While capable of rendering live stages and supporting avatar-based participation, existing VL platforms such as <xref ref-type="bibr" rid="B40">VRChat (2022)</xref>, <xref ref-type="bibr" rid="B4">Cluster (2023)</xref>, and Project Sekai (<xref ref-type="bibr" rid="B30">SEGA Co., Ltd, 2024</xref>) have been reported to exhibit limitations in user-perceived presence in the audience role. Industry surveys and user studies suggest that audience avatars often lack sufficient expressiveness, and interactions are constrained by uniform movement patterns that fail to convey individual emotions or personality cues (<xref ref-type="bibr" rid="B42">Waltemate et al., 2018</xref>).</p>
<p>One promising direction is to adjust audience interaction in virtual live events to more closely approximate real-world social dynamics, thereby improving subjective presence (<xref ref-type="bibr" rid="B10">Gonzalez-Franco and Lanier, 2017</xref>; <xref ref-type="bibr" rid="B29">Sebanz et al., 2006</xref>). Although full-body motion capture systems (e.g., Xsens or Perception Neuron) can offer richer individual movements and expressive potential (<xref ref-type="bibr" rid="B41">VRChat, 2023</xref>), such approaches entail high costs and technical barriers and do not guarantee enhanced social interaction among audience avatars by themselves. Research on group synchrony indicates that merely enabling realistic motion is insufficient.Rather, context-sensitive and diversity-oriented synchronization mechanisms are required to foster emotional engagement (<xref ref-type="bibr" rid="B34">Tarr et al., 2012</xref>; <xref ref-type="bibr" rid="B12">Hatfield et al., 1993</xref>). Therefore, improving presence requires the exploration of cost-effective technical solutions in tandem with carefully designed interaction conditions. In this exploratory pilot study, we examine how varying the diversity of audience avatar movements and synchronization patterns affects the subjective presence in VL event settings.</p>
<p>Despite advances in immersive VR performances, systematic methods to quantify avatar synchronization effects remain scarce. To address this gap, we propose a mixed-methods framework combining validated presence scales with a synchronization index, and conduct a pilot study to test how user-driven versus controlled (human-driven) avatars affect presence. Based on these findings, we outline initial design implications for large-scale VR audiences.</p>
<p>In this study, we expand the definition of presence to encompass not only spatial immersion but also emotional engagement elicited through avatar interactions. Building on established instruments such as the Slater-Usoh-Steed questionnaire and the Igroup Presence Questionnaire (<xref ref-type="bibr" rid="B38">Usoh et al., 2000</xref>; <xref ref-type="bibr" rid="B27">Schubert et al., 2001a</xref>), we incorporate additional exploratory measures related to avatar movement diversity and synchronization patterns. Details of these measures, including the combination of subjective presence scores with motion-capture data to assess avatar-driven interactions, are described later in the paper.</p>
<p>In this study, we assess presence in VL events through subjective immersion and emotional engagement measures collected via validated questionnaires and complemented by exploratory motion-capture metrics. Here, presence is conceptualized primarily as the audience&#x2019;s emotional involvement and sense of co-presence facilitated by avatar interactions, with a focus on movement diversity and synchronization quality rather than mere interaction frequency. Building on established methodologies, our questionnaire-based approach employs adapted items from the Presence Questionnaire (<xref ref-type="bibr" rid="B43">Witmer and Singer, 1998</xref>) alongside additional questions tailored to capture perceptions of avatar movement realism and social connection. Details of the questionnaire design and analysis procedures are described later in the paper.</p>
<p>The objective of this exploratory pilot study is to examine how varying the diversity of audience avatar movements, from uniform scripted patterns to richer, user-like variations, affects subjective presence and emotional engagement in VL event settings. By addressing the challenge of limited audience interaction in current VL platforms, we investigate movement diversity as a factor that influences presence. Although conducted with a small sample and focused interaction modality (e.g., synchronized light-stick gestures), this study aims to generate preliminary insights and propose initial design guidelines for VR event experiences that support enhanced presence. These findings will inform future large-scale studies and platform design strategies.</p>
<p>Building on foundational frameworks of social synchrony, notably the joint action theory (<xref ref-type="bibr" rid="B29">Sebanz et al., 2006</xref>) and emotional contagion models (<xref ref-type="bibr" rid="B12">Hatfield et al., 1993</xref>), we posit that coordinated and diverse avatar movements can amplify co-presence beyond mere visual immersion. Recent advances in entertainment computing further demonstrate that interactive media platforms generate powerful illusions of social unity (<xref ref-type="bibr" rid="B10">Gonzalez-Franco and Lanier, 2017</xref>). In this study, we (1) introduce a robust evaluation framework combining psychometric presence scales with a novel motion-capture-based Synchronization Index, (2) test how varying proportions of synchronized versus dummy (scripted) avatars modulate emotional engagement and presence, and (3) extract design principles for scalable, cost-effective audience interaction in VL events. Together, these contributions both extend VR presence theory and provide practical guidelines for next-generation digital entertainment systems.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related work: research on face-to-face and avatar movement coordination</title>
<p>Research on remote group synchrony indicates that simple gestures, such as virtual &#x201c;high-fives&#x201d; or collective waves, can foster a shared sense of unity among distributed viewers and boost immersive engagement (<xref ref-type="bibr" rid="B24">PutturVenkatraj et al., 2024</xref>; <xref ref-type="bibr" rid="B29">Sebanz et al., 2006</xref>). In large physical audiences, emergent synchronization of handclapping has been linked to collective emotional alignment (<xref ref-type="bibr" rid="B21">N&#xe9;da et al., 2000</xref>), and synchronized arousal between performers and spectators has been demonstrated in ritual contexts (<xref ref-type="bibr" rid="B16">Konvalinka et al., 2011</xref>). In VR sports viewing, studies show that these synchronized non-verbal cues convey emotional states even in the absence of speech, thereby enhancing subjective presence (<xref ref-type="bibr" rid="B15">Kimmel et al., 2024</xref>; <xref ref-type="bibr" rid="B12">Hatfield et al., 1993</xref>). Nevertheless, the precise pathways through which real-time emotional expression via avatar movement heightens presence remain underexplored.</p>
<p>In parallel, seminal work by (<xref ref-type="bibr" rid="B7">Dahl and Friberg, 2007</xref>) established that movement parameters, particularly velocity and amplitude, critically shape emotional transmission in physical performances, however their findings have to be fully translated to VR settings yet. Subsequent VR research has demonstrated that avatar motions aligned with musical or performative cues can function as proxies for emotional expression and elevate presence levels (<xref ref-type="bibr" rid="B31">Shlizerman et al., 2018</xref>), but systematic investigation of these effects under varying synchronization conditions is lacking.</p>
<p>Extending beyond lab-scale experiments, large-scale virtual concerts such as Ariana Grande&#x2019;s Fortnite performance have exemplified how massive, real-time avatar synchronization can produce powerful collective experiences, suggesting design principles for next-generation VL platforms (<xref ref-type="bibr" rid="B13">Hatmaker, 2021</xref>). Meanwhile, (<xref ref-type="bibr" rid="B25">Rogers et al., 2022</xref>) compared real-time full-body avatar interactions with face-to-face communication and found comparable levels of user comfort and engagement, yet their study did not directly measure presence or unity in live-event contexts.</p>
<p>Entertainment computing research further underscores that interactive VR environments can generate compelling illusions of presence (<xref ref-type="bibr" rid="B10">Gonzalez-Franco and Lanier, 2017</xref>), and that user satisfaction in simulated live performances correlates with avatar responsiveness and environmental interactivity (<xref ref-type="bibr" rid="B18">Liaw et al., 2020</xref>).Together, these bodies of work point to a gap in understanding how real-time diversification and synchronization of audience avatar movements jointly contribute to emotional expression, social unity, and the subjective sense of &#x201c;being there.&#x201d; Cross-correlation approaches to quantify movement coordination have been established in prior work (<xref ref-type="bibr" rid="B5">Cornejo et al., 2018</xref>; <xref ref-type="bibr" rid="B6">2023</xref>). Our study extends this line by applying avatar&#x2013;audio coupling in VL audience settings and by integrating beat-locked motion features into a Synchronization Index (SI. Recent advances in AI-driven avatar animation further illustrate this potential: (<xref ref-type="bibr" rid="B35">Ullal et al., 2021</xref>; <xref ref-type="bibr" rid="B36">Ullal et al., 2022</xref>) proposed multi-objective optimization frameworks for expressive gestures in AR/MR; <xref ref-type="bibr" rid="B14">Juravsky et al., 2024</xref> introduced SuperPADL, enabling scalable language-directed physics-based motion control; and <xref ref-type="bibr" rid="B1">Bourgault et al. (2025)</xref> presented Narrative Motion Blocks, combining natural language with direct manipulation for animation creation. Our study therefore investigates how varying degrees of movement diversity among audience avatars modulate presence and unity in VL events, addressing this critical research gap.</p>
<p>Based on previous studies, the real-time reflection of human movements in audience avatars during VL events may create experiences similar to those of actual live events. This study aims to explore how diversifying the movement patterns of audience avatars enriches emotional expression and influences presence in VL events.</p>
</sec>
<sec id="s3">
<label>3</label>
<title>Experimental verification of avatar proportion and presence in digital entertainment</title>
<p>To investigate how varying the proportion of avatar types influences the presence of audience avatars, this section outlines the experimental design and methods employed in this study.</p>
<p>Terminology: In this paper, we use consistent terms for audience avatars. Program-controlled avatars are referred to as &#x201c;controlled (human-driven) avatars&#x201d; (replacing earlier labels such as &#x201c;pre-programmed&#x201d; or simply &#x201c;dummy&#x201d;). Avatars driven by live participants are referred to as &#x201c;user-controlled (human-driven) avatars&#x201d; (replacing earlier terms such as &#x201c;user avatars&#x201d; or &#x201c;human avatars&#x201d;). These unified terms are applied consistently throughout the manuscript, including figure captions.</p>
<sec id="s3-1">
<label>3.1</label>
<title>Evaluation framework</title>
<p>In order to rigorously assess avatar synchronization effects, we developed a dual-track evaluation framework. First, subjective presence was measured using the Igroup Presence Questionnaire (IPQ) (<xref ref-type="bibr" rid="B28">Schubert et al., 2001b</xref>), comprising Spatial Presence, Involvement, and Experienced Realism subscales scored on a seven-point (0&#x2013;6) scale. Internal consistency in our pilot yielded Cronbach&#x2019;s <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.84</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. In addition, we included exploratory &#x201c;unity&#x201d; and &#x201c;empathy&#x201d; items, which were not part of the validated IPQ but were designed to probe potential interpersonal aspects of VL presence.Second, we define a Synchronization Index (SI) as the maximum normalized cross-correlation coefficient between each participant&#x2019;s glow-stick motion and the <italic>group-averaged audience motion</italic> time series, computed within a fixed lag window and normalized to [0,1]. In parallel, we report beat-alignment CCFs between individual motion and the 140 BPM reference sequence as a separate, exploratory objective metric (not part of SI). Autocorrelation functions (ACF) quantify periodicity, while CCF peaks locate optimal lag alignment (<xref ref-type="bibr" rid="B3">Chatfield, 2003</xref>). This framework ensures both psychometric validity and objective quantification of movement coupling prior to hypothesis testing. The Synchronization Index (SI) was defined as the maximum normalized cross-correlation coefficient between each participant&#x2019;s motion and the group-averaged signal, whereas <xref ref-type="fig" rid="F10">Figures 10</xref>, <xref ref-type="fig" rid="F11">11</xref> present pairwise CCF values between participants and individual avatars.</p>
</sec>
<sec id="s3-2">
<label>3.2</label>
<title>Objective</title>
<p>This pilot investigation tests two formal hypotheses. H1: Increasing the proportion of user-controlled, rhythm-synchronized avatars will produce significantly higher subjective presence and social unity scores than uniform, dummy (scripted) avatars. H2: The novel Synchronization Index (SI), defined via peak cross-correlation between glow-stick motion and audio beats, will correlate strongly (<inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) with Presence Questionnaire scores. We evaluate these hypotheses through a mixed-methods design: (a) validated psychometric instruments (Igroup Presence Questionnaire (<xref ref-type="bibr" rid="B43">Witmer and Singer, 1998</xref>), Emotional Empathy Scale (<xref ref-type="bibr" rid="B8">Davis, 1983</xref>)) and (b) quantitative motion-capture analyses (ACF/CCF). This structured approach clarifies our theoretical contributions and sets the stage for subsequent confirmatory studies.</p>
<p>In this study, special emphasis is placed on the movement of glow sticks held by the avatars. A VL environment is constructed to focus on this design aspect. As shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, dummy avatars are programmed to perform repetitive, unsynchronized glow stick movements, while user-controlled (human-driven) avatars mimic real-time user movements in the VR space. In the VL environment, an experimental collaborator operates an avatar surrounded by two types of avatars: (1) dummy avatars that wave glow sticks with constant, non-synchronized motions, and (2) user-controlled (human-driven) avatars that display real-time synchronized movements with the music rhythm. This setup enables us to examine how differences between dummy and user-controlled (human-driven) avatar behaviors affect audience interaction and the overall sense of presence.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Comparison of dummy avatars and user-controlled (human-driven) avatars. Dummy avatars perform pre-programmed motions, while user-controlled (human-driven) avatars reflect real-time human movements, enabling enhanced emotional expressiveness and interaction.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g001.tif">
<alt-text content-type="machine-generated">A comparison between a dummy (scripted) avatar and a user-controlled (human-driven) avatar in a virtual live space. The top image shows a dummy avatar performing pre-scripted glow-stick waving motions, indicated by arrows showing fixed movement directions. The bottom image shows a user-controlled avatar whose glow-stick movements reflect real-time human controller input, with arrows illustrating dynamic motion. The scene highlights the contrast between program-generated animation and live human motion in a virtual audience context.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-3">
<label>3.3</label>
<title>Experimental environment</title>
<p>The experiment was implemented in Unity 2019.4.3f1 (Unity Technologies) using the asset &#x201c;Unity-chan Live Stage! &#x2013; Candy Rock Star &#x2013; (v1.0)&#x201d; (<xref ref-type="bibr" rid="B37">UNITY-CHAN!OFFICIAL-WEBSITE, 2014</xref>). Avatar synchronization employed Photon Unity Networking 2 (PUN2) (<xref ref-type="bibr" rid="B23">Photon Unity Networking, 2022</xref>). All experimental sessions in this study were conducted under wired LAN conditions (mean &#x3d; 15&#xa0;m, SD &#x3d; 3&#xa0;m) to minimize variability. Wireless VR setups were measured only during pilot testing and are reported for reference. No predictive interpolation was applied during the experiment; instead, <italic>post hoc</italic> cross-correlation analysis was performed with <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mo>&#xb1;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 1&#xa0;s lag windows, which absorbed small transmission delays. This approach ensured consistency across sessions conducted under wired LAN. These bounds informed our data interpretation, as preliminary analysis confirmed that CCF peaks remained robust within this latency range. We discuss potential interpolation strategies to mitigate residual desynchronization in Section 6. The virtual stage comprised one performer avatar with pre-scripted animations synchronized to a 140 BPM audio track, surrounded by five audience avatars arranged in a semicircle as shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. Dummy avatars executed fixed glow-stick waving motions at 2&#xa0;Hz, whereas user-controlled (human-driven) avatars mirrored participants&#x2019; real-time Oculus Quest two controller rotations (sampled at 90&#xa0;Hz). All glow-stick orientation data were recorded via PUN2 for subsequent objective analysis.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>A diagram showing an experimental environment. This includes the spatial arrangement of avatars, differentiating between user-controlled avatars and dummy avatars performing pre-programmed motions. The layout clarifies the positions of each avatar type during the VL event.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g002.tif">
<alt-text content-type="machine-generated">Diagram illustrating the experimental setup. The left panel shows the physical space with a participant wearing a head-mounted display and holding motion controllers, tracked on a table. The right panel shows the mapped virtual environment with a performer avatar on stage and audience avatars positioned around. User-controlled avatars are shown in black and dummy (scripted) avatars in blue. Arrows show correspondence between physical participant position and virtual avatar position.</alt-text>
</graphic>
</fig>
<p>As illustrated in <xref ref-type="fig" rid="F3">Figure 3</xref>, the stimuli used in this study included a performer avatar and audience avatars with uniform motions. This figure, originally placed in the Introduction, is moved here to clarify that it represents the experimental setup specific to our study. As illustrated in <xref ref-type="fig" rid="F4">Figure 4</xref>, the immersive virtual live venue includes a stage-front perspective and audience area context, providing participants with a realistic concert environment.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Stimuli used in the present study: performer avatar and audience avatars with uniform motions. This figure is part of the experimental setup rather than prior literature examples.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g003.tif">
<alt-text content-type="machine-generated">Virtual stage scene featuring a performer avatar standing under stage lighting and star decorations. Two audience avatars in silhouette hold glow-sticks in the foreground, representing spectators in a virtual live performance environment.</alt-text>
</graphic>
</fig>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>The experimental virtual live stage. The stage design includes a performer avatar and a group of audience avatars, showcasing their arrangement and the interaction environment created for the study.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g004.tif">
<alt-text content-type="machine-generated">Three-dimensional virtual concert stage with an anime-style performer avatar illuminated by spotlights. Audience silhouettes holding glow-sticks face the stage. Lighting towers and speakers surround the performer, representing an immersive live concert environment.</alt-text>
</graphic>
</fig>
<p>As depicted in <xref ref-type="fig" rid="F5">Figure 5</xref>, the participants can control the user-controlled (human-driven) avatars by moving the motion controller as if they were holding a glow stick. The glow stick movements correspond to those of the Oculus Quest 2 VR controllers, with the black circular part serving as a grip handle.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>A first-person view of the controller setup used in the experimental environment. The figure illustrates the participant&#x2019;s perspective while holding the glow stick using the Oculus Quest two controller. The black circular part serves as the handle for gripping, enabling precise movements synchronized with the virtual environment.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g005.tif">
<alt-text content-type="machine-generated">Top image shows a virtual avatar on stage with 3D axes displaying X, Y, and Z rotational directions linked to glow-stick movement. Bottom image shows a user holding an Oculus Quest 2 motion controller with the same axis markers, illustrating how real-world controller rotation maps directly to virtual avatar glow-stick motion.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-4">
<label>3.4</label>
<title>Experimental design</title>
<p>To systematically evaluate synchronization effects, we employed three conditions of user-to-dummy avatar ratios: 0%, 50%, and 100%. In each session, one human participant was present and controlled their avatar in real time. The remaining audience avatars were either dummy (scripted) avatars or user-controlled (human-driven) avatars replaying the participant&#x2019;s real-time movements. Thus, the 50% and 100% conditions indicate the proportion of user-driven avatars among the four audience avatars. The 50% condition was chosen to represent mixed interactions, reflecting typical medium-sized VR audiences. As shown in <xref ref-type="fig" rid="F6">Figure 6</xref>, the VL environment includes one performer avatar, one avatar reflecting an experiment participant&#x2019;s movements, and four audience avatars. We used a Latin-square design to counterbalance condition order across participants, minimizing sequence effects. Each participant completed three 3-min sessions, with condition assignment randomized via computer script prior to each run. This design ensures equitable exposure to each synchronization level and controls for order bias.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>The ratio and spatial arrangement of avatars in the experimental virtual live stage. This layout features one performer avatar and audience avatars (both user-controlled and dummy), highlighting their positions under different experimental conditions (0%, 50%, 100% user-controlled (human-driven) avatars) to evaluate their impact on presence.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g006.tif">
<alt-text content-type="machine-generated">Diagram showing three audience avatar ratio conditions in a virtual live stage: 0% user-controlled (five dummy avatars), 50% user-controlled (two dummy and two user-controlled avatars), and 100% user-controlled (five live avatars). The performer avatar appears on stage in all conditions.</alt-text>
</graphic>
</fig>
<p>During the experiment, the song package &#x2018;Unity-chan Live Stage! -Candy Rock Star-(version 1.0)&#x2019; (<xref ref-type="bibr" rid="B37">UNITY-CHAN!OFFICIAL-WEBSITE, 2014</xref>) was used. The performer avatar, Unity-chan, provided consistent musical and motion stimuli, while the focus was on analyzing how real-time synchronized and diverse audience avatar gestures contribute to an enhanced sense of presence beyond the performer&#x2019;s actions. The experiment varied the ratio of actual users (user-controlled (human-driven) avatars) and program-controlled avatars (dummy avatars) across three conditions (0%, 50%, 100% user-controlled (human-driven) avatars), with participants controlling avatars via Oculus Quest 2 HMDs.</p>
</sec>
<sec id="s3-5">
<label>3.5</label>
<title>Evaluation design for audience presence in VL environment</title>
<p>In addition to IPQ, we included three author-designed items intended to probe perceived &#x201c;unity/empathy&#x201d; during audience co-action. These items are exploratory (non-validated) and are reported descriptively; they are not used for confirmatory hypothesis testing.</p>
<p>Our evaluation integrates (1) psychometric reliability checks (Cronbach&#x2019;s <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.84 for presence, <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.81 for unity) and (2) rigorous statistical analyses. A one-way repeated-measures ANOVA revealed a significant effect of avatar synchronization ratio on presence (F (2,12) &#x3d; 5.23, p &#x3d; 0.022, <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.47). Given the small N, we additionally conducted Kruskal&#x2013;Wallis tests on individual Likert items (H &#x3d; 7.12, p &#x3d; 0.028), confirming the ANOVA pattern without normality assumptions (<xref ref-type="bibr" rid="B17">Kruskal and Wallis, 1952</xref>). Finally, a <italic>post hoc</italic> power analysis (G&#x2a;Power 3.1) indicated approximately 60% power to detect large effects (Cohen&#x2019;s <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). The repeated-measures ANOVA yielded <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>2,12</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5.23</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, which exceeded the critical value (<inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">crit</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3.89</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). However, <italic>post hoc</italic> pairwise comparisons did not consistently reach significance after correction, and thus the results should be interpreted as exploratory. Effect sizes (<inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.47</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) are reported alongside <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-values to contextualize the observed trends.</p>
<p>For objective measures, we computed ACF and CCF on glow-stick rotation data following Chatfield&#x2019;s methodology (<xref ref-type="bibr" rid="B3">Chatfield, 2003</xref>). We focused our motion analysis on the x-axis rotation, as the glow-stick waving gesture was primarily lateral and this axis accounted for the greatest proportion of variance with stable periodic peaks. The y- and z-axes were more affected by controller drift and noise, and were excluded to ensure robustness. Raw angular displacement values were analyzed rather than angular velocity. The mean peak CCF increased linearly across conditions (0%: M &#x3d; 0.12; 50%: M &#x3d; 0.25; 100%: M &#x3d; 0.40), with a significant linear trend in ANOVA (F (1,6) &#x3d; 9.78, p &#x3d; 0.019). Although SI showed a moderate positive correlation with presence scores (<inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.42</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.13</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>), this relationship did not reach significance. Therefore, H2 was only partially supported. We interpret SI as a complementary indicator of presence that captures aspects of timing-based coupling but requires refinement through additional features such as inter-avatar phase-locking and tempo-variance measures. Additionally, the participants&#x2019; sense of presence was evaluated by using Scheffe&#x2019;s paired comparison method (<xref ref-type="bibr" rid="B26">Scheff&#xe9;, 1952</xref>), as illustrated in <xref ref-type="fig" rid="F7">Figure 7</xref>.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>The experimental evaluation scale used to measure the sense of presence in different VL event scenarios. The scale employs Scheffe&#x2019;s paired comparison method, asking participants to compare two scenes and rate which one provides a higher sense of presence on a five-point scale.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g007.tif">
<alt-text content-type="machine-generated">Evaluation scale used to compare perceived presence between live scenes. The scale has three items rated from &#x2212;1 to &#x002B;1, where participants select which scene provided a stronger sense of presence. Labels range from &#x201c;Absolutely the first&#x201d; to &#x201c;Absolutely the third.&#x201d;</alt-text>
</graphic>
</fig>
<p>For the motion analysis, motion data from the avatars&#x2019; glow sticks were analyzed using autocorrelation (ACF) (<xref ref-type="bibr" rid="B38">Usoh et al., 2000</xref>; <xref ref-type="bibr" rid="B2">Brockwell and Davis, 1991</xref>) and cross-correlation functions (CCF) (<xref ref-type="bibr" rid="B38">Usoh et al., 2000</xref>; <xref ref-type="bibr" rid="B27">Schubert et al., 2001a</xref>) to quantify synchronization with the music and among avatars. Peaks in the ACF and variations in the CCF provided metrics for determining whether participants&#x2019; movements were periodic and synchronized with the music&#x2019;s rhythm. A decrease in correlation coefficients with increasing proportions of user-controlled (human-driven) avatars was also investigated (<xref ref-type="bibr" rid="B39">Vivanco and Jayasumana, 2007</xref>).</p>
</sec>
<sec id="s3-6">
<label>3.6</label>
<title>Experimental procedure</title>
<p>Participants first completed a brief tutorial phase in which dummy avatars executed unsynchronized glow-stick motions at fixed intervals, and user-controlled (human-driven) avatars moved in tandem with a 140 BPM rhythm under direct controller input. A brief detection task confirmed the participants&#x2019; ability to distinguish human-driven avatars from programmatic avatars. Subsequently, each participant performed three experimental sessions under 0%, 50%, and 100% user-avatar conditions in a Latin-square counterbalanced order. Each session lasted 3&#xa0;minutes. Immediately after each session, participants completed the presence and empathy questionnaires, and brief semi-structured interviews were conducted to collect qualitative feedback. Finally, all motion and questionnaire data were aggregated for statistical analysis.</p>
</sec>
<sec id="s3-7">
<label>3.7</label>
<title>Participants</title>
<p>Seven university students (five male and two female; mean age &#x3d; 23.1 years, SD &#x3d; 1.4; range 21&#x2013;25 years) were recruited via campus-wide email. None had prior experience with VL events, which helped reduce expectancy effects. As this pilot study was conducted entirely within our laboratory setting and involved minimal risk procedures, the study did not undergo a formal ethics review process at Ritsumeikan University. All participants provided written informed consent after receiving a clear explanation of the study&#x2019;s objectives, procedures, and data management. A <italic>post hoc</italic> power analysis using G&#x2a;Power (<xref ref-type="bibr" rid="B9">Faul et al., 2007</xref>) showed approximately 60% power to detect large effects (Cohen&#x2019;s d &#x3d; 0.8) in repeated-measures analyses, highlighting the exploratory nature of the current sample size.</p>
<p>Sensitivity analysis: Given the within-subjects design and <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, a sensitivity analysis (G&#x2a;Power 3.1; <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mn>05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) indicated that only large effects would be detectable with adequate power in our setting. We therefore frame the study as a pilot and emphasize estimation and hypothesis generation over strict null-hypothesis significance testing.</p>
</sec>
<sec id="s3-8">
<label>3.8</label>
<title>Results</title>
<p>This section presents the findings of our study, emphasizing both subjective and objective evaluations of presence under varying experimental conditions and discussing their implications for digital entertainment design.</p>
<sec id="s3-8-1">
<label>3.8.1</label>
<title>Subjective evaluation</title>
<p>In this pilot study, we report presence and unity as the primary subjective outcomes. Exploratory subscales (emotional susceptibility, independence, interdependence) were inspected qualitatively and showed the same directional pattern, but they are not used for hypothesis testing given the small sample size. The average emotional susceptibility scores ranged from 2.5 to 3.6 (<xref ref-type="table" rid="T1">Table 1</xref>), while independence and interdependence scores showed expected variations among individuals. Internal consistency was satisfied (Cronbach&#x2019;s <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.78&#x2013;0.85). A one-way repeated-measures ANOVA on emotional susceptibility revealed a significant main effect of avatar synchronization ratio (F (2,12) &#x3d; 4.56, p &#x3d; 0.034, <inline-formula id="inf17">
<mml:math id="m17">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 0.43), with <italic>post hoc</italic> Scheff&#xe9; tests showing higher susceptibility in the 100% user-avatar condition (M &#x3d; 3.60, SD &#x3d; 0.35) than in the 0% condition (M &#x3d; 2.50, SD &#x3d; 0.45; p &#x3d; 0.021). Independence did not differ significantly across conditions (F (2,12) &#x3d; 2.10, p &#x3d; 0.154), nor did interdependence (F (2,12) &#x3d; 1.85, p &#x3d; 0.196). These results complement the overall presence findings and highlight that emotional susceptibility is particularly sensitive to avatar synchronization levels. Complementary Kruskal&#x2013;Wallis tests on individual Likert-scale items confirmed these patterns without assuming normality (<xref ref-type="bibr" rid="B17">Kruskal and Wallis, 1952</xref>). Preference data derived from Scheff&#xe9;&#x2019;s paired comparison (<xref ref-type="table" rid="T2">Table 2</xref>; <xref ref-type="fig" rid="F8">Figure 8</xref>) further showed that nearly half of the participants favored the all-user-avatar scenario in terms of presence (mean preference &#x3d; 0.28) compared to the mixed and dummy-only conditions (<xref ref-type="bibr" rid="B26">Scheff&#xe9;, 1952</xref>; <xref ref-type="bibr" rid="B20">Miranda, 2000</xref>). These converging results underscore that higher synchronization enhances both emotional engagement and the subjective sense of presence.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Participant questionnaire evaluation results regarding the sense of presence under different avatar movement ratio conditions. All values are means across participants.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Question/Subject</th>
<th align="left">Influence of feelings (average)</th>
<th align="left">Mutual independence (average)</th>
<th align="left">Mutual harmonization (average)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Subject 1</td>
<td align="center">2.6</td>
<td align="center">3</td>
<td align="center">4</td>
</tr>
<tr>
<td align="center">Subject 2</td>
<td align="center">3.6</td>
<td align="center">3.5</td>
<td align="center">3.5</td>
</tr>
<tr>
<td align="center">Subject 3</td>
<td align="center">2.8</td>
<td align="center">3.25</td>
<td align="center">3</td>
</tr>
<tr>
<td align="center">Subject 4</td>
<td align="center">2.8</td>
<td align="center">4.25</td>
<td align="center">2</td>
</tr>
<tr>
<td align="center">Subject 5</td>
<td align="center">3.2</td>
<td align="center">3.5</td>
<td align="center">3.25</td>
</tr>
<tr>
<td align="center">Subject 6</td>
<td align="center">2.8</td>
<td align="center">4</td>
<td align="center">3.5</td>
</tr>
<tr>
<td align="center">Subject 7</td>
<td align="center">2.6</td>
<td align="center">4</td>
<td align="center">3.5</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>1:Strongly disagree 2:disagree 3:Not sure 4: Agree 5:Strongly agree.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Results of the integral comparison method for presence evaluation under different avatar ratio conditions. Higher scores indicate a greater sense of presence.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Conditions/Pairwise comparison</th>
<th align="center">0</th>
<th align="center">50</th>
<th align="center">100</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0</td>
<td align="center">&#x2013;</td>
<td align="center">1</td>
<td align="center">2</td>
</tr>
<tr>
<td align="left">50</td>
<td align="center">&#x2212;1</td>
<td align="center">&#x2013;</td>
<td align="center">4</td>
</tr>
<tr>
<td align="left">100</td>
<td align="center">&#x2212;2</td>
<td align="center">&#x2212;4</td>
<td align="center">&#x2013;</td>
</tr>
<tr>
<td align="left">Xi</td>
<td align="center">&#x2212;3</td>
<td align="center">&#x2212;3</td>
<td align="center">6</td>
</tr>
<tr>
<td align="left">Xj</td>
<td align="center">3</td>
<td align="center">3</td>
<td align="center">&#x2212;6</td>
</tr>
<tr>
<td align="left">Xi&#x2013;Xj</td>
<td align="center">&#x2212;6</td>
<td align="center">&#x2212;6</td>
<td align="center">12</td>
</tr>
<tr>
<td align="left">Average Liking</td>
<td align="center">&#x2212;0.14</td>
<td align="center">&#x2212;0.14</td>
<td align="center">0.28</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Evaluation results using Scheffe&#x2019;s paired comparison method (<xref ref-type="bibr" rid="B26">Scheff&#xe9;, 1952</xref>) for differences in the sense of presence.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g008.tif">
<alt-text content-type="machine-generated">Line chart showing presence ratings for three avatar ratio conditions. A red triangle represents 0% user-controlled avatars, a blue circle represents 50% user-controlled avatars, and a yellow square represents 100% user-controlled avatars, plotted on a &#x2212;1 to &#x002B;1 presence scale.</alt-text>
</graphic>
</fig>
<p>Post-experiment interviews confirm these findings. Many participants indicated that an increased number of user-controlled avatarsthat exhibit synchronized movements with the music, enhanced their sense of presence. Several participants noted that synchronization with the music rhythm significantly improved the immersive experience, while non-synchronized movements felt unnatural. These subjective insights underline the importance of synchronization in enhancing interaction, which is critical to designing engaging digital live events.</p>
</sec>
<sec id="s3-8-2">
<label>3.8.2</label>
<title>Objective evaluation</title>
<p>To quantify movement synchronization, glow-stick rotation time series were analyzed using autocorrelation (ACF) and cross-correlation (CCF) methods as described by Chatfield and Collins (<xref ref-type="bibr" rid="B3">Chatfield, 2003</xref>). Autocorrelation functions (<xref ref-type="fig" rid="F9">Figure 9</xref>) showed clear peaks at approximately 43 lags, corresponding to the 140 BPM beat and confirming the participants&#x2019; periodic movements. The cross-correlation coefficients between each participant&#x2019;s movements and the group average increased systematically with higher synchronization ratios (0%: M &#x3d; 0.12, SD &#x3d; 0.05; 50%: M &#x3d; 0.25, SD &#x3d; 0.08; 100%: M &#x3d; 0.40, SD &#x3d; 0.10). A linear contrast in a repeated-measures ANOVA confirmed a significant linear trend (F (1,6) &#x3d; 9.78, p &#x3d; 0.019), indicating that greater avatar synchronization yields stronger objective alignment. Throughout Results, &#x201c;CCF with individual avatars&#x201d; refers to pairwise cross-correlations for interpretation, whereas SI always denotes coupling with the group-average motion, not with the audio beat. While perfect synchrony was not observed, distinct CCF peaks (<xref ref-type="fig" rid="F10">Figures 10</xref>, <xref ref-type="fig" rid="F11">11</xref>) provide quantifiable evidence of interaction dynamics. Notably, under the 50% condition, participants sometimes synchronized more strongly with dummy avatars than with the live user-controlled (human-driven) avatar. This does not contradict the overall finding that group-level SI increased with higher human ratios, since SI aggregates across all avatars and reflects alignment with the group average. We interpret this as evidence that local synchrony with scripted avatars may occur even when global synchrony is enhanced by the presence of more human-controlled avatars. We flag this observation explicitly as exploratory given the small <inline-formula id="inf18">
<mml:math id="m18">
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and we avoid generalizing beyond the present sample; future work should test this pattern with larger and more heterogeneous cohorts. This distinction between individual CCF values and the group-level SI highlights the multi-layered nature of synchrony and presence, and suggests an important direction for future research. <xref ref-type="fig" rid="F9">Figures 9</xref>&#x2013;<xref ref-type="fig" rid="F11">11</xref> illustrate participants one and 6, chosen because their Synchronization Index values were representative of the median and interquartile range, respectively, thus serving to illustrate typical synchronization patterns.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Autocorrelation functions of x-axis rotational data for Subjects one and 6.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g009.tif">
<alt-text content-type="machine-generated">Autocorrelation function (ACF) plots for two representative participants (Subjects 1 and 6) at avatar ratios of 0%, 50%, and 100%. Each graph shows autocorrelation across lag values for glow-stick movement signals sampled at 20 Hz, demonstrating rhythmic periodicity.</alt-text>
</graphic>
</fig>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Cross-correlation functions of x-axis rotational data for participant 1, showing synchronization with the &#x201c;left front&#x201d; and &#x201c;right front&#x201d; avatars. Labels now explicitly indicate whether each avatar was human-controlled or dummy. The interval from 15 to 60&#xa0;s was analyzed to exclude the initial adaptation period and final fade-out, focusing on the segment of stable rhythm and engagement.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g010.tif">
<alt-text content-type="machine-generated">Cross-correlation function (CCF) plots for Subject 1 from 15 to 60 seconds. Columns represent avatar ratio conditions (0%, 50%, 100%). Rows show correlation with avatars positioned left-front, left-back, right-front, and right-back. Lag values range from &#x2212;200 to 200.</alt-text>
</graphic>
</fig>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Cross-correlation functions of x-axis rotational data for participant 6. The interval from 15 to 60&#xa0;s was analyzed to exclude the initial adaptation period and final fade-out, focusing on the segment of stable rhythm and engagement.</p>
</caption>
<graphic xlink:href="frvir-06-1655545-g011.tif">
<alt-text content-type="machine-generated">Cross-correlation function plots for Subject 6 from 15 to 60 seconds. Columns represent avatar ratio conditions (0% only dummy avatars, 50% mixed, 100% only user-controlled avatars). Rows represent left-front, left-back, right-front, and right-back avatar positions. Lag values range from &#x2212;200 to 200.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="s3-9">
<label>3.9</label>
<title>Discussion</title>
<p>This pilot study examined our two formal hypotheses in a VR-based virtual live (VL) event: H1 predicted that higher proportions of synchronized avatars would increase subjective presence, and H2 predicted that the novel Synchronization Index (SI) would correlate strongly with questionnaire scores. The results provide clear support for H1: participants reported significantly higher presence in the 100% synchronization condition compared to the 0% synchronization condition, with the 50% synchronization condition yielding intermediate ratings. In contrast, H2 was not supported under our <italic>a priori</italic> threshold: although SI values showed a moderate positive correlation with presence scores (<inline-formula id="inf19">
<mml:math id="m19">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.42</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.13</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>), this did not meet the predefined criterion (<inline-formula id="inf20">
<mml:math id="m20">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). As a robustness check, we also computed Spearman&#x2019;s <inline-formula id="inf21">
<mml:math id="m21">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, which likewise did not reach significance.</p>
<p>Theoretically, these findings extend social synchrony frameworks&#x2014;where coordinated movements foster emotional contagion (<xref ref-type="bibr" rid="B12">Hatfield et al., 1993</xref>; <xref ref-type="bibr" rid="B29">Sebanz et al., 2006</xref>)&#x2014;into VR contexts, and they corroborate physical performance research on movement parameters shaping emotional transmission (<xref ref-type="bibr" rid="B7">Dahl and Friberg, 2007</xref>). The partial validation of H2 suggests that while our SI captures meaningful aspects of avatar&#x2013;audio coupling, further refinement and validation across diverse rhythmic patterns are needed for robust application.</p>
<p>Study limitations include the small sample size (<inline-formula id="inf22">
<mml:math id="m22">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, university students) and the focus on a single gesture modality (glow-stick movements). These factors constrain the statistical power and generalizability of the findings. The controlled laboratory setting with a small audience may also not reflect the dynamics of real-world VL platforms. Broader demographic and cultural samples, as well as multi-gesture behaviors (e.g., clapping, dancing, cheering), will be necessary for future validation.</p>
<p>Future work should recruit larger and more diverse participant pools, incorporate varied interaction forms such as full-body dance movements, and deploy the system on commercial VR platforms to test ecological validity. In addition, although our present analysis averaged across both dummy and user-controlled avatars, future studies should explicitly compare synchronization patterns with each type. This distinction will help clarify whether participants entrain preferentially to scripted avatars or to human-driven avatars, thereby improving the interpretation of presence outcomes. Technical enhancements, including predictive interpolation and adaptive latency compensation (<xref ref-type="bibr" rid="B11">G&#xfc;l et al., 2020</xref>), should also be explored to strengthen SI reliability.</p>
<p>Practically, our mixed-methods framework offers preliminary, cost-effective design suggestions: while full synchronization maximizes presence, moderate variability may enhance emotional engagement without sacrificing co-presence. Interestingly, contrary to our initial hypothesis, some participants reported higher presence when more dummy avatars were present. One possible explanation is that participants occasionally perceived the dummy avatars&#x2019; 120 BPM rhythm as aligned with the musical beat. This suggests that beat misperception and entrainment to an alternative rhythm may have contributed to the observed effects. These insights lay the groundwork for scalable, adaptive synchronization mechanisms in next-generation VR entertainment systems.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s4">
<label>4</label>
<title>Conclusion</title>
<p>In this exploratory pilot study, we developed a VR-based environment simulating a VL event with varied avatar synchronization ratios. Our findings robustly support H1, confirming that higher synchronization increases subjective presence, and offer initial&#x2014;but inconclusive&#x2014;evidence for H2 regarding the Synchronization Index&#x2019;s correlation with presence scores. Despite sample and modality limitations, these results align with social synchrony theory (<xref ref-type="bibr" rid="B29">Sebanz et al., 2006</xref>) and emotional contagion models (<xref ref-type="bibr" rid="B12">Hatfield et al., 1993</xref>). We tentatively suggest that VR platform developers explore adaptive synchronization mechanisms&#x2014;such as rhythmic feedback loops, latency-aware animation blending, and lightweight predictive interpolation&#x2014;as potentially cost-effective means to enhance audience immersion. While our pilot findings highlight the promise of synchronization-based design, larger-scale validation across diverse user groups and interaction modalities will be required before deriving definitive design standards.</p>
</sec>
<sec id="s5">
<label>5</label>
<title>Future work</title>
<p>Future work should also explore frequency-domain approaches such as cross-spectral analysis to determine whether participants entrain more strongly to the musical beat (140 BPM) or to the dummy avatars&#x2019; rhythm (120 BPM). Quantifying deviations from the intended beat would provide an additional layer of evidence to explain why participants sometimes reported stronger presence in dummy-rich conditions. Building on the insights gained, future research should systematically vary performer behaviors. For example, gesture amplitude and tempo variation could be manipulated to test their impact. Stage design elements, including lighting dynamics and spatial audio cues, should then be adjusted to assess how they moderate audience synchronization and presence. Furthermore, other presence-related factors merit investigation. Sound design (e.g., spatialized music, bass emphasis), avatar personalization, and enhanced social affordances such as cheering or chat interactions may substantially contribute to immersive presence. We plan to examine these in combination with synchronization effects in future work.</p>
<p>In parallel, it will be important to examine how network latency and animation interpolation strategies influence perceived synchronicity, employing objective measures like cross-correlation analysis of motion data (<xref ref-type="bibr" rid="B3">Chatfield, 2003</xref>) alongside subjective questionnaires. Rather than treating synchronization as a binary variable, nuanced exploration of timing offsets and variability may reveal optimal parameter ranges for maximizing presence. Additionally, extending the participant pool to more diverse demographics and increasing the number of simultaneous avatars will be crucial for generalizing the findings beyond the pilot context. This work will not only clarify the relationship between avatar interaction patterns and co-presence but also inform concrete design guidelines for next-generation digital live event platforms that balance technical feasibility with user-centered experience design. Future research should expand participant diversity, incorporate full-body motion tracking, and evaluate synchronization algorithms in real-world network environments to validate and extend these preliminary results.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="ethics-statement" id="s7">
<title>Ethics statement</title>
<p>Ethical approval was not required for the studies involving humans because under Ritsumeikan University&#x2019;s policies for minimal-risk VR experiments at the time, formal IRB review was not required. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>YG: Conceptualization, Methodology, Data curation, Formal Analysis, Investigation, Project administration, Visualization, Writing &#x2013; original draft. SS: Supervision, Validation, Writing &#x2013; review and editing. KM: Supervision, Validation, Writing &#x2013; review and editing. YO: Supervision, Validation, Writing &#x2013; review and editing.</p>
</sec>
<ack>
<title>Acknowledgements</title>
<p>I would like to express my sincere gratitude to everyone who contributed to this research, especially those who participated in the experiments.</p>
</ack>
<sec sec-type="COI-statement" id="s10">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s11">
<title>Generative AI statement</title>
<p>The author(s) declare that Generative AI was used in the creation of this manuscript. During the preparation of this work, the authors used ChatGPT (<xref ref-type="bibr" rid="B22">OpenAI, 2025</xref>) for language refinement. After using this tool, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s12">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn fn-type="custom" custom-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1207632/overview">Justine Saint-Aubert</ext-link>, Inria Rennes - Bretagne Atlantique Research Centre, France</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2548152/overview">Timothy John Pattiasina</ext-link>, Institut Informatika Indonesia, Indonesia</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2798939/overview">Yoshiko Arima</ext-link>, Kyoto University of Advanced Science (KUAS), Japan</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Bourgault</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>L.-Y.</given-names>
</name>
<name>
<surname>Jacobs</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kazi</surname>
<given-names>R. H.</given-names>
</name>
</person-group> (<year>2025</year>). &#x201c;<article-title>Narrative motion blocks: combining direct manipulation and natural language interactions for animation creation</article-title>,&#x201d; in <source>
<italic>Proceedings of the 2025 ACM designing Interactive Systems Conference (DIS &#x2019;25)</italic> (ACM)</source>. <pub-id pub-id-type="doi">10.1145/3715336.3735766</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Brockwell</surname>
<given-names>P. J.</given-names>
</name>
<name>
<surname>Davis</surname>
<given-names>R. A.</given-names>
</name>
</person-group> (<year>1991</year>). <source>Time series: theory and methods</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name>. <pub-id pub-id-type="doi">10.1007/978-1-4757-3843-5</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Chatfield</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2003</year>). <source>The analysis of time series: an introduction</source>. <edition>6th edn</edition>. <publisher-loc>Boca Raton, FL</publisher-loc>: <publisher-name>CRC Press</publisher-name>.</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<collab>Cluster</collab> (<year>2023</year>). <article-title>Cluster event&#x2014;virtual live</article-title>. <source>Clust. Creat. Blog</source> <volume>2023</volume>.</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cornejo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Javiera</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Himmbler</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Temporal interpersonal synchrony and movement coordination</article-title>. <source>Front. Psychol.</source> <volume>9</volume>, <fpage>1546</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2018.01546</pub-id>
<pub-id pub-id-type="pmid">30210391</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cornejo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Cuadros</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Carr&#xe9;</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hurtado</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Olivares</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Dynamics of interpersonal coordination: a cross-correlation approach</article-title>. <source>Front. Psychol.</source> <volume>14</volume>, <fpage>1264504</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2023.1264504</pub-id>
<pub-id pub-id-type="pmid">38292530</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dahl</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Friberg</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Visual perception of expressiveness in musicians&#x2019; body movements</article-title>. <source>Music Percept.</source> <volume>24</volume>, <fpage>433</fpage>&#x2013;<lpage>454</lpage>. <pub-id pub-id-type="doi">10.1525/mp.2007.24.5.433</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Davis</surname>
<given-names>M. H.</given-names>
</name>
</person-group> (<year>1983</year>). <article-title>Measuring individual differences in empathy: evidence for a multidimensional approach</article-title>. <source>J. Personality Soc. Psychol.</source> <volume>44</volume>, <fpage>113</fpage>&#x2013;<lpage>126</lpage>. <pub-id pub-id-type="doi">10.1037/0022-3514.44.1.113</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Faul</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Erdfelder</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Lang</surname>
<given-names>A.-G.</given-names>
</name>
<name>
<surname>Buchner</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Gpower 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences</article-title>. <source>Behav. Res. Methods</source> <volume>39</volume>, <fpage>175</fpage>&#x2013;<lpage>191</lpage>. <pub-id pub-id-type="doi">10.3758/BF03193146</pub-id>
<pub-id pub-id-type="pmid">17695343</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gonzalez-Franco</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lanier</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Model of illusions and virtual reality</article-title>. <source>Front. Psychol.</source> <volume>8</volume>, <fpage>1125</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.01125</pub-id>
<pub-id pub-id-type="pmid">28713323</pub-id>
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>G&#xfc;l</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bosse</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Podborski</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Schierl</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Hellge</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Kalman filter&#x2013;based head motion prediction for cloud-based mixed reality</article-title>. <source>Presence Teleoperators and Virtual Environ.</source> <volume>29</volume>, <fpage>30</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1145/3394171.3413699</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hatfield</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Cacioppo</surname>
<given-names>J. T.</given-names>
</name>
<name>
<surname>Rapson</surname>
<given-names>R. L.</given-names>
</name>
</person-group> (<year>1993</year>). <article-title>Emotional contagion</article-title>. <source>Curr. Dir. Psychol. Sci.</source> <volume>2</volume>, <fpage>96</fpage>&#x2013;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1111/1467-8721.ep10770953</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hatmaker</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Fortnite&#x2019;s ariana grande concert offers a taste of music in the metaverse</article-title>. <source>TechCrunch</source>.</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Juravsky</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fidler</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>X. B.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Superpadl: scaling language-directed physics-based control with progressive supervised distillation</article-title>,&#x201d; in <source>
<italic>Proceedings of the ACM SIGGRAPH 2024 Conference</italic> (ACM)</source>. <pub-id pub-id-type="doi">10.1145/3641519.3657492</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kimmel</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Heuten</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Landwehr</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Kinetic connections: exploring the impact of realistic body movements on social presence in collaborative virtual reality</article-title>. <source>Proc. ACM Human-Computer Interact.</source> <volume>8</volume>, <fpage>1</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1145/3686910</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Konvalinka</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Xygalatas</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Bulbulia</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Schj&#xf8;dt</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Jegind&#xf8;</surname>
<given-names>E.-M.</given-names>
</name>
<name>
<surname>Wallot</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>Synchronized arousal between performers and related spectators in a fire-walking ritual</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>108</volume>, <fpage>8514</fpage>&#x2013;<lpage>8519</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1016955108</pub-id>
<pub-id pub-id-type="pmid">21536887</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kruskal</surname>
<given-names>W. H.</given-names>
</name>
<name>
<surname>Wallis</surname>
<given-names>W. A.</given-names>
</name>
</person-group> (<year>1952</year>). <article-title>Use of ranks in one-criterion variance analysis</article-title>. <source>J. Am. Stat. Assoc.</source> <volume>47</volume>, <fpage>583</fpage>&#x2013;<lpage>621</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.1952.10483441</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liaw</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ooi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Rusli</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Lau</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Tam</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Chua</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Nurse-physician communication team training in virtual reality <italic>versus</italic> live simulations: randomized controlled trial on team communication and teamwork attitudes</article-title>. <source>J. Med. Internet Res.</source> <volume>22</volume>, <fpage>e17279</fpage>. <pub-id pub-id-type="doi">10.2196/17279</pub-id>
<pub-id pub-id-type="pmid">32267235</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lombard</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ditton</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>At the heart of it all: the concept of presence</article-title>. <source>J. Computer-Mediated Commun.</source> <volume>3</volume>, <fpage>0</fpage>. <pub-id pub-id-type="doi">10.1111/j.1083-6101.1997.tb00072.x</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Miranda</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2000</year>). &#x201c;<article-title>An evaluation of the paired comparisons method for software sizing</article-title>,&#x201d; in <source>Proceedings of the 22nd international conference on software engineering</source> (<publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>597</fpage>&#x2013;<lpage>604</lpage>. <pub-id pub-id-type="doi">10.1145/337180.337477</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>N&#xe9;da</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ravasz</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Brechet</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Vicsek</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Barab&#xe1;si</surname>
<given-names>A.-L.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>The sound of many hands clapping</article-title>. <source>Nature</source> <volume>403</volume>, <fpage>849</fpage>&#x2013;<lpage>850</lpage>. <pub-id pub-id-type="doi">10.1038/35002660</pub-id>
<pub-id pub-id-type="pmid">10706271</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="web">
<collab>OpenAI</collab> (<year>2025</year>). <article-title>ChatGPT (june 2025 version)</article-title>. <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://openai.com">https://openai.com</ext-link> (Accessed July 24, 2025)</comment>.</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<collab>Photon Unity Networking</collab> (<year>2022</year>). <article-title>Pun-2 introduction</article-title>. <source>Phot. Engine Doc</source>.</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>PutturVenkatraj</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Meijer</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Perusquia-Hernandez</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Huisman</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>El Ali</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>Shareyourreality: investigating haptic feedback and agency in virtual avatar co-embodiment</article-title>,&#x201d; in <source>
<italic>CHI &#x2019;24: proceedings of the CHI conference on human factors in computing systems</italic> (ACM)</source>. <pub-id pub-id-type="doi">10.1145/3613904.3642425</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rogers</surname>
<given-names>S. L.</given-names>
</name>
<name>
<surname>Broadbent</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Fraser</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Speelman</surname>
<given-names>C. P.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Realistic motion avatars are the future for social interaction in virtual reality</article-title>. <source>Front. Virtual Real.</source> <volume>2</volume>, <fpage>750729</fpage>. <pub-id pub-id-type="doi">10.3389/frvir.2021.750729</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scheff&#xe9;</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>1952</year>). <article-title>An analysis of variance for paired comparisons</article-title>. <source>J. Am. Stat. Assoc.</source> <volume>47</volume>, <fpage>381</fpage>&#x2013;<lpage>400</lpage>. <pub-id pub-id-type="doi">10.2307/2281310</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schubert</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Friedmann</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Regenbrecht</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2001a</year>). <article-title>The experience of presence: factor analytic insights</article-title>. <source>Presence teleoper. Virtual Environ.</source> <volume>10</volume>, <fpage>266</fpage>&#x2013;<lpage>281</lpage>. <pub-id pub-id-type="doi">10.1162/10574601300343603</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schubert</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Friedmann</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Regenbrecht</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2001b</year>). <article-title>The experience of presence: factor analytic insights</article-title>. <source>Presence Teleoperators Virtual Environ.</source> <volume>10</volume>, <fpage>266</fpage>&#x2013;<lpage>281</lpage>. <pub-id pub-id-type="doi">10.1162/105474601300343603</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sebanz</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Bekkering</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Prinz</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Joint action: bodies and minds moving together</article-title>. <source>Trends Cognitive Sci.</source> <volume>10</volume>, <fpage>70</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2005.12.009</pub-id>
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal">
<collab>SEGA Co., Ltd.</collab> (<year>2024</year>). <article-title>Project sekai virtual live</article-title>. <source>Proj. SEKAI Off. Syst. Guide</source>.</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Shlizerman</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Dery</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Schoen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kemelmacher-Shlizerman</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Audio to body dynamics</article-title>,&#x201d; in <source>2018 IEEE/CVF conference on computer vision and pattern recognition</source>, <fpage>7574</fpage>&#x2013;<lpage>7583</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2018.00790</pub-id>
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Slater</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wilbur</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>A framework for immersive virtual environments (five): speculations on the role of presence in virtual environments</article-title>. <source>Presence Teleoperators and Virtual Environ.</source> <volume>6</volume>, <fpage>603</fpage>&#x2013;<lpage>616</lpage>. <pub-id pub-id-type="doi">10.1162/pres.1997.6.6.603</pub-id>
</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Swarbrick</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Seibt</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Grinspun</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Vuoskoski</surname>
<given-names>J. K.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Corona concerts: the effect of virtual concert characteristics on social connection and kama muta</article-title>. <source>Front. Psychol.</source> <volume>12</volume>, <fpage>648448</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2021.648448</pub-id>
<pub-id pub-id-type="pmid">34239478</pub-id>
</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tarr</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Launay</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dunbar</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Synchrony and exertion during dance independently raise pain threshold and encourage social bonding</article-title>. <source>Biol. Lett.</source> <volume>8</volume>, <fpage>106</fpage>&#x2013;<lpage>109</lpage>. <pub-id pub-id-type="doi">10.1098/rsbl.2015.0767</pub-id>
</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ullal</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Watkins</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Sarkar</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>A dynamically weighted multi-objective optimization approach to positional interactions in remote&#x2013;local augmented/mixed reality</article-title>,&#x201d; in <source>2021 IEEE international conference on artificial intelligence and virtual reality (AIVR)</source> (<publisher-name>IEEE</publisher-name>), <fpage>29</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1109/AIVR52153.2021.00014</pub-id>
</mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ullal</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Watkins</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sarkar</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2022</year>). &#x201c;<article-title>A multi-objective optimization framework for redirecting pointing gestures in remote&#x2013;local mixed/augmented reality</article-title>,&#x201d; in <source>Proceedings of the 2022 ACM Symposium on Spatial User Interaction (SUI &#x2019;22)</source> (<publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>ACM</publisher-name>). <pub-id pub-id-type="doi">10.1145/3565970.3567681</pub-id>
</mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="web">
<collab>UNITY-CHAN!OFFICIAL-WEBSITE</collab> (<year>2014</year>). <article-title>UNITY-CHAN LIVE STAGE! -Candy rock Star-</article-title>. <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://unity-chan.com/download/releaseNote.php?id=CandyRockStar">https://unity-chan.com/download/releaseNote.php?id&#x3d;CandyRockStar</ext-link>.</comment>
</mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Usoh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Catena</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Arman</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Slater</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Using presence questionnaires in reality</article-title>. <source>Presence Teleoperators Virtual Environ.</source> <volume>9</volume>, <fpage>497</fpage>&#x2013;<lpage>503</lpage>. <pub-id pub-id-type="doi">10.1162/105474600566989</pub-id>
</mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Vivanco</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Jayasumana</surname>
<given-names>A. P.</given-names>
</name>
</person-group> (<year>2007</year>). &#x201c;<article-title>A measurement-based modeling approach for network-induced packet delay</article-title>,&#x201d; in <source>32nd IEEE conference on local computer networks</source>, <fpage>175</fpage>&#x2013;<lpage>182</lpage>. <pub-id pub-id-type="doi">10.1109/LCN.2007.136</pub-id>
</mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="book">
<collab>VRChat</collab> (<year>2022</year>). &#x201c;<article-title>Vrchat blog</article-title>,&#x201d; in <source>
<italic>VRChat Blog</italic> accessed</source>.</mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal">
<collab>VRChat</collab> (<year>2023</year>). <article-title>Full-body tracking and user engagement in virtual live events</article-title>. <source>VRChat Tech. Rep</source>.</mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waltemate</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Gall</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Roth</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Botsch</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Latoschik</surname>
<given-names>M. E.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>The impact of avatar personalization and immersion on virtual body ownership, presence, and emotional response</article-title>. <source>IEEE Trans. Vis. Comput. Graph.</source> <volume>24</volume>, <fpage>1643</fpage>&#x2013;<lpage>1652</lpage>. <pub-id pub-id-type="doi">10.1109/TVCG.2018.2794629</pub-id>
<pub-id pub-id-type="pmid">29543180</pub-id>
</mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Witmer</surname>
<given-names>B. G.</given-names>
</name>
<name>
<surname>Singer</surname>
<given-names>M. J.</given-names>
</name>
</person-group> (<year>1998</year>). <article-title>Measuring presence in virtual environments: a presence questionnaire</article-title>. <source>Presence Teleoperators and Virtual Environ.</source> <volume>7</volume>, <fpage>225</fpage>&#x2013;<lpage>240</lpage>. <pub-id pub-id-type="doi">10.1162/105474698565686</pub-id>
</mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="journal">
<collab>XR Association</collab> (<year>2023</year>). <article-title>State of the industry report 2023</article-title>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>