<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Hum. Neurosci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Human Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Hum. Neurosci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1662-5161</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnhum.2026.1744449</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Spontaneous production rates in music and speech: Effector systems or domain specificity?</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Coleman</surname>
<given-names>Nicole C.</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3378659"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Palmer</surname>
<given-names>Caroline</given-names>
</name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/30596"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Pfordrsher</surname>
<given-names>Peter Q.</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/30337"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Psychology, University at Buffalo</institution>, <city>Buffalo</city>, <state>NY</state>, <country country="us">United States</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Psychology, McGill University</institution>, <city>Montreal</city>, <state>QC</state>, <country country="ca">Canada</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Peter Q. Pfordrsher, <email xlink:href="mailto:pqp@buffalo.edu">pqp@buffalo.edu</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-26">
<day>26</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>20</volume>
<elocation-id>1744449</elocation-id>
<history>
<date date-type="received">
<day>11</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>13</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>29</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Coleman, Palmer and Pfordrsher.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Coleman, Palmer and Pfordrsher</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-26">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Individuals perform many tasks at an optimal rate that is consistent within but not between individuals, evidenced by the spontaneous rate at which one performs a task in the absence of external rate cues. We tested three hypotheses concerning how spontaneous production rates (SPRs) are generated and associated across language and music tasks: Biomechanical constraints associated with effector systems (vocalized/fingered: H1), reliance on auditory feedback (presence/absence: H2), and domain-specific constraints (speech/music: H3). We tested these hypotheses by having participants produce music and speech sequences, sequences that used vocalized or fingered effectors, plus a Silent finger-tapping condition to test the influence of auditory feedback on spontaneous production rates. SPRs were significantly correlated across all tasks that involved production of auditory feedback, regardless of effector or domain. However, Silent Tapping rates were not significantly correlated with any SPRs that produced auditory feedback. Together, these findings suggest that the generation of auditory feedback plays a critical role in the spontaneous rate at which participants engage in rhythmic motor actions, more so than the biomechanical constraints of effector systems.</p>
</abstract>
<kwd-group>
<kwd>domain specificity</kwd>
<kwd>music</kwd>
<kwd>singing</kwd>
<kwd>speech</kwd>
<kwd>spontaneous production rates</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. Financial support was provided by NSERC grant 298173 to the second author and NSF grant BCS-2419763 to the third author.</funding-statement>
</funding-group>
<counts>
<fig-count count="5"/>
<table-count count="4"/>
<equation-count count="0"/>
<ref-count count="38"/>
<page-count count="11"/>
<word-count count="8554"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Speech and Language</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<title>Introduction</title>
<p>People engage in a variety of rhythmic motor behaviors, including walking and running, which are defined in part by a regular period or rate at which the rhythm recurs. A wide range of movement rates are available, but the range is always limited and in general there is an optimal rate that an individual gravitates toward for any given behavior. These optimal rates are estimated using the <italic>spontaneous production rate</italic> (SPR), which is the rate at which one produces sequences of sounds in the absence of external cues. We address rhythmic behaviors that are used to communicate via audition: music and speech. These behaviors are associated with different cognitive <italic>domains</italic>, referring to a class of information that may be associated with a distinct cognitive network, possibly associated with distinct modules (<xref ref-type="bibr" rid="ref11">Fodor, 1983</xref>). Music and language may constitute separate domains according to some claims (e.g., <xref ref-type="bibr" rid="ref25">Patel, 2008</xref>; <xref ref-type="bibr" rid="ref27">Peretz and Coltheart, 2003</xref>). With respect to the rate at which people produce music and speech, two questions emerge: First, is there any association between an individual&#x2019;s optimal rate for speech and their optimal rate for music production, and&#x2014;if so&#x2014;the second question involves the basis on which that association is made.</p>
<p>The second question is of particular interest given that the production of music and language can vary both in terms of the actions used to generate <italic>auditory feedback</italic>, and in terms of the <italic>effector system</italic> used to produce the actions. We define effector systems as a collection of muscles and joints, such as finger or tongue movements, that function in a coordinated way in order to execute the goals of an action, forming synergies (<xref ref-type="bibr" rid="ref3">Bernstein, 1967</xref>). The present study addresses the role of these factors in the Spontaneous Production rates (SPRs) at which people produce rhythmic sequences like speech or music.</p>
<p>We focus primarily on how individual differences are associated across music and speech production tasks, and not on the average rates at which each task is produced. Recent evidence suggests that music and language from around the world differ with respect to the rate at which they are produced, with considerably faster rates for the rhythms formed by successive spoken syllables than by the onsets of musical notes in a melody (<xref ref-type="bibr" rid="ref7">Ding et al., 2017</xref>; <xref ref-type="bibr" rid="ref24">Ozaki et al., 2024</xref>). The fact that these differences exist does not necessarily indicate that different timing systems contribute to an individual&#x2019;s optimal rates of production across domains. It is well known that different gait patterns within species, including horses (<xref ref-type="bibr" rid="ref16">Hoyt and Taylor, 1981</xref>) and humans (<xref ref-type="bibr" rid="ref20">Kung et al., 2018</xref>) are associated with different optimal rates, for example with faster optimal rates for running and slower optimal rates for walking. These differences can be attributed to changes in the attractor landscape within a single system when a control pattern (here, rate) varies, and do not require any attribution to the use of different timing systems for the different coordination patterns (e.g., <xref ref-type="bibr" rid="ref18">Kelso, 1995</xref>). In a similar way, it is possible that SPRs for speech and music production may vary in overall rate while being based on the same timing system. Given this flexibility, we focus on whether an individual who chooses a relatively &#x201C;fast&#x201D; rate for one task (compared to other individuals) also chooses a &#x201C;fast&#x201D; rate on a different task, even if the two rates differ considerably from each other. If so, we may say that this individual timekeeper engages a &#x201C;fast&#x201D; timekeeper during both motor tasks, while varying rates flexibly from instance to instance in response to other contextual factors.</p>
<p>Results from previous studies offer mixed evidence concerning whether SPRs are correlated across music and language production. Whereas <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> found no significant correlation between SPRs for speech production and the production of melodies on a keyboard, <xref ref-type="bibr" rid="ref6">Coleman et al. (2024)</xref> found significant associations between production of speech and production of sung melodies. These results are consistent with the view that SPRs are based on the effector system used, given that the correlated SPRs in Coleman et al. were found when speech and music were produced via the same system. Similarly, <xref ref-type="bibr" rid="ref9010">Engler et al. (2024)</xref> found that tapping rates were more strongly correlated when participants tapped under conditions of varying auditory feedback (taps producing a melody, a single tone, or silent) in contrast to correlations between tapping rates with either hand clapping or walking. Constraints on optimal rates due to effector system may reflect the biomechanics associated with effector systems that function synergistically when one engages in rhythmic movements. The integrated movements of limbs that vary in mass, flexibility and muscular dexterity significantly influence the optimal rate of rhythmic movements. These effects have been modeled in past research by varying masses of pendula that are moved rhythmically (e.g., <xref ref-type="bibr" rid="ref35">Turvey et al., 1986</xref>). These studies did not, however, manipulate the presence or absence of auditory feedback, an important dimension of timing in music and speech.</p>
<p>Other results suggest that auditory feedback plays a role in governing optimal rates. <xref ref-type="bibr" rid="ref34">Tranchant et al. (2022)</xref> found that the spontaneous rate of each participant&#x2019;s productions of musical sequences using finger movements were not significantly correlated with their spontaneous motor tempo (fingered tapping at consistent rates in the absence of auditory feedback; <xref ref-type="bibr" rid="ref9">Drake et al., 2000</xref>; <xref ref-type="bibr" rid="ref23">McAuley et al., 2006</xref>), whereas other studies have found significant correlations for similar conditions (<xref ref-type="bibr" rid="ref9010">Engler et al., 2024</xref>; <xref ref-type="bibr" rid="ref29">Pfordresher et al., 2021</xref>). <xref ref-type="bibr" rid="ref9010">Engler et al. (2024)</xref> found stronger correlations between tasks with shared auditory information&#x2014;tapping to produce different melodies, or rhythmic actions that generated no sound (silent tapping versus hand clapping)&#x2014;than across tasks that generate different kinds of auditory patterns. These results suggest that auditory information associated with produced actions&#x2014;here the simple presence or absence of auditory feedback&#x2014;may determine optimal production rates.</p>
<p>Other theoretical approaches associate spontaneous production rates with information based on principles of information transmission, possibly leading to differences in production rates across music and speech (<xref ref-type="bibr" rid="ref1">Assaneo and Poeppel, 2018</xref>; <xref ref-type="bibr" rid="ref8">Doelling et al., 2019</xref>). Thus, differences in SPRs may result for actions that lead to music or speech, independent of effector system.</p>
<p>The present study tests three hypotheses concerning the basis on which timekeepers may or may not be distinguished. The first (H1) is that spontaneous rates are constrained by the biomechanics of the effector system used to generate a rhythm. Some aforementioned results support H1: Whereas <xref ref-type="bibr" rid="ref6">Coleman et al. (2024)</xref> found significantly correlated SPRs across speaking and singing, which share the same (vocal) effector system, <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> found no significant association (and a near zero correlation) for SPRs from speaking and playing melodies on a piano. The current study tests whether this dissociation is replicated within a single study.</p>
<p>The second hypothesis (H2) is that SPRs are influenced by the presence of auditory feedback generated from actions. Recent studies provide mixed evidence. Whereas <xref ref-type="bibr" rid="ref34">Tranchant et al. (2022)</xref> found no association between silent tapping and a condition in which tapping led to a sequence of melody pitches, <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> found a significant but modest correlation between silent tapping and SPRs for piano production. As shown in <xref ref-type="table" rid="tab1">Table 1</xref>, H2 predicts strong correlations between spontaneous rates among sung melodies, speech, and tapped melodies, for which feedback is present, but weak or non-significant correlations between any of these conditions and silent tapping.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Research hypotheses and conditions.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Condition</th>
<th align="left" valign="top">Effector (H1)</th>
<th align="left" valign="top">Auditory feedback (H2)</th>
<th align="left" valign="top">Domain (H3)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Speech</td>
<td align="left" valign="top">Vocal</td>
<td align="left" valign="top">Present</td>
<td align="left" valign="top">Language</td>
</tr>
<tr>
<td align="left" valign="top">Sung Melody</td>
<td align="left" valign="top">Vocal</td>
<td align="left" valign="top">Present</td>
<td align="left" valign="top">Music</td>
</tr>
<tr>
<td align="left" valign="top">Tapped Melody</td>
<td align="left" valign="top">Fingered</td>
<td align="left" valign="top">Present</td>
<td align="left" valign="top">Music</td>
</tr>
<tr>
<td align="left" valign="top">Silent Tapping</td>
<td align="left" valign="top">Fingered</td>
<td align="left" valign="top">Absent</td>
<td align="left" valign="top">n/a</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Columns after experimental conditions organized by factors that are related to the three hypotheses being tested. &#x2018;Effector&#x2019; refers to the effector system used for a task, relevant to hypothesis H1. &#x2018;Auditory Feedback&#x2019; refers to the presence or absence of auditory feedback, relevant to hypothesis H2. Domain refers to whether the auditory information conforms to the domain of Language or Music. Tests of hypotheses are based on whether correlations of production rates are based on one or more of these factors.</p>
</table-wrap-foot>
</table-wrap>
<p>The third hypothesis (H3) is that SPRs are based on the acoustic patterning of domain-specific information that is produced. This hypothesis predicts that SPRs will be correlated within domain (i.e., across production of music via singing and on a keyboard), but not across domains (i.e., music versus speech production). Although the results of <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> are consistent with H3, the results of <xref ref-type="bibr" rid="ref6">Coleman et al. (2024)</xref> are not. However, neither of these studies included comparisons within a domain across different effector systems. A recent study that examined synchronization that crossed effector system (hand claps versus articulation) with the domain associated with the auditory target (syllable versus tone), was consistent with the idea that timekeeping is independent of effector and domain for accurate synchronizers (<xref ref-type="bibr" rid="ref21">Mares et al., 2023</xref>). The current study uses a comparable design and can address whether this generality extends to SPRs.</p>
<p>We report an experiment that tested these hypotheses by having participants generate SPRs for short memorized sequences. Participants included those with and without musical training. Tasks that could be produced easily (without formal training) were presented to each participant in four conditions: Speaking short phrases, singing familiar melodies on the syllable &#x2018;da&#x2019; (Sung Melody), producing familiar melodies by tapping on a keyboard (Tapped Melody), or Silent Tapping. Participants were given different melodies for the Sung Melody and Tapped Melody conditions to avoid carryover effects across conditions. <xref ref-type="table" rid="tab1">Table 1</xref> illustrates the relationship between these experimental conditions and the three hypotheses. Whereas Speech and Sung Melody conditions both involve the vocal effector system, Tapped Melody and Silent Tapping conditions both involve the hand and finger effector system. H1 predicts shared SPRs within the verbal and within the tapping conditions, but not across these systems. The Silent Tapping condition was used to test H2: If the presence of auditory feedback determines SPRs, then production rates in this condition should not correlate with the Tapped Melody condition. H3 predicts a significant association between Sung and Tapped Melody conditions, but no correlation between these conditions and the Speech condition.</p>
</sec>
<sec sec-type="methods" id="sec2">
<title>Method</title>
<sec id="sec3">
<title>Participants</title>
<p>Eighty-one students were recruited through the University at Buffalo&#x2019;s online participant recruiting software (SONA) and given partial course credit for Introduction to Psychology in exchange for their time. Participants gave informed consent to participate in the study and indicated they had not been diagnosed with any hearing or speaking disabilities. Twenty-three participants were excluded because of unfamiliarity with the melodic stimuli (<italic>n</italic>&#x202F;=&#x202F;20) or errors/dysfluencies during production of speech stimuli (<italic>n</italic>&#x202F;=&#x202F;3).</p>
<p>Fifty-eight participants (mean age&#x202F;=&#x202F;19.07, range&#x202F;=&#x202F;18&#x2013;29) were included in the final analysis. These participants had on average 0.66&#x202F;years (range&#x202F;=&#x202F;0&#x2013;16&#x202F;years) of formal training in singing (e.g., private lessons) and 2.91&#x202F;years (range&#x202F;=&#x202F;0&#x2013;17) in organized singing (e.g., choir). Participants also averaged 2.85&#x202F;years (range&#x202F;=&#x202F;0&#x2013;15&#x202F;years) of formal training on a primary instrument. Six participants (10% of the sample) were non-L1 English speakers. Two participants listed Chinese (including Cantonese, Mandarin, and other Chinese languages) and two participants listed Korean as their native language. Two participants listed &#x2018;other&#x2019; without specifying their native language. Of the individuals who were L1 English speakers, 28 (48% of the sample) were multilingual.</p>
</sec>
<sec id="sec4">
<title>Apparatus</title>
<p>Participants were seated in front of a computer monitor (Dell ST2310F) and a M-Audio Keystation 49e keyboard. The computer monitor was connected to and mirrored the screen of a windows computer (Dell OptiPlex 7490 AIO) outside of the sound booth. The windows computer ran the Speech and Sung Melody tasks through MATLAB (version R2021b) using Psychtoolbox-3 (<xref ref-type="bibr" rid="ref9011">Kleiner et al., 2007</xref>). A three-key mini keypad in the booth allowed the participants to progress through the Speech and Sung Melody tasks at their own pace. Participants spoke or sang into a microphone (Shure SM58). Audio for these tasks was recorded through an audio interface (Focusrite Scarlett 2i2).</p>
<p>Participants tapped on the M-Audio Keystation 49e, using a single key (see Procedure) for the melody tapping and Silent Tapping trials. Each tap generated the next tone in the melody, allowing musically trained and untrained participants to control the melody rate. Melodies were played through a Roland RD-700 SX midi tone generator over Sennheiser HD 280 pro (64 ohm) headphones. FTAP, a software package used to collect MIDI data and control MIDI output with millisecond temporal precision (<xref ref-type="bibr" rid="ref10">Finney, 2001</xref>), was used to output the experimental orders and acquire tapping responses. FTAP was run on a Linux operating system on a Dell Optiplex 790 computer outside of the booth.</p>
<p>A microphone (AKG C1000S) outside of the booth allowed experimenters to talk to the participant and a pair of headphones (Sennheiser HD 280 pro) allowed the experimenter to hear the participant. Qualtrics (Provo, UT, USA) was used to administer the questionnaire at the end of the experiment.</p>
</sec>
<sec id="sec5">
<title>Materials and design</title>
<p>There were four spontaneous rate tasks: Speaking, singing (Sung Melody), production of melodies by tapping (Tapped Melody), and tapping without auditory feedback (Silent Tapping). These conditions varied with respect to effector system used for production (vocal: Speech and Sung Melody, fingers: Tapped Melody and Silent Tapping), presence of auditory feedback (absent for Silent Tapping, present for the rest), and&#x2014;for conditions with auditory feedback&#x2014;domain represented by auditory information (speech versus music). Participants produced two stimulus items for each task generating auditory feedback and completed three successive trials for each stimulus item. Tasks were grouped together in blocks to avoid confusion for participants.</p>
<p>Each participant was randomly assigned to one of two experimental orders for the succession of tasks. Both orders followed two constraints. First, musical production tasks (Sung and Tapped Melody tasks) did not follow each other in direct succession to minimize potential within-domain carryover effects. Second, each order began with a Silent Tapping trial to provide an estimate for this task that was free of any possible carryover of auditory information from previous tasks. Order 1 was Silent Tapping, Sung Melody Trials, Silent Tapping, Speech Trials, and Tapped Melody Trials. Order 2 was Silent Tapping, Tapped Melody Trials, Speech Trials, Silent Tapping, and Sung Melody Trials. <xref ref-type="fig" rid="fig1">Figure 1</xref> illustrates the flow of trials for an example participant in Order 1.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Illustration of trial orders. This is an example depicting Order 1 for a given participant. There are 6 trials total for each Speech or Sung Melody condition and each condition began with practice trials. For the music tasks (Sung and Tapped Melody), four melodies were assigned. Two of the melodies were produced in the Sung Melody trials and the remaining melodies were produced in the Tapped Melody trials, as indicated by the red outlines for each stimulus.</p>
</caption>
<graphic xlink:href="fnhum-20-1744449-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart illustrates three experimental trials: Sung Melody, Speech, and Tapped Melody. Each trial includes instructions, warm-ups, practice, and randomized song or phrase tasks, followed by a questionnaire at the end.</alt-text>
</graphic>
</fig>
<p>During the Speech task, two short phrases served as stimulus items. Participants were asked to memorize all two phrases at the start of the task. The phrases were: <italic>Good things come to those who wait</italic>, and <italic>She who pays the piper calls the tune</italic>. These phrases were chosen because they are simple, have a binary meter, are simple in structure, and the stressed syllables are easy to distinguish when marking onsets during analysis (see <italic>Data Analysis</italic>).</p>
<p>For the Sung Melody task, participants were asked to produce two familiar melodies from either list A or list B from memory: <italic>Mary Had a Little Lamb</italic> and <italic>Twinkle Twinkle Little Star (List A)</italic> or <italic>Frosty the Snowman</italic> and <italic>Are You Sleeping? (List B)</italic>, to ensure that the same melodies were not used for both sung and Tapped Melody trials for a given participant. List order was counterbalanced with experimental order (Order 1 and Order 2). Participants only had to be familiar with the tune and not the song lyrics; all participants sang the melodies on the syllable &#x201C;da.&#x201D; The songs were chosen because they are highly familiar to most participants and have a binary meter. Additionally, these songs are not associated with a canonical pre-recorded version and thus will not be associated with a specific tempo (as would be the case for prerecorded popular music) to avoid effects related to tempo recall (e.g., <xref ref-type="bibr" rid="ref14">Hine et al., 2025</xref>). If they were unfamiliar with a melody, they were asked to choose an alternate familiar melody from the same list for the remaining three trials.</p>
<p>For the Tapped Melody task, participants heard the melody they had chosen as familiar from list A or B. On each trial, participants tapped a single key on the keyboard and each tap generated the next pitch in the melody, based on MIDI pitch information presented by the FTAP software package (<xref ref-type="bibr" rid="ref10">Finney, 2001</xref>). Successive presses of the key thus allowed the participant to control the melody&#x2019;s rate. Participants practiced tapping the melody in a practice trial until they felt comfortable to move on to the experimental trials.</p>
<p>For the Silent Tapping task, participants were instructed to tap on a single key on the keyboard in front of them at a comfortable and consistent rate until instructed by the experimenter to stop. The audio output from the keyboard was turned off, and no keyboard sounds resulted from their actions. This task occurred before and after the first block of trials and consisted of 39 taps on each trial.</p>
<p>A background questionnaire measured participants&#x2019; ages at which they started learning any non-native languages as well as their musical background, using self-reported years of training in organized singing (including choir, musical ensembles, musical theater, etc.), formal training in singing, and formal training on any primary instrument. We screened for the possibility of congenital amusia, a disorder affecting the perception of pitch and overall recognition of music (<xref ref-type="bibr" rid="ref2">Ayotte et al., 2002</xref>; <xref ref-type="bibr" rid="ref28">Peretz and Vuvan, 2017</xref>), by asking the level of difficulty a participant had in recognizing music when lyrics are absent, and we screened for possible beat-deafness (<xref ref-type="bibr" rid="ref30">Phillips-Silver et al., 2011</xref>) by asking participants how hard it was for them to find the beat of the music.</p>
</sec>
<sec id="sec6">
<title>Procedure</title>
<p>Participants were directed to enter the sound-attenuated booth. Once the participant read and provided informed consent, the experimenter sat down at the desk outside of the sound booth to start the experiment. The participant and experimenter communicated through the microphone and headphone set-up on both sides of the sound booth. The experimenter then guided the participant through the correct order of tasks, reading the instructions for each task and asking if the participant had any questions along the way. In each task, participants were instructed to produce the taps or vocalizations at a consistent and comfortable tempo. None of the participants had difficulty understanding or complying with this general instruction, monitored by an experimenter.</p>
<p>Each order began with the Silent Tapping condition. Participants were instructed to tap at a comfortable and consistent rate using the pointer finger of their dominant hand until the experimenter instructed them to stop. Although no sound was played through the headphones, participants were instructed to keep the headphones on to minimize any sound occurring from any keyboard movement. The participant completed one trial of the Silent Tapping task before moving on to the next task.</p>
<p>Using Order1 as an example, participants next began the Sung Melody trials by producing a warm-up melody, <italic>Happy Birthday</italic>. At the beginning of each Sung Melody trial, song lyrics were displayed on the screen to aid participants&#x2019; memory for the melody until the participant decided to progress to the production task at which point the lyrics were removed. The participant was then instructed to repeat the melody at a comfortable and consistent rate on the syllable &#x201C;da&#x201D; to reduce memory demands, provide a clear consonant sound to mark syllable onsets, and prevent the length of the phonemes from influencing their spontaneous singing rate, (e.g., <xref ref-type="bibr" rid="ref31">Racette and Peretz, 2007</xref>). After completing one warm-up trial, the participant completed three consecutive trials for each of the two chosen melodies, which were presented in a randomized order.</p>
<p>The participant then repeated another trial of the Silent Tapping and was told to tap at a comfortable and consistent rate until instructed to stop by the experimenter.</p>
<p>Next, the participant completed a block of Speech trials. A single <italic>trial</italic> for speech was defined as 4 successive repetitions of a short well-known phrase. Four repetitions of each spoken phrase in the Speech trials approximately matched the duration of the Sung Melody trials. On each trial the participant viewed the text of the target phrase on the computer screen. They were instructed to read and memorize the phrase on the screen and then proceed to the recording screen when they were ready, at which time the phrase disappeared from the screen. They spoke the phrase from memory to avoid the effects of reading on spontaneous speaking rate and to encourage a more natural speaking rate. The participant was instructed to repeat the phrase at a consistent and comfortable rate from memory. If the participant needed more time to see or practice the stimulus, the phrase appeared again on the screen until the participant was ready to record. The block of Speech trials began with a warm-up trial that involved producing from memory <italic>I pledge allegiance to the flag of the United States of America</italic> four times, as shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>. The warm-up trial, similar to the Happy Birthday Sung Melody warmup, helped to ensure that participants understood the instructions and performed the tasks correctly. There were three consecutive trials for each phrase in the Speech task using the target phrases described earlier.</p>
<p>The participant then proceeded to the Tapped Melody task. This block of trials was configured identically to the Sung Melody trials, only using the alternate list of melodies (i.e., if Sung Melody trials used list A, then Tapped Melody trials would use list B). Participants produced the melodies by tapping a single key on the keyboard. Participants were instructed to produce the melody at a comfortable and consistent rate. Warm-up for this task involved tapping a practice melody until the participant tapped the rhythm correctly.</p>
<p>At the end of the experiment, the participant was instructed to sit at the windows computer outside of the sound booth where the questionnaire appeared on the screen. The participant then completed the background questionnaire. When the questionnaire was submitted, the researcher granted course credit to the participant and ended the experiment session. The entire experiment lasted 60&#x202F;min.</p>
</sec>
<sec id="sec7">
<title>Data analysis</title>
<p>Participants&#x2019; SPRs were assessed by marking sung tone onsets and syllable onsets and by calculating the mean of all inter-onset intervals (IOIs) for each trial (one melody repetition for music; four repetitions of one spoken phrase for speech). For speech production, the onset of each syllable was identified using Gentle (Version 0.10.01), a forced aligner program that uses media files and audio transcripts to mark the onset of each phoneme automatically and precisely. Analyses of vocal and fingered melodies omitted the IOI defined by the onset of the final syllable of a phrase and the onset of the first syllable of the next phrase in a trial because participants tended to pause in-between repetitions (cf. <xref ref-type="bibr" rid="ref29">Pfordresher et al., 2021</xref>). The inclusion of these pauses would inaccurately inflate an individual&#x2019;s SPR.</p>
<p>For Sung Melody production, the onset of each &#x201C;da&#x201D; tone was marked manually using Praat (<xref ref-type="bibr" rid="ref9012">Boersma and Weenink, 2022</xref>). The researcher used visual cues from the spectral content and amplitude envelope of the sound in addition to auditory cues to hand-mark the onset of each &#x2018;Da&#x2019;. The denominational unit we used to establish SPR was the quarter note. For Sung Melody and Tapped Melody trials, IOIs that included denominations of half notes or longer were omitted because participants tend to shorten long durations, making the ratio of half notes to quarter notes closer to 1:1.7 rather than the correct notated ratio of 1:2 (<xref ref-type="bibr" rid="ref9014">Bengtsson and Gabrielsson, 1980</xref>; <xref ref-type="bibr" rid="ref9013">Repp et al., 2013</xref>). Likewise, we summed across IOIs for 8th notes that followed in immediate succession, so that the resulting measure mapped onto the quarter note duration. We removed 8th note IOIs that were isolated (e.g., the 8th note that follows the dotted quarter note in <italic>Frosty the Snowman</italic>). For trials on which the participant misinterpreted the rhythm (which occurred on 17% of all Sung Melody trials, but not on any Tapped Melody trials), we created a template of notated durations based on the rhythm produced by the participant, rather than the ideal notated rhythm for the melody. After these adjustments, all IOIs in the remaining trials were used in the Sung Melody analysis as participants did not add additional pauses in between notes or phrases. The timing data through FTAP for the Tapped Melody trials and the Silent Tapping trials provided the information to calculate the IOI for each tap. No IOIs were excluded from the Silent Tapping tasks as the participants did not pause during their productions.</p>
<p>The mean IOI was first calculated for each trial, following removal of outliers. The mean and SE were then computed across the trial means and constituted the primary dependent variable. Outliers were defined separately for each participant and task (Speech or Sung Melody) as being more than three standard deviations away from the mean across all trials in the IOIs used to analyze a single trial. This process led to negligible loss of data (<italic>N</italic>&#x202F;=&#x202F;55 IOIs removed, 67 from Speech trials and 8 from Sung Melody trials, 0.002% of all IOIs). Any dysfluencies were removed in the process of removing outliers.</p>
</sec>
</sec>
<sec sec-type="results" id="sec8">
<title>Results</title>
<p><xref ref-type="fig" rid="fig2">Figure 2</xref> show the distribution of mean IOIs across participants. Bars in each panel represent the mean IOI across trials, and error bars reflect variability across trials for that participant and task. Participants are ordered in each panel according to their rank order for Sung Melody IOIs, and axis scales vary to show the range of IOIs in the task (e.g., note that smaller <italic>Y</italic>-axis range for Sung Melody trials than music trials). These plots verify that individuals varied significantly from each other on each task, thus validating the use of correlation to assess the degree of covariation across tasks.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Mean IOI by participant for each task. Mean (SE) SPR across item using mean IOI for each individual for all 4 production tasks. Participants are ordered on the <italic>x</italic>-axis in all graphs based on the Sung Melody task. <italic>Y</italic>-axis scale varies across graphs to highlight the range of individual differences within each task. Mean and SE values are based on the mean across trials for each task (<italic>n</italic>&#x202F;=&#x202F;6 for Sung Melody, Speech, and Tapped Melody, <italic>n</italic>&#x202F;=&#x202F;2 for Silent Tapping).</p>
</caption>
<graphic xlink:href="fnhum-20-1744449-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Four grouped bar graphs compare mean inter-onset interval (IOI) per trial in seconds across fifty-eight participants for sung melody, speech, tapped melody, and silent tapping tasks. Sung melody shows a steady increase, speech displays greater variability, tapped melody has moderate variation, and silent tapping exhibits the largest overall spread and highest means. Error bars indicate variability for each participant.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="fig3">Figure 3</xref> displays associations between SPRs across tasks. Each point in each plot represents the joint mean SPR for two conditions. Because each condition participates in three different bivariate correlations, we adopted a Bonferroni correction and considered correlations at <italic>p</italic>&#x202F;&#x003C;&#x202F;0.017 to be statistically significant. Pearson correlations are shown in the leftmost column of <xref ref-type="table" rid="tab2">Table 2</xref>. All correlations that compared conditions in which actions led to auditory feedback (SPR tasks), yielded significant correlations. By contrast, all correlations with Silent Tapping (bottom row of <xref ref-type="fig" rid="fig3">Figure 3</xref>) did not yield significant correlations (<italic>p</italic>&#x202F;&#x003E;&#x202F;0.10 for each).</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Bivariate correlations across tasks. All axis units are in seconds per IOI, on average. Axis scales vary by panel based on the range of X and Y values. Panels <bold>(A&#x2013;C)</bold> (top row) show correlations among SPR tasks that include auditory feedback. Panels <bold>(D&#x2013;F)</bold> (bottom row) show correlations of SPR tasks with Silent Tapping.</p>
</caption>
<graphic xlink:href="fnhum-20-1744449-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Six-panel figure with scatterplots showing correlations between melodic or tapping performance and speech or melody outcomes. Panels A, B, and C show significant positive correlations between sung melody or tapped melody and speech or tapped melody, while panels D, E, and F show non-significant relationships involving silent tapping. Each plot includes a regression line and reported r and p values.</alt-text>
</graphic>
</fig>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Pearson <italic>r</italic> and Bayes Factor values for associations across tasks.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Correlation</th>
<th align="center" valign="top">Pearson <italic>r</italic></th>
<th align="center" valign="top">Bayes Factor<sub>10</sub></th>
<th align="center" valign="top">Bayes Factor <sub>01</sub></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Sung Melody/Speech</td>
<td align="char" valign="bottom" char=".">0.47&#x002A;&#x002A;</td>
<td align="center" valign="bottom">279.46+++</td>
<td align="center" valign="bottom">0.004</td>
</tr>
<tr>
<td align="left" valign="middle">Sung Melody/Tapped Melody</td>
<td align="char" valign="bottom" char=".">0.42&#x002A;&#x002A;</td>
<td align="center" valign="bottom">59.938++</td>
<td align="center" valign="bottom">0.017</td>
</tr>
<tr>
<td align="left" valign="middle">Speech/Tapped Melody</td>
<td align="char" valign="bottom" char=".">0.29&#x002A;</td>
<td align="center" valign="bottom">3.558+</td>
<td align="center" valign="bottom">0.281</td>
</tr>
<tr>
<td align="left" valign="middle">Sung Melody/Silent Tapping</td>
<td align="char" valign="bottom" char=".">0.03</td>
<td align="center" valign="bottom">0.137</td>
<td align="center" valign="bottom">7.278&#x2212;&#x2212;</td>
</tr>
<tr>
<td align="left" valign="middle">Tapped Melody/Silent Tapping</td>
<td align="char" valign="bottom" char=".">0.16</td>
<td align="center" valign="bottom">0.123</td>
<td align="center" valign="bottom">8.120&#x2212;&#x2212;</td>
</tr>
<tr>
<td align="left" valign="middle">Speech/Silent Tapping</td>
<td align="char" valign="bottom" char=".">0.04</td>
<td align="center" valign="bottom">0.119</td>
<td align="center" valign="bottom">8.408&#x2212;&#x2212;</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>All tests one-tailed based on anticipated positive association. Bayse Factor<sub>10</sub> assesses support for alternative over null hypothesis, whereas Bayes Factor<sub>01</sub> assesses support for null over alternative hypothesis. For Pearson <italic>r</italic>: &#x002A;&#x002A;<italic>p</italic>&#x202F;&#x003C;&#x202F;0.001, &#x002A;<italic>p</italic>&#x202F;&#x003C;&#x202F;0.017. For Bayes Factor<sub>10</sub>: +++ extreme evidence, ++ very strong evidence, + moderate. For Bayes Factor<sub>01</sub>: &#x2212;&#x2212; very strong evidence.</p>
</table-wrap-foot>
</table-wrap>
<p>We also analyzed the results shown in <xref ref-type="fig" rid="fig3">Figure 3</xref> using the Bayes Factor criterion (<xref ref-type="bibr" rid="ref32">Rouder et al., 2009</xref>). Using the Bayesian correlation function in JASP version 0.17.2.1 (<xref ref-type="bibr" rid="ref17">JASP Team, 2023</xref>), the Bayes Factor criterion (BF) tests the likelihood of the alternative hypothesis (significant correlation, BF<sub>10</sub>) over the null (no correlation) or support for the null over the alternative (BF<sub>01</sub>). Because only positive associations of SPRs across tasks are interpretable, we treated all tests as one-tailed. BF values, shown in the middle and right columns of <xref ref-type="table" rid="tab2">Table 2</xref>, indicated overwhelming support of the alternative hypothesis (i.e., a significant correlation) for correlations of Sung Melody with both Tapped Melody and Speech, and the association between Speech and Tapped Melody SPRs yielded strong support for the alternative hypothesis. By contrast, all associations involving Silent Taping yielded strong support for the null hypothesis. These results were consistent with those found for frequentist (Pearson) correlations, while providing additional insights on whether results support the null hypothesis.</p>
<p>We also analyzed whether this pattern of results differed by language experience or musical background. The current sample included a sizable proportion of multilingual participants and included predominantly nonmusicians. Although the vast majority of multilinguals in our sample were L1 English speakers, it is possible that experience learning the sound patterns of other languages may influence speech SPRs. <xref ref-type="table" rid="tab3">Table 3</xref> shows the across-task correlations for all participants in comparison with multilingual and monolingual participants. The same basic pattern is found in all groups, although the correlation between Tapped Melodies and Speech (more modest in size for the entire group than the other significant correlations) was non-significant for the two smaller subsets (despite being of higher magnitude for multilingual participants than for the whole sample). We take this to be simply a byproduct of sample size. We also analyzed correlations for those participants who were categorized as non-musicians, defined as having fewer than 6&#x202F;years of formal training on any musical instrument or singing. These correlations also yielded the same pattern as found for the entire sample. We did not run correlations on participants defined as musicians given how few participants were categorized as such (<italic>n</italic>&#x202F;=&#x202F;10).</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Pearson <italic>r</italic> values for participant groups.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Correlation</th>
<th align="center" valign="top">Multilingual (<italic>n</italic>&#x202F;=&#x202F;20)</th>
<th align="center" valign="top">Monolingual (<italic>n</italic>&#x202F;=&#x202F;38)</th>
<th align="center" valign="top">Non-musicians (<italic>n</italic>&#x202F;=&#x202F;48)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="bottom">Sung Melody/Speech</td>
<td align="char" valign="bottom" char="."><bold>0.55</bold></td>
<td align="char" valign="bottom" char="."><bold>0.39</bold></td>
<td align="char" valign="bottom" char="."><bold>0.54</bold></td>
</tr>
<tr>
<td align="left" valign="bottom">Sung Melody/Tapped Melody</td>
<td align="char" valign="bottom" char="."><bold>0.49</bold></td>
<td align="char" valign="bottom" char="."><bold>0.36</bold></td>
<td align="char" valign="bottom" char="."><bold>0.45</bold></td>
</tr>
<tr>
<td align="left" valign="bottom">Speech/Tapped Melody</td>
<td align="char" valign="bottom" char=".">0.37</td>
<td align="char" valign="bottom" char=".">0.23</td>
<td align="char" valign="bottom" char="."><bold>0.31</bold></td>
</tr>
<tr>
<td align="left" valign="bottom">Sung Melody/Silent Tapping</td>
<td align="char" valign="bottom" char=".">&#x2212;0.03</td>
<td align="char" valign="bottom" char=".">0.05</td>
<td align="char" valign="bottom" char=".">0.03</td>
</tr>
<tr>
<td align="left" valign="bottom">Tapped Melody/Silent Tapping</td>
<td align="char" valign="bottom" char=".">0.02</td>
<td align="char" valign="bottom" char=".">0.21</td>
<td align="char" valign="bottom" char=".">0.11</td>
</tr>
<tr>
<td align="left" valign="bottom">Speech/Silent Tapping</td>
<td align="char" valign="bottom" char=".">&#x2212;0.03</td>
<td align="char" valign="bottom" char=".">0.06</td>
<td align="char" valign="bottom" char=".">0.03</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold indicates correlations significant at <italic>p</italic> &#x003C; 0.017.</p>
</table-wrap-foot>
</table-wrap>
<p>We analyzed correlations of the Silent Tapping condition with other conditions based on its relative position within the order of tasks. One complication with interpreting this condition as representing the absence of feedback is that in certain situations participants may imagine auditory feedback resulting from their taps (<xref ref-type="bibr" rid="ref4">Brown and Palmer, 2012</xref>, <xref ref-type="bibr" rid="ref5">2013</xref>; <xref ref-type="bibr" rid="ref13">Highben and Palmer, 2004</xref>). This is particularly likely for participants who experience the second silent tapping trial after the Melody Tapping condition, which is similar to Silent Tapping with respect to the effector system. This happens in the second of two orders that a participant could experience (see <italic>Methods</italic>). Thus, we re-analyzed all correlations of Silent Tapping with the other conditions, separately for the first and second trials of Silent Tapping (trial 1 always began the session), and further separating correlations with trial 2 for participants who experienced trial Order 1 (Silent Tapping precedes Melody Tapping) versus those who experienced trial Order 2 (Silent Tapping follows Melody Tapping). Spontaneous tapping rates for the second trial of Silent Tapping did correlate significantly with Melody Tapping for participants who experienced trial Order 2, <italic>r</italic>(27)&#x202F;=&#x202F;0.35, <italic>p</italic>&#x202F;=&#x202F;0.03, suggesting that auditory imagery associated with Melody Tapping may have carried over to the later Silent Tapping trials in these cases. No other correlations of any SPR trials with Silent Tapping were significant (all <italic>p</italic>&#x202F;&#x003E;&#x202F;0.10).</p>
<p>We next addressed whether SPRs are reliable across trials within a given task (<xref ref-type="fig" rid="fig4">Figure 4</xref>). It is possible, for instance, that the lack of correlation for the Silent Tapping conditions in most cases may reflect generally unreliable production rates for this task. We assessed reliability by averaging IOIs across the first trial for each item in a task (for Silent Tapping, this is just the mean produced rate for a single trial) and comparing this with the comparable mean for the final trial (trial 3 for production of Speech, Sung Melody, or Tapped Melody; trial 2 for Silent Tapping). All correlations were strong, positive, and statistically significant, indicating high reliability. High reliability between Silent Tapping trial 1 (before any auditory feedback trials) and trial 2 (in middle of study) indicates that any effects of auditory imagery did not significantly modulate the pattern of individual differences in tapping rate for this task. The plot in <xref ref-type="fig" rid="fig4">Figure 4D</xref> removes one participant from the Tapped Melody condition whose mean IOI for trial 1 (1.98) was highly discrepant from trial 3 (0.35), in contrast to other participants. When this participant was included, the correlation was much smaller but still significant at <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, <italic>r</italic>(56)&#x202F;=&#x202F;0.25, <italic>p</italic>&#x202F;=&#x202F;0.027.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Bivariate correlations across trials and within tasks (<bold>A</bold> = Vocal Melody, <bold>B</bold> = Song, <bold>C</bold> = Tapped Melody, <bold>D</bold> = Silent Tapping). All axis units are in seconds per IOI, on average. Axis scales vary by panel based on the range of X and Y values.</p>
</caption>
<graphic xlink:href="fnhum-20-1744449-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Four-panel scientific figure showing scatter plots with regression lines for correlations between trials in four conditions: A, Vocal Melody (r equals 0.94), B, Song (r equals 0.97), C, Tapped Melody (r equals 0.93), and D, Silent Tapping (r equals 0.81), each with p less than 0.001.</alt-text>
</graphic>
</fig>
<p>Although our main focus is on correlations across tasks with respect to individual differences in SPR, we also analyzed mean SPR across tasks to assess whether the present data replicate results found in other studies. The mean IOI across all trials and items for each participant and task were analyzed via a within-participants single-factor ANOVA, which yielded a strong and significant effect of task, <italic>F</italic>(3, 171)&#x202F;=&#x202F;113.43, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001, <inline-formula>
<mml:math id="M1">
<mml:msubsup>
<mml:mi>&#x03B7;</mml:mi>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:math>
</inline-formula> = 0.67. Means and distributions are shown in <xref ref-type="fig" rid="fig5">Figure 5</xref>. Post-hoc tests (Bonferroni corrected) revealed significant contrasts across all pairs of conditions. Mean IOIs thus proceeded in an ordered manner with Speech (<italic>M</italic>&#x202F;=&#x202F;0.223, SE&#x202F;=&#x202F;0.005)&#x202F;&#x003C;&#x202F;Sung Melody (<italic>M</italic>&#x202F;=&#x202F;0.339, SE&#x202F;=&#x202F;0.008)&#x202F;&#x003C;&#x202F;Tapped Melody (<italic>M</italic>&#x202F;=&#x202F;0.387, SE&#x202F;=&#x202F;0.009)&#x202F;&#x003C;&#x202F;Silent Tapping (<italic>M</italic>&#x202F;=&#x202F;0.763, SE&#x202F;=&#x202F;0.042). This reflects what has been reported elsewhere (<xref ref-type="bibr" rid="ref6">Coleman et al., 2024</xref>; <xref ref-type="bibr" rid="ref29">Pfordresher et al., 2021</xref>). It is worth noting that none of the conditions approximates the 0.6&#x202F;sec IOI rate first reported by <xref ref-type="bibr" rid="ref12">Fraisse (1982)</xref>, and often repeated in the literature (e.g., <xref ref-type="bibr" rid="ref15">Honing, 2013</xref>), as shown by the light gray line. The correlations across tasks that we report thus were found in the context of mean differences in production rate that are typical for the tasks we used.</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Mean production rate across tasks. Each dot indicates each participant&#x2019;s mean IOI (in seconds) across all trials and items for a given task. The bold lines show means and upper and lower lines represent 95% confidence intervals for each task. <italic>Y</italic> axis units are in seconds per IOI.</p>
</caption>
<graphic xlink:href="fnhum-20-1744449-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Strip chart comparing mean inter-onset interval (IOI) across four conditions: Speech, Sung Melody, Tapped Melody, and Silent Tapping. Speech, Sung Melody, and Tapped Melody show similar low mean IOI values with small variance, while Silent Tapping shows higher mean IOI values and greater variability. Dark black horizontal lines indicate means for each group, and light black horizontal lines show confidence intervals. Individual data points for each condition are shown as grey dots with random jitter added to x-axis values to prevent occlusion.</alt-text>
</graphic>
</fig>
</sec>
<sec sec-type="discussion" id="sec9">
<title>Discussion</title>
<p>We used a set of tasks to assess whether individual differences in spontaneous production rates (SPRs) reflect shared timekeeping mechanisms based on the effector system used to generate rhythmic actions and/or the auditory feedback generated by those actions. We tested three hypotheses concerning the basis for the selection of an optimal production rate, observed via SPRs: basis on shared effector systems (H1), basis on the presence or absence of auditory feedback (H2), or basis on domain-specific differences in the structuring of information within a domain (H3). <xref ref-type="table" rid="tab4">Table 4</xref> provides a qualitative summary of how well the pattern of correlations across tasks fit these hypotheses, using both positive (significant correlation) and negative (nonsignificant correlation) evidence. Results clearly support the auditory feedback hypothesis over the other hypotheses. As shown in <xref ref-type="table" rid="tab4">Table 4</xref>, all significant and non-significant correlations are consistent with this hypothesis, and with the results of <xref ref-type="bibr" rid="ref34">Tranchant et al. (2022)</xref>. By contrast, the pattern of results is only consistent with other hypotheses at rates suggesting chance performance.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Qualitative assessment of support for hypotheses.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Correlation</th>
<th align="left" valign="top" rowspan="2">Significant</th>
<th align="center" valign="top" colspan="3">Consistent with?</th>
</tr>
<tr>
<th align="center" valign="top">Effector (H1)</th>
<th align="center" valign="top">Feedback (H2)</th>
<th align="center" valign="top">Domain (H3)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Sung Melody/Speech</td>
<td align="left" valign="top">Y</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2713;</td>
</tr>
<tr>
<td align="left" valign="top">Sung/Tapped Melody</td>
<td align="left" valign="top">Y</td>
<td align="center" valign="top">&#x2717;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2713;</td>
</tr>
<tr>
<td align="left" valign="top">Speech/Tapped Melody</td>
<td align="left" valign="top">Y</td>
<td align="center" valign="top">&#x2717;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2717;</td>
</tr>
<tr>
<td align="left" valign="top">Sung Melody/Silent Tap</td>
<td align="left" valign="top">N</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">(n/a)</td>
</tr>
<tr>
<td align="left" valign="top">Tapped Melody /Silent Tap</td>
<td align="left" valign="top">N</td>
<td align="center" valign="top">&#x2717;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">(n/a)</td>
</tr>
<tr>
<td align="left" valign="top">Speech/Silent Tap</td>
<td align="left" valign="top">N</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">&#x2713;</td>
<td align="center" valign="top">(n/a)</td>
</tr>
<tr>
<td align="left" valign="top"><italic>N Support</italic></td>
<td/>
<td align="center" valign="top">3/6</td>
<td align="center" valign="top">6/6</td>
<td align="center" valign="top">1/3</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Each row reflects a correlation illustrated in <xref ref-type="fig" rid="fig2">Figure 2</xref>. Significance coded based on <italic>p</italic>-values from those correlations. Check marks indicate results that are consistent with a hypothesis, whereas <italic>x</italic>-marks indicate lack of support. Conditions include repetition of spoken phrases (Speech), singing familiar melodies on &#x2018;da&#x2019; (Sung Melody), producing a melody on a keyboard by tapping a single key repeatedly (Tapped Melody), or tapping on a keyboard with no auditory feedback (Silent Tapping). H1: effector specificity, H2: auditory feedback presence, H3: domain specificity.</p>
</table-wrap-foot>
</table-wrap>
<p>The primary finding of this experiment (H2) was that spontaneous rates vary across individuals in similar ways for tasks in which production generates auditory feedback (such as the SPR, spontaneous production rate), but not for tasks in which production does not generate auditory feedback (such as the SMT, spontaneous motor tempo). These findings are consistent with <xref ref-type="bibr" rid="ref34">Tranchant et al. (2022)</xref> who speculated that differences in SPR and SMT values within individuals may be related to the fact that the auditory feedback present in SPR tasks provides error correction for timing. Another possibility is that SPR tasks are used to generate auditory patterns that may convey meaningful information to a perceiver. It is possible that the engagement of motor systems in tasks that contain a goal to create perceptual patterns may be less strongly associated with biomechanical constraints than tasks that do not contain perceptual goals.</p>
<p>The findings reported here differ from two other studies that we know of that reported positive correlations between SMT and SPR tasks. <xref ref-type="bibr" rid="ref9010">Engler et al. (2024)</xref> reported positive correlations between conditions similar to our Tapped Melody condition and silent tapping condition, and <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> reported a significant correlation between spontaneous rates of production of melodies on a piano and silent tapping. We suggest that these differences reflect sensitivity of SMT tasks to carryover effects from condition orders. Similar to what we reported in the present <italic>Results</italic> section, <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> found that the significant correlations across conditions varied depending on whether SMT was measured before or after the piano production task, with a significant correlation found for SPR and SMT only when SMT was measured after SPR. Although <xref ref-type="bibr" rid="ref9010">Engler et al. (2024)</xref> did not report such an analysis, they note that their SMT condition never occurred at the start of the session, but instead always occurred at a position intermingled with several other conditions (including SPR). Thus, we propose that other positive correlations between SMT and SPR may reflect a carryover of auditory imagery and/or motor imagery commonly seen in musicians (<xref ref-type="bibr" rid="ref13">Highben and Palmer, 2004</xref>). An important implication of these results is that the use of SMT as a generalizable measure of internal timekeeping must be handled with care, given its potential vulnerability to order effects.</p>
<p>The current findings are also consistent with a timekeeping mechanism that is domain- and effector-general (contrary to H1 and H3), consistent with cognitive clock-counter models of timekeeping (e.g., <xref ref-type="bibr" rid="ref19">Killeen and Weiss, 1987</xref>; <xref ref-type="bibr" rid="ref36">Wing and Kristofferson, 1973</xref>). The correlation reported here between SPRs arising in Sung Melody and Speech production is consistent with the results of <xref ref-type="bibr" rid="ref6">Coleman et al. (2024)</xref> and suggests domain-general timekeeping when sequences are produced with the same (vocal) effector system. Correlations of these tasks with the Tapped Melody condition suggest timekeeping that generalizes across effectors both within domain (Tapped with Sung Melody) and even across domains (Tapped Melody with Speech). The significant (although modest compared to others) correlation between Tapped Melody and Speech production was not consistent with the null (near zero) correlation between piano and speech production reported in <xref ref-type="bibr" rid="ref29">Pfordresher et al. (2021)</xref> and deserves consideration. Some possible explanations for these different results include the use of familiar tunes that were associated with lyrics (used in the current study but not the previous study), the complexity of the fingered motor task (low in the current study&#x2014;single-finger tapping&#x2014;relative to the previous study), and degree of musical training (untrained participants in the current study, pianists in the previous study).</p>
<p>In sum, the present findings suggest that optimal rates may be defined by <italic>the incorporation of sensorimotor plans to form action-sound contingencies</italic>, based on the integration of motor actions with auditory feedback (cf. <xref ref-type="bibr" rid="ref34">Tranchant et al., 2022</xref>). This conclusion is supported both by positive evidence (correlations across SPR tasks) as well as by Bayesian statistics that suggest support for the null hypothesis with respect to correlations across SPR and SMT tasks. The fact that music can be produced by different effector systems (including vocal and arm movements) makes this domain uniquely suitable for decoupling auditory from effector-based information.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec10">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors without undue reservation.</p>
</sec>
<sec sec-type="ethics-statement" id="sec11">
<title>Ethics statement</title>
<p>The studies involving humans were approved by University at Buffalo Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their verbal consent to participate in the study.</p>
</sec>
<sec sec-type="author-contributions" id="sec12">
<title>Author contributions</title>
<p>NC: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. CP: Conceptualization, Formal analysis, Methodology, Supervision, Visualization, Writing &#x2013; review &#x0026; editing. PP: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec sec-type="COI-statement" id="sec13">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec14">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec15">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Assaneo</surname><given-names>M. F.</given-names></name> <name><surname>Poeppel</surname><given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>The coupling between auditory and motor cortices is rate-restricted: evidence for an intrinsic speech-motor rhythm</article-title>. <source>Sci. Adv.</source> <volume>4</volume>:<fpage>eaao3842</fpage>. doi: <pub-id pub-id-type="doi">10.1126/sciadv.aao3842</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ayotte</surname><given-names>J.</given-names></name> <name><surname>Peretz</surname><given-names>I.</given-names></name> <name><surname>Hyde</surname><given-names>K.</given-names></name></person-group> (<year>2002</year>). <article-title>Congenital amusia: a group study of adults afflicted with a music-specific disorder</article-title>. <source>Brain</source> <volume>125</volume>, <fpage>238</fpage>&#x2013;<lpage>251</lpage>. doi: <pub-id pub-id-type="doi">10.1093/brain/awf028</pub-id>, <pub-id pub-id-type="pmid">11844725</pub-id></mixed-citation></ref>
<ref id="ref9014"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bengtsson</surname><given-names>I.</given-names></name> <name><surname>Gabrielsson</surname><given-names>A.</given-names></name></person-group> (<year>1980</year>). <article-title>Methods for analyzing performance of musical rhythm</article-title>. <source>Scandinavian Journal of Psychology</source> <volume>21</volume>, <fpage>257</fpage>&#x2013;<lpage>268</lpage>. doi: <pub-id pub-id-type="doi">10.1111/j.1467-9450.1980.tb00369.x</pub-id></mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Bernstein</surname><given-names>N.</given-names></name></person-group> (<year>1967</year>). <source>The coordination and regulation of movements</source>. <publisher-loc>Oxford, UK</publisher-loc>: <publisher-name>Pergammon Press</publisher-name>.</mixed-citation></ref>
<ref id="ref9012"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Boersma</surname><given-names>P.</given-names></name> <name><surname>Weenink</surname><given-names>D.</given-names></name></person-group> (<year>2022</year>). <article-title>Praat: doing phonetics by computer [Computer program]</article-title>. Available online at: <ext-link xlink:href="https://www.fon.hum.uva.nl/paul/praat.html" ext-link-type="uri">https://www.fon.hum.uva.nl/paul/praat.html</ext-link></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brown</surname><given-names>R. M.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name></person-group> (<year>2012</year>). <article-title>Auditory-motor learning influences auditory memory for music</article-title>. <source>Mem. Cogn.</source> <volume>40</volume>, <fpage>567</fpage>&#x2013;<lpage>578</lpage>. doi: <pub-id pub-id-type="doi">10.3758/s13421-011-0177-x</pub-id>, <pub-id pub-id-type="pmid">22271265</pub-id></mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brown</surname><given-names>R. M.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name></person-group> (<year>2013</year>). <article-title>Auditory and motor imagery modulate learning in music performance</article-title>. <source>Front. Hum. Neurosci.</source> <volume>7</volume>:<fpage>320</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnhum.2013.00320</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Coleman</surname><given-names>N. C.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name> <name><surname>Pfordresher</surname><given-names>P. Q.</given-names></name></person-group> (<year>2024</year>). <article-title>Spontaneous production rates in song and speech</article-title>. <source>Music Sci.</source> <volume>7</volume>:<fpage>20592043241279056</fpage>. doi: <pub-id pub-id-type="doi">10.1177/20592043241279056</pub-id></mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname><given-names>N.</given-names></name> <name><surname>Patel</surname><given-names>A. D.</given-names></name> <name><surname>Chen</surname><given-names>L.</given-names></name> <name><surname>Butler</surname><given-names>H.</given-names></name> <name><surname>Luo</surname><given-names>C.</given-names></name> <name><surname>Poeppel</surname><given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>Temporal modulations in speech and music</article-title>. <source>Neurosci. Biobehav. Rev.</source> <volume>81</volume>, <fpage>181</fpage>&#x2013;<lpage>187</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neubiorev.2017.02.011</pub-id>, <pub-id pub-id-type="pmid">28212857</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Doelling</surname><given-names>K. B.</given-names></name> <name><surname>Assaneo</surname><given-names>M. F.</given-names></name> <name><surname>Bevilacqua</surname><given-names>D.</given-names></name> <name><surname>Pesaran</surname><given-names>B.</given-names></name> <name><surname>Poeppel</surname><given-names>D.</given-names></name></person-group> (<year>2019</year>). <article-title>An oscillator model better predicts cortical entrainment to music</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>116</volume>, <fpage>10113</fpage>&#x2013;<lpage>10121</lpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.1816414116</pub-id></mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Drake</surname><given-names>C.</given-names></name> <name><surname>Jones</surname><given-names>M. R.</given-names></name> <name><surname>Baruch</surname><given-names>C.</given-names></name></person-group> (<year>2000</year>). <article-title>The development of rhythmic attending in auditory sequences: attunement, referent period, focal attending</article-title>. <source>Cognition</source> <volume>77</volume>, <fpage>251</fpage>&#x2013;<lpage>288</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0010-0277(00)00106-2</pub-id>, <pub-id pub-id-type="pmid">11018511</pub-id></mixed-citation></ref>
<ref id="ref9010"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Engler</surname><given-names>B. H.</given-names></name> <name><surname>Zamm</surname><given-names>A.</given-names></name> <name><surname>M&#x00F8;ller</surname><given-names>C.</given-names></name></person-group> (<year>2024</year>). <article-title>Spontaneous rates exhibit high intra-individual stability across movements involving different biomechanical systems and cognitive demands</article-title>. <source>Sci Rep.</source> <volume>14</volume>:<fpage>14876</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-65788-6</pub-id></mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Finney</surname><given-names>S. A.</given-names></name></person-group> (<year>2001</year>). <article-title>FTAP: a Linux-based program for tapping and music experiments</article-title>. <source>Behav. Res. Methods Instrum. Comput.</source> <volume>33</volume>, <fpage>65</fpage>&#x2013;<lpage>72</lpage>. doi: <pub-id pub-id-type="doi">10.3758/bf03195348</pub-id>, <pub-id pub-id-type="pmid">11296721</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Fodor</surname><given-names>J.</given-names></name></person-group> (<year>1983</year>). <source>Modularity of mind</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fraisse</surname><given-names>P.</given-names></name></person-group> (<year>1982</year>). <article-title>Rhythm and tempo</article-title>. <source>Psychol. Music</source> <volume>1</volume>, <fpage>149</fpage>&#x2013;<lpage>180</lpage>.</mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Highben</surname><given-names>Z.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name></person-group> (<year>2004</year>). <article-title>Effects of auditory and motor mental practice in memorized piano performance</article-title>. <source>Bull. Counc. Res. Music. Educ.</source> <volume>159</volume>, <fpage>58</fpage>&#x2013;<lpage>65</lpage>.</mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hine</surname><given-names>K.</given-names></name> <name><surname>Wakana</surname><given-names>Y.</given-names></name> <name><surname>Nakauchi</surname><given-names>S.</given-names></name></person-group> (<year>2025</year>). <article-title>How we remember music tempo: the role of spontaneous motor tempo in recall and preference</article-title>. <source>Front. Psychol.</source> <volume>16</volume>:<fpage>1631625</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpsyg.2025.1631625</pub-id></mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Honing</surname><given-names>H.</given-names></name></person-group> (<year>2013</year>). <article-title>Structure and interpretation of rhythm in music</article-title>. <source>Psychol. Music</source> <volume>3</volume>, <fpage>369</fpage>&#x2013;<lpage>404</lpage>. doi: <pub-id pub-id-type="doi">10.1016/B978-0-12-381460-9.00009-2</pub-id></mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hoyt</surname><given-names>D.</given-names></name> <name><surname>Taylor</surname><given-names>C.</given-names></name></person-group> (<year>1981</year>). <article-title>Gait and the energetics of locomotion in horses</article-title>. <source>Nature</source> <volume>292</volume>, <fpage>239</fpage>&#x2013;<lpage>240</lpage>.</mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="book"><person-group person-group-type="author"><collab id="coll1">JASP Team</collab></person-group> (<year>2023</year>). <source>JASP version 0.17.2.1 [computer software]</source>. <publisher-loc>Amsterdam, ND</publisher-loc>: <publisher-name>University of Amsterdam</publisher-name>.</mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Kelso</surname><given-names>J. A. S.</given-names></name></person-group> (<year>1995</year>). <source>Dynamical patterns: the self-organization of brain and behavior</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Killeen</surname><given-names>P. R.</given-names></name> <name><surname>Weiss</surname><given-names>N. A.</given-names></name></person-group> (<year>1987</year>). <article-title>Optimal timing and the weber function</article-title>. <source>Psychol. Rev.</source> <volume>94</volume>, <fpage>274</fpage>&#x2013;<lpage>295</lpage>.</mixed-citation></ref>
<ref id="ref9011"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kleiner</surname><given-names>M.</given-names></name> <name><surname>Brainard</surname><given-names>D.</given-names></name> <name><surname>Pelli</surname><given-names>D.</given-names></name></person-group> (<year>2007</year>). <article-title>What&#x2019;s new in Psychtoolbox-3?</article-title> <source>Perception, ECVP Abstract Supplement.</source> <volume>36</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>.</mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kung</surname><given-names>S. M.</given-names></name> <name><surname>Fink</surname><given-names>P. W.</given-names></name> <name><surname>Legg</surname><given-names>S. J.</given-names></name> <name><surname>Ali</surname><given-names>A.</given-names></name> <name><surname>Shultz</surname><given-names>S. P.</given-names></name></person-group> (<year>2018</year>). <article-title>What factors determine the preferred gait transition speed in humans? A review of the triggering mechanisms</article-title>. <source>Hum. Mov. Sci.</source> <volume>57</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.humov.2017.10.023</pub-id>, <pub-id pub-id-type="pmid">29121506</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mares</surname><given-names>C.</given-names></name> <name><surname>Echavarr&#x00ED;a Solana</surname><given-names>R.</given-names></name> <name><surname>Assaneo</surname><given-names>M. F.</given-names></name></person-group> (<year>2023</year>). <article-title>Auditory-motor synchronization varies among individuals and is critically shaped by acoustic features</article-title>. <source>Commun. Biol.</source> <volume>6</volume>:<fpage>658</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s42003-023-04976-y</pub-id>, <pub-id pub-id-type="pmid">37344562</pub-id></mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McAuley</surname><given-names>J. D.</given-names></name> <name><surname>Jones</surname><given-names>M. R.</given-names></name> <name><surname>Holub</surname><given-names>S.</given-names></name> <name><surname>Johnston</surname><given-names>H. M.</given-names></name> <name><surname>Miller</surname><given-names>N. S.</given-names></name></person-group> (<year>2006</year>). <article-title>The time of our lives: lifespan development of timing and event tracking</article-title>. <source>J. Exp. Psychol. Gen.</source> <volume>135</volume>, <fpage>348</fpage>&#x2013;<lpage>367</lpage>. doi: <pub-id pub-id-type="doi">10.1037/0096-3445.135.3.348</pub-id>, <pub-id pub-id-type="pmid">16846269</pub-id></mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ozaki</surname><given-names>Y.</given-names></name> <name><surname>Tierney</surname><given-names>A.</given-names></name> <name><surname>Pfordresher</surname><given-names>P. Q.</given-names></name> <name><surname>McBride</surname><given-names>J. M.</given-names></name> <name><surname>Benetos</surname><given-names>E.</given-names></name> <name><surname>Proutskouva</surname><given-names>P.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Globally, songs and instrumental melodies are slower, higher, and use more stable pitches than speech: a registered report. Science</article-title>. <source>Advances</source> <volume>10</volume>:<fpage>eadm9797</fpage>. doi: <pub-id pub-id-type="doi">10.1126/sciadv.adm9797</pub-id></mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Patel</surname><given-names>A. D.</given-names></name></person-group> (<year>2008</year>). <source>Music, language and the brain</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peretz</surname><given-names>I.</given-names></name> <name><surname>Coltheart</surname><given-names>M.</given-names></name></person-group> (<year>2003</year>). <article-title>Modularity of music processing</article-title>. <source>Nat. Neurosci.</source> <volume>6</volume>, <fpage>688</fpage>&#x2013;<lpage>691</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nn1083</pub-id>, <pub-id pub-id-type="pmid">12830160</pub-id></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peretz</surname><given-names>I.</given-names></name> <name><surname>Vuvan</surname><given-names>D. T.</given-names></name></person-group> (<year>2017</year>). <article-title>Prevalence of congenital amusia</article-title>. <source>Eur. J. Hum. Genet.</source> <volume>25</volume>, <fpage>625</fpage>&#x2013;<lpage>630</lpage>. doi: <pub-id pub-id-type="doi">10.1038/ejhg.2017.15</pub-id>, <pub-id pub-id-type="pmid">28224991</pub-id></mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pfordresher</surname><given-names>P. Q.</given-names></name> <name><surname>Greenspon</surname><given-names>E. B.</given-names></name> <name><surname>Friedman</surname><given-names>A. L.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name></person-group> (<year>2021</year>). <article-title>Spontaneous production rates in music and speech</article-title>. <source>Front. Psychol.</source> <volume>12</volume>:<fpage>611867</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpsyg.2021.611867</pub-id></mixed-citation></ref>
<ref id="ref30"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Phillips-Silver</surname><given-names>J.</given-names></name> <name><surname>Toiviainen</surname><given-names>P.</given-names></name> <name><surname>Gosselin</surname><given-names>N.</given-names></name> <name><surname>Piche</surname><given-names>O.</given-names></name> <name><surname>Nozaradan</surname><given-names>S.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name> <etal/></person-group>. (<year>2011</year>). <article-title>Born to dance but beat deaf: a new form of congenital amusia</article-title>. <source>Neuropsychologia</source> <volume>49</volume>, <fpage>961</fpage>&#x2013;<lpage>969</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuropsychologia.2011.02.002</pub-id>, <pub-id pub-id-type="pmid">21316375</pub-id></mixed-citation></ref>
<ref id="ref31"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Racette</surname><given-names>A.</given-names></name> <name><surname>Peretz</surname><given-names>I.</given-names></name></person-group> (<year>2007</year>). <article-title>Learning lyrics: to sing or not to sing?</article-title> <source>Mem. Cogn.</source> <volume>35</volume>, <fpage>242</fpage>&#x2013;<lpage>253</lpage>. doi: <pub-id pub-id-type="doi">10.3758/bf03193445</pub-id>, <pub-id pub-id-type="pmid">17645165</pub-id></mixed-citation></ref>
<ref id="ref9013"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Repp</surname><given-names>B. H.</given-names></name> <name><surname>London</surname><given-names>J.</given-names></name> <name><surname>Keller</surname><given-names>P. E.</given-names></name></person-group> (<year>2013</year>). <article-title>Systematic Distortions in Musicians&#x2019; Reproduction of Cyclic Three-Interval Rhythms</article-title>. <source>Music Perception</source> <volume>30</volume>, <fpage>291</fpage>&#x2013;<lpage>305</lpage>. doi: <pub-id pub-id-type="doi">10.1525/mp.2012.30.3.291</pub-id></mixed-citation></ref>
<ref id="ref32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rouder</surname><given-names>J. N.</given-names></name> <name><surname>Speckman</surname><given-names>P. L.</given-names></name> <name><surname>Sun</surname><given-names>D.</given-names></name> <name><surname>Morey</surname><given-names>R. D.</given-names></name> <name><surname>Iverson</surname><given-names>G.</given-names></name></person-group> (<year>2009</year>). <article-title>Bayesian t tests for accepting and rejecting the null hypothesis</article-title>. <source>Psychon. Bull. Rev.</source> <volume>16</volume>, <fpage>225</fpage>&#x2013;<lpage>237</lpage>. doi: <pub-id pub-id-type="doi">10.3758/PBR.16.2.225</pub-id>, <pub-id pub-id-type="pmid">19293088</pub-id></mixed-citation></ref>
<ref id="ref34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tranchant</surname><given-names>P.</given-names></name> <name><surname>Scholler</surname><given-names>E.</given-names></name> <name><surname>Palmer</surname><given-names>C.</given-names></name></person-group> (<year>2022</year>). <article-title>Endogenous rhythms influence musicians&#x2019; and non-musicians&#x2019; interpersonal synchrony</article-title>. <source>Sci. Rep.</source> <volume>12</volume>:<fpage>12973</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-022-16686-2</pub-id></mixed-citation></ref>
<ref id="ref35"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Turvey</surname><given-names>M. T.</given-names></name> <name><surname>Rosenblum</surname><given-names>L. D.</given-names></name> <name><surname>Schmidt</surname><given-names>R. C.</given-names></name> <name><surname>Kugler</surname><given-names>P. N.</given-names></name></person-group> (<year>1986</year>). <article-title>Fluctuations and phase symmetry in coordinated rhythmic movements</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>12</volume>, <fpage>564</fpage>&#x2013;<lpage>583</lpage>, <pub-id pub-id-type="pmid">2946808</pub-id></mixed-citation></ref>
<ref id="ref36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wing</surname><given-names>A. M.</given-names></name> <name><surname>Kristofferson</surname><given-names>A. B.</given-names></name></person-group> (<year>1973</year>). <article-title>Response delays and the timing of discrete motor responses</article-title>. <source>Percept. Psychophys.</source> <volume>14</volume>, <fpage>5</fpage>&#x2013;<lpage>12</lpage>.</mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1304912/overview">John Houde</ext-link>, University of California, San Francisco, United States</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/737799/overview">Kyoko Hine</ext-link>, Toyohashi University of Technology, Japan</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/928203/overview">Lu Luo</ext-link>, Beijing Sport University, China</p>
</fn>
</fn-group>
</back>
</article>