<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2026.1751421</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Neural tracking of continuous speech reveals enhanced late responses to degraded speech</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Na</surname>
<given-names>Youngmin</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1939311"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Quan</surname>
<given-names>Luong Do Anh</given-names>
</name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1779526"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Joo</surname>
<given-names>Hyosung</given-names>
</name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1938476"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Choi</surname>
<given-names>Inyong</given-names>
</name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/974163"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Woo</surname>
<given-names>Jihwan</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1527734"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Biomedical Engineering, University of Ulsan</institution>, <city>Ulsan</city>, <country country="kr">Republic of Korea</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Electrical, Electronic and Computer Engineering, University of Ulsan</institution>, <city>Ulsan</city>, <country country="kr">Republic of Korea</country></aff>
<aff id="aff3"><label>3</label><institution>Department of Communication Sciences and Disorders, University of Iowa</institution>, <city>Iowa City</city>, <state>IA</state>, <country country="us">United States</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Jihwan Woo, <email xlink:href="mailto:jhwoo@ulsan.ac.kr">jhwoo@ulsan.ac.kr</email></corresp>
<fn fn-type="equal" id="fn0001">
<label>&#x2020;</label>
<p>These authors have contributed equally to this work</p>
</fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-06">
<day>06</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>20</volume>
<elocation-id>1751421</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>17</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>21</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Na, Quan, Joo, Choi and Woo.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Na, Quan, Joo, Choi and Woo</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-06">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Comprehending degraded speech demands greater cognitive effort. While previous studies have identified the neural pathways involved in processing degraded speech signals, the temporal dynamics of these neural networks remain unclear.</p>
</sec>
<sec>
<title>Methods</title>
<p>This study investigated the time course of neural responses to clean and degraded (i.e., noise-vocoded) speech signals using temporal response functions (TRFs).</p>
</sec>
<sec>
<title>Results</title>
<p>Our findings reveal that early TRF components (N1TRF and P2TRF) exhibited greater amplitude and temporal precision for clean speech. In contrast, degraded speech elicited additional cortical responses with a longer delay, designated as P400TRF. Subsequent source localization analyses showed that the P400TRF component originates from language processing areas within the temporal and frontal lobes.</p>
</sec>
<sec>
<title>Discussion</title>
<p>These findings highlight the role of delayed neural mechanisms in maintaining speech comprehension when intelligibility is compromised, offering novel insights that broaden our understanding of auditory cortical processing under challenging listening conditions.</p>
</sec>
</abstract>
<kwd-group>
<kwd>compensatory neural processing</kwd>
<kwd>degraded speech</kwd>
<kwd>neural tracking</kwd>
<kwd>speech perception</kwd>
<kwd>temporal response functions</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the National Research Foundation of Korea (NRF) grants from the Korean government (NRF-RS-2024-00461617, RS-2024-00338148). This work was supported by U.S. Department of Defense (DoD) Hearing Restoration and Rehabilitation Program grant awarded to I.C. (HT9425-23-1-0912), National Institute on Deafness and Other Communication Disorders (NIDCD) P50 (DC000242) awarded to I.C., T.D.G., B.M., and Bruce J. Gantz, and National Science Foundation Grant No. 2306331 awarded to Chipara, Adhikari, Wu, and Choi.</funding-statement>
</funding-group>
<counts>
<fig-count count="4"/>
<table-count count="3"/>
<equation-count count="0"/>
<ref-count count="109"/>
<page-count count="13"/>
<word-count count="10995"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Auditory Cognitive Neuroscience</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>When speech signals become noisy, distorted, or challenging to comprehend, the human auditory system compensates by combining high-level cognitive processes and low-level auditory mechanisms. The brain engages a complex network of neural systems to process degraded speech, integrating sensory, cognitive, and linguistic brain regions in a coordinated manner (<xref ref-type="bibr" rid="ref72">Obleser and Kotz, 2010</xref>; <xref ref-type="bibr" rid="ref49">Karunathilake et al., 2023b</xref>). Additionally, studies have highlighted that phase-locked responses in the auditory cortex are enhanced during speech comprehension, underscoring the role of neural synchronization in processing complex auditory information (<xref ref-type="bibr" rid="ref76">Peelle et al., 2013</xref>; <xref ref-type="bibr" rid="ref42">Hauswald et al., 2022</xref>; <xref ref-type="bibr" rid="ref54">K&#x00F6;sem et al., 2023</xref>).</p>
<p>In the context of degraded speech, auditory cortical areas, such as the primary auditory cortex (A1) and secondary auditory regions, are primarily responsible for processing the degraded acoustic features. These regions encode basic elements like pitch, duration, and spectral properties, even when noise or reverberation is present. However, when speech intelligibility is compromised&#x2014;due to background noise or low-quality signals&#x2014;the brain recruits higher-level cortical regions, particularly in the prefrontal cortex and parietal areas, which involve attentional control, working memory, and top-down modulation. These regions enable the brain to fill perceptual gaps, stabilize auditory representations, and improve the listener&#x2019;s ability to extract meaningful speech information (<xref ref-type="bibr" rid="ref17">Davis and Johnsrude, 2007</xref>; <xref ref-type="bibr" rid="ref72">Obleser and Kotz, 2010</xref>; <xref ref-type="bibr" rid="ref49">Karunathilake et al., 2023b</xref>). Moreover, robust cortical entrainment to the speech envelope, which relies on the spectro-temporal fine structure, plays a critical role in maintaining speech intelligibility in challenging environments (<xref ref-type="bibr" rid="ref21">Ding et al., 2014</xref>; <xref ref-type="bibr" rid="ref42">Hauswald et al., 2022</xref>).</p>
<p>The neural mechanisms extend beyond simple auditory processing by integrating semantic and linguistic expectations to enhance speech intelligibility. For instance, the left inferior frontal gyrus (IFG), associated with syntactic and semantic processing, has been shown to become more active when listeners are attempting to make sense of speech under challenging conditions (<xref ref-type="bibr" rid="ref74">Obleser et al., 2007</xref>; <xref ref-type="bibr" rid="ref77">Peelle et al., 2010</xref>; <xref ref-type="bibr" rid="ref79">Peelle and Wingfield, 2022</xref>). Additionally, the cortical entrainment response is significantly influenced by prior knowledge and speech intelligibility, indicating that listeners&#x2019; expectations and contextual understanding are vital for effective neural tracking of speech (<xref ref-type="bibr" rid="ref3">Baltzell et al., 2017</xref>; <xref ref-type="bibr" rid="ref49">Karunathilake et al., 2023b</xref>).</p>
<p>While temporal response function (TRF) studies have predominantly used extended speech stimuli, recent research supports the efficacy of shorter sentences (<xref ref-type="bibr" rid="ref103">Vanthornhout et al., 2019</xref>; <xref ref-type="bibr" rid="ref16">Das et al., 2020</xref>; <xref ref-type="bibr" rid="ref68">Muncke et al., 2022</xref>; <xref ref-type="bibr" rid="ref93">Slaats et al., 2023</xref>; <xref ref-type="bibr" rid="ref80">Pendyala et al., 2024</xref>). Short sentences have been shown to provide reliable neural tracking responses and are effective in studying speech intelligibility (<xref ref-type="bibr" rid="ref68">Muncke et al., 2022</xref>; <xref ref-type="bibr" rid="ref93">Slaats et al., 2023</xref>).</p>
<p>Nevertheless, an important aspect that remains underexplored is how the timing of neural responses contributes to the comprehension of degraded speech. Most previous studies investigating the neural processing of degraded speech have primarily focused on the spatial localization of neural activity, which identifies which brain regions are activated in response to degraded speech. However, neural timing plays a crucial role in speech perception, as it determines how efficiently the brain processes degraded signals over time. Examining the onset, duration, and temporal patterns of neural responses can offer valuable insights into how the brain adapts to distorted auditory input. Understanding these time-varying dynamics is key to revealing how auditory signals are integrated and compensated for, particularly in noisy or unclear conditions.</p>
<p>Important findings into the time course of speech perception have emerged from eye-tracking experiments using the Visual World Paradigm (VWP; <xref ref-type="bibr" rid="ref45">Huettig et al., 2011</xref>). VWP posits that language is processed incrementally. In spoken words, the initial portions of a speech signal often do not provide enough information to immediately identify the word&#x2019;s meaning. This initial uncertainty leads to temporary but significant ambiguity among several possible words, resulting in a process known as lexical competition. Multiple potential word candidates are activated during this process and compete for selection. Lexical competition is critical in real-time language processing and is crucial for understanding how listeners decipher speech in everyday environments. Even when a word is clearly pronounced, listeners with normal hearing still experience temporary competition among these possible lexical candidates. The mechanisms through which typical listeners resolve these ambiguities have been extensively studied (e.g., <xref ref-type="bibr" rid="ref66">McQueen et al., 1999</xref>; <xref ref-type="bibr" rid="ref14">Dahan and Magnuson, 2006</xref>).</p>
<p>Immediate competition has been observed across various groups, including infants (<xref ref-type="bibr" rid="ref96">Swingley et al., 1999</xref>; <xref ref-type="bibr" rid="ref30">Fernald et al., 2001</xref>), adolescents (<xref ref-type="bibr" rid="ref90">Sekerina and Brooks, 2007</xref>; <xref ref-type="bibr" rid="ref87">Rigler et al., 2015</xref>), and postlingually deafened adults using cochlear implants (CIs) (<xref ref-type="bibr" rid="ref29">Farris-Trimble et al., 2014</xref>). However, CI users face unique challenges due to the degraded auditory signal provided by these devices (<xref ref-type="bibr" rid="ref106">Wilson and Dorman, 2008</xref>). These challenges highlight the differences in how CI users, especially prelingually deaf children, might process spoken language compared to typical listeners. In particular, prelingually deaf children using CIs may experience word recognition processes that differ significantly from typical listeners (<xref ref-type="bibr" rid="ref65">McMurray et al., 2017</xref>). Despite research on lexical competition across various groups, there is a notable lack of studies investigating the neural mechanisms underlying these processes, especially in cochlear implant users. Addressing this gap is crucial, as understanding the neural basis of these strategies could lead to improvements in CI technology and rehabilitation methods. In recent years, several groups have advanced our understanding of neural responses in cochlear implant users using electrophysiological and neuroimaging approaches (<xref ref-type="bibr" rid="ref20">Dimitrijevic et al., 2019</xref>; <xref ref-type="bibr" rid="ref60">Langner et al., 2021</xref>; <xref ref-type="bibr" rid="ref84">Prince et al., 2021</xref>). These studies provide important context for interpreting neural data in CI users and underscore the relevance of investigating neural processing alongside behavioral outcomes.</p>
<p>Recent auditory neuroscience frameworks suggest that predictive coding may provide a comprehensive explanation for how the brain processes degraded speech (<xref ref-type="bibr" rid="ref86">Rao and Ballard, 1999</xref>; <xref ref-type="bibr" rid="ref35">Friston, 2005</xref>). Predictive coding proposes that the brain continuously generates predictions about upcoming sensory input and minimizes the difference between expected and actual input through a process of error correction (<xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref43">Heilbron et al., 2022</xref>). In degraded listening conditions, listeners may rely more heavily on top-down predictions to compensate for the reduced acoustic detail. Importantly, predictive coding mechanisms can operate even during passive listening, supporting the idea that linguistic and auditory processing continues automatically without requiring active attention (<xref ref-type="bibr" rid="ref100">Todorovic et al., 2011</xref>; <xref ref-type="bibr" rid="ref2">Auksztulewicz and Friston, 2016</xref>; <xref ref-type="bibr" rid="ref89">Sedley et al., 2016</xref>; <xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>). This framework aligns with the observed neural tracking patterns in both early and late time windows under degraded speech conditions, offering a strong theoretical basis for understanding speech processing when auditory input is compromised.</p>
<p>Understanding the temporal dynamics of speech perception in CI users is crucial for developing more effective auditory prostheses and improving communication abilities (<xref ref-type="bibr" rid="ref92">Sharma et al., 2002b</xref>). To address this gap, it is necessary to delve into how the brain responds to various types of speech input, particularly under conditions that replicate the auditory experiences of CI users. Investigating neural responses to different forms of lexical competition in listeners requires analyzing their reactions to continuous speech. In response to speech features, the brain demonstrates neural tracking during language processing (<xref ref-type="bibr" rid="ref59">Lalor and Foxe, 2010</xref>). This neural tracking can provide insights into how different populations process speech, especially when the auditory signal is less than optimal.</p>
<p>Studies have used speech envelopes to measure neural responses to continuous speech, as these envelopes capture essential temporal dynamics of speech that correlate with EEG signals (e.g., <xref ref-type="bibr" rid="ref22">Ding and Simon, 2012a</xref>; <xref ref-type="bibr" rid="ref19">Di Liberto et al., 2015</xref>; <xref ref-type="bibr" rid="ref71">O&#x2019;Sullivan et al., 2015</xref>). Recent research showed that older adults exhibit stronger tracking of speech envelopes even in conditions with poorer signal-to-noise ratios (<xref ref-type="bibr" rid="ref18">Decruy et al., 2021</xref>; <xref ref-type="bibr" rid="ref48">Karunathilake et al., 2023a</xref>), possibly due to compensatory mechanisms to maintain comprehension (<xref ref-type="bibr" rid="ref83">Presacco et al., 2016</xref>). These findings suggest that the brain may adapt to degraded auditory conditions by enhancing certain neural processes to maintain speech understanding. This enhanced neural tracking likely reflects the increased cognitive effort required to resolve lexical competition in challenging listening conditions. Additionally, neural tracking has been studied to indicate how the brain processes new linguistic information, particularly in high lexical competition scenarios. Such studies often reveal specific neural signatures, such as the N400 event-related potential, which reflect the cognitive effort involved in integrating challenging or unexpected words within a sentence (<xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>; <xref ref-type="bibr" rid="ref55">Koskinen et al., 2020</xref>; <xref ref-type="bibr" rid="ref105">Weissbart et al., 2020</xref>; <xref ref-type="bibr" rid="ref101">Toffolo et al., 2022</xref>). These late neural tracking responses are typically delayed relative to word onset and are associated with higher cognitive effort required for semantic integration, especially under conditions of high lexical competition (<xref ref-type="bibr" rid="ref61">Lau et al., 2008</xref>; reviews on N400 responses: <xref ref-type="bibr" rid="ref57">Kutas and Federmeier, 2011</xref>; <xref ref-type="bibr" rid="ref32">Frank et al., 2015</xref>; <xref ref-type="bibr" rid="ref33">Frank and Willems, 2017</xref>).</p>
<p>Researchers often use noise-vocoded speech to replicate the auditory experience of cochlear implant users (<xref ref-type="bibr" rid="ref34">Friesen et al., 2001</xref>; <xref ref-type="bibr" rid="ref88">Rosen et al., 2013</xref>). This method allows researchers to simulate the auditory challenges faced by CI users and study how these challenges affect language processing at the neural level. In this study, we employed EEG recordings and speech envelope tracking techniques while presenting participants with both clear and degraded speech sentences. We aim to understand better how degraded auditory input impacts lexical processing by investigating the neural dynamics within the predictive coding framework. We specifically test whether degraded speech primarily elicits increased reliance on top-down predictions and enhanced error correction mechanisms that facilitate speech processing in challenging listening conditions. The findings from this study could contribute to developing more effective and targeted auditory training and rehabilitation strategies for CI users.</p>
<p>Here, we compare the TRF differences between two speech intelligibility conditions. Using scalp EEG, within-subject comparisons between cortical neural tracking of highly intelligible natural sentences and barely intelligible vocoded sentences were performed by extracting temporal envelope-related responses. In the early time range (approximately 100&#x202F;ms after sound onset), the N100 component (N1), typically associated with early auditory processing and sensory abstraction stages, is expected to show stronger responses in the natural sentences condition (<xref ref-type="bibr" rid="ref69">N&#x00E4;&#x00E4;t&#x00E4;nen, 2001</xref>; <xref ref-type="bibr" rid="ref56">Krumbholz et al., 2003</xref>; <xref ref-type="bibr" rid="ref73">Obleser et al., 2006</xref>). Similarly, the P2 component (approximately 200&#x202F;ms after sound onset), which follows the N1 and reflects higher-level auditory processing including attention and stimulus evaluation, is also anticipated to exhibit enhanced responses in the natural speech condition. Although traditionally linked with early sensory processing, stronger N1 and P2 responses in the natural sentence condition could indicate more efficient or immediate engagement with the speech signal, facilitating quicker lexical competition and selection (<xref ref-type="bibr" rid="ref12">Commuri et al., 2023</xref>).</p>
<p>Moreover, we anticipate stronger late responses in the vocoded sentences condition for the typical trajectory of semantic integration effort (N400, occurring around 300&#x2013;500&#x202F;ms after sentence onset). Recent studies employing TRF analyses have revealed not only early components such as N1 and P2, but also prominent later responses (often termed &#x201C;P400&#x201D; TRFs) that are thought to parallel the classic N400 ERP associated with semantic processing (e.g., <xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>). We hypothesized that degraded speech would elicit increased reliance on top-down predictions and enhanced error correction mechanisms, reflected in delayed and enhanced late neural responses, as the brain works to resolve increased prediction error (<xref ref-type="bibr" rid="ref2">Auksztulewicz and Friston, 2016</xref>; <xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref44">Heilbron and Chait, 2018</xref>).</p>
<p>We first tested the above hypothesis using sensor-space signals from central electrodes (i.e., FCz, Cz, and CPz) that are commonly employed to investigate auditory cortex activities related to speech perception (<xref ref-type="bibr" rid="ref81">Phillips et al., 2000</xref>; <xref ref-type="bibr" rid="ref69">N&#x00E4;&#x00E4;t&#x00E4;nen, 2001</xref>; <xref ref-type="bibr" rid="ref10">&#x010C;eponien et al., 2002</xref>; <xref ref-type="bibr" rid="ref102">Tremblay et al., 2003</xref>; <xref ref-type="bibr" rid="ref63">Martin et al., 2008</xref>; <xref ref-type="bibr" rid="ref6">Brandmeyer et al., 2013</xref>; <xref ref-type="bibr" rid="ref50">Kayser et al., 2015</xref>; <xref ref-type="bibr" rid="ref46">Jafarpisheh et al., 2016</xref>; <xref ref-type="bibr" rid="ref52">Khalighinejad et al., 2017</xref>; <xref ref-type="bibr" rid="ref95">Steinmetzger and Rosen, 2017</xref>; <xref ref-type="bibr" rid="ref25">Drennan and Lalor, 2019</xref>; <xref ref-type="bibr" rid="ref28">Etard and Reichenbach, 2019</xref>; <xref ref-type="bibr" rid="ref62">Mai and Wang, 2019</xref>; <xref ref-type="bibr" rid="ref97">Synigal et al., 2020</xref>). Subsequently, we further tested the same hypothesis in source space within six bilateral cortical regions of interest (ROI) that were selected based on previous studies that demonstrated the involvement of bilateral Heschl&#x2019;s gyrus (HG), the planum polare (PP), the planum temporale (PT), the supramarginal gyrus (SMG), inferior frontal gyrus (IFG), and the middle temporal gyrus (MTG).</p>
</sec>
<sec sec-type="materials|methods" id="sec2">
<label>2</label>
<title>Materials and methods</title>
<sec id="sec3">
<label>2.1</label>
<title>Participants</title>
<p>Fifty normal-hearing subjects aged between 20 and 33 (mean, 24.1&#x202F;years; standard deviation, 2.4; 25 men and 25 women) participated in this study. All study procedures were reviewed and approved by the Institutional Review Board of the University of Ulsan. All participants signed an informed consent form, and the study was carried out under approved guidelines.</p>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Behavioral speech intelligibility (SI) test and stimuli</title>
<p>The 4-channel vocoded and natural conditions were employed. The Korean Sentence Recognition Test for adults was conducted to obtain the behavioral SI scores prior to EEG data acquisition (<xref ref-type="bibr" rid="ref47">Jang et al., 2008</xref>). A 4-channel vocoder degraded the sentences to provide a lower speech intelligibility condition, as <xref ref-type="bibr" rid="ref107">Wilson et al. (1991)</xref> outlined. The vocoder implementation used a logarithmically-spaced filter bank spanning 200&#x2013;5,000&#x202F;Hz. The envelope of each frequency band was extracted using a 200&#x202F;Hz low-pass filter and used to modulate Gaussian white noise carriers, which were then synthesized sequentially to reconstruct the speech signal (<xref ref-type="bibr" rid="ref67">Mehta and Oxenham, 2017</xref>). This process preserved temporal envelope cues while removing spectral fine structure, resulting in moderate spectral degradation characteristic of 4-channel vocoded speech. The SI scores of the 4-channel vocoded conditions were significantly lower than those in the natural speech condition (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05; paired <italic>t</italic>-test, <xref ref-type="table" rid="tab1">Table 1</xref>). Ten continuous Korean sentences, each lasting less than 2.5&#x202F;s, were selected from the Korean Standard Sentence Lists for Adults (KS-SL-A) (<xref ref-type="bibr" rid="ref47">Jang et al., 2008</xref>). To avoid stimulus overlap between behavioral and EEG testing, the behavioral SI test employed 10 sentence sets that did not contain any of the sentences selected for the EEG experiment.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Means and standard deviations for word-recognition scores for both speech intelligibility conditions.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Speech conditions</th>
<th align="center" valign="top" colspan="2">Word-recognition score (%)</th>
<th align="center" valign="top" rowspan="2"><italic>p</italic>-value</th>
</tr>
<tr>
<th align="center" valign="top">Mean</th>
<th align="center" valign="top">SD</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">4-channels vocoded</td>
<td align="center" valign="top">78.6</td>
<td align="center" valign="top">12.56</td>
<td align="center" valign="top" rowspan="2">&#x002A;</td>
</tr>
<tr>
<td align="left" valign="middle">Natural</td>
<td align="center" valign="top">99.65</td>
<td align="center" valign="top">1.01</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>&#x002A;<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, SD, standard deviation.</p>
<p><italic>p</italic>-values were obtained from the paired <italic>t</italic>-test between the natural and 4-channels vocoded condition.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>EEG recording and processing</title>
<p>Participants were seated approximately 1&#x202F;m apart from two loudspeakers while watching a silent movie. Sentences were randomly played at 65&#x202F;dB SPL through the loudspeakers in a soundproof room under two conditions: passive listening to vocoded speech (degraded condition) and natural speech (clean condition). This presentation level was selected to ensure comfortable listening while maintaining consistent audibility across all participants. To minimize training effects and ensure that participants&#x2019; performance in the degraded condition was not influenced by prior exposure to clear speech, all participants first completed the degraded condition, followed by the clean condition. Critically, the identical set of 10 sentences was used in both the vocoded and natural conditions. Each sentence was repeated 100 times with an inter-stimulus interval of 3&#x202F;s in a randomized order within each condition block. This design allowed for direct comparison of neural responses to the same linguistic content under different acoustic intelligibility conditions. Brain activity during the passive listening tasks was recorded using a 64-channel EEG system (Biosemi Active 2 system, Biosemi Co., Netherlands) at a sampling rate of 2,048&#x202F;Hz. The EEG data were band-pass filtered between 1 and 57 Hz using a finite impulse response filter. Signals were subsequently downsampled to 256&#x202F;Hz and segmented into epochs of 3&#x202F;s, starting 0.5&#x202F;s before stimulus onset. The estimated potential was computed by averaging 100 epochs and filtered through a 1&#x2013;15&#x202F;Hz band-pass filter for further analysis.</p>
</sec>
<sec id="sec6">
<label>2.4</label>
<title>Control analysis: alpha power as a marker of alertness</title>
<p>To address potential confounding effects of block order on alertness and fatigue, we conducted a control analysis examining alpha power (8&#x2013;12&#x202F;Hz), a well-established electrophysiological marker of vigilance and cognitive engagement (<xref ref-type="bibr" rid="ref82">Pivik and Harman, 1995</xref>; <xref ref-type="bibr" rid="ref58">Lai et al., 2022</xref>). Alpha band activity (8&#x2013;12&#x202F;Hz) was extracted from the preprocessed EEG and averaged across parietal sites (Pz, POz). Mean alpha power was compared between vocoded and natural condition using paired <italic>t</italic>-tests (two-tailed, degrees of freedom&#x202F;=&#x202F;49, <italic>n</italic>&#x202F;=&#x202F;50). The absence of significant differences would indicate that condition effects are not confounded by fatigue/ alertness changes.</p>
</sec>
<sec id="sec7">
<label>2.5</label>
<title>TRF estimation</title>
<p>An auditory spectrogram was created, containing 128 spectrally resolved sub-band envelopes of the speech signals, using the NSL toolbox (<xref ref-type="bibr" rid="ref11">Chi et al., 2005</xref>). These sub-bands are logarithmically spaced between approximately 90 and 4,000&#x202F;Hz. The speech envelopes were obtained by averaging all sub-band envelopes across frequency bands. The speech envelopes were downsampled to a sampling rate of 256&#x202F;Hz to match the sampling rate of the EEG signals. Baseline normalization was performed by subtracting the baseline mean from the EEG signals and then dividing by the baseline standard deviation. The baseline period (0.5&#x202F;s before stimulus onset) was pooled across epochs and channels to estimate the baseline mean and standard deviation. This normalization was applied across epochs and channels to ensure consistency across epochs and preserve the relative power differences across channels.</p>
<p>The TRFs of natural and vocoded condition were estimated using the multivariate temporal response function (mTRF) Toolbox (<xref ref-type="bibr" rid="ref23">Ding and Simon, 2012b</xref>; <xref ref-type="bibr" rid="ref13">Crosse et al., 2016</xref>). The time window ranged from &#x2212;100 to 600&#x202F;ms. The ridge parameter (<italic>&#x03BB;</italic>), which control for overfitting, was searched over a range from 10<sup>&#x2212;6</sup> to 10<sup>6</sup>. For each &#x03BB;, TRF model performance was validated using a <italic>leave-one-out</italic> cross-validation (across epochs) approach, with performance quantified by the mean squared error (MSE) (averaged across channels) on unseen data (<xref ref-type="bibr" rid="ref9">Browne, 2000</xref>). Model performance for each condition at each &#x03BB; was then averaged over subjects. The &#x03BB; value that yielded the lowest MSE for both conditions was 750 and was subsequently used as the optimal ridge parameter. Finally, a grand average of TRFs was obtained for each condition.</p>
<p>The global field power (GFP) of TRFs was computed for each participant in both the natural and 4-channel vocoded speech conditions. For sensor-level analysis, paired <italic>t</italic>-tests (two-tailed, degrees of freedom&#x202F;=&#x202F;49, <italic>n</italic>&#x202F;=&#x202F;50) were performed at each time point to compare GFP between conditions. The resulting <italic>p</italic>-values were corrected for multiple comparisons using the false discovery rate (FDR) method (<xref ref-type="bibr" rid="ref4">Benjamini and Hochberg, 1995</xref>). Three continuous clusters of significant time periods (80&#x2013;103&#x202F;ms, 131&#x2013;185&#x202F;ms, and 376&#x2013;408&#x202F;ms) were identified, corresponding to the N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub> components, respectively, following conventional auditory evoked potential nomenclature. For each period, <italic>t</italic>-values comparing natural and vocoded conditions are shown in the figures using color coding. Additionally, GFPs for each condition were compared with their respective baselines (average amplitude from &#x2212;100&#x202F;ms to 0&#x202F;ms) to identify significant response periods (two-tailed paired t-test, degrees of freedom&#x202F;=&#x202F;49, <italic>n</italic>&#x202F;=&#x202F;50).</p>
</sec>
<sec id="sec8">
<label>2.6</label>
<title>TRF components validation: permutation testing</title>
<p>To validate that identified TRF components (N1, P2, and P400) represent genuine stimulus&#x2013;response coupling rather than chance-level noise or artifacts, permutation testing was conducted. Stimulus envelopes were shuffled relative to EEG data to generate a null distribution of TRF amplitudes expected under the assumption of no true stimulus&#x2013;response relationship. For each subject and condition (vocoded and natural), TRFs were computed using the identical analysis pipeline as the original analysis, except stimulus envelopes were shuffled relative to the EEG recordings. Valid TRFs (computed from original, non-shuffled stimulus-EEG pairings) were compared against this null distribution (shuffled TRFs) at each time point and each condition. Peak amplitudes of TRF components (N1, P2, and P400) were compared between valid TRFs and shuffled TRFs using paired <italic>t</italic>-tests (two-tailed, <italic>&#x03B1;</italic>&#x202F;=&#x202F;0.05, with FDR correction for multiple comparisons). Significantly higher valid TRF amplitudes would demonstrate that components reflect genuine neural tracking of the speech signal rather than noise.</p>
</sec>
<sec id="sec9">
<label>2.7</label>
<title>TRF at source level using the inverse problem</title>
<p>The inverse problem was solved to calculate the source level of the evoked potentials using the MNE-python toolbox with 64-electrode EEG signals at the sensor level. First, the linear inverse operator was assembled using the noise covariance computed from the baseline period of each epoch (&#x2212;500&#x202F;ms to 0&#x202F;ms) and the forward operators provided by FreeSurfer (<xref ref-type="bibr" rid="ref31">Fischl, 2012</xref>). Then, the inverse operator was used to project EEG signals to 20,484 voxels at the source level using dSPM noise normalization (<xref ref-type="bibr" rid="ref15">Dale et al., 2000</xref>). Only the orientation perpendicular to the cortical surface was extracted to compute the TRF at the source level.</p>
<p>The TRF of the source level was generated using the mTRF toolbox between the source current intensity and the speech temporal envelope at the same time points corresponding to the significant periods identified at the sensor level (N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub>) for each voxel. The ridge parameter <italic>&#x03BB;</italic> was set to 750, the same as the sensor level. For source-level analysis, paired <italic>t</italic>-tests (two-tailed, degrees of freedom&#x202F;=&#x202F;49, <italic>n</italic>&#x202F;=&#x202F;50) were performed at each of 20,484 voxels comparing natural and 4-channel vocoded conditions, with statistical significance set at <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05 and FDR correction applied across all voxels. The <italic>t</italic>-values are denoted by colors in the figures. Regions of interest were defined using the Destrieux atlas implemented in FreeSurfer (<xref ref-type="bibr" rid="ref31">Fischl, 2012</xref>). Six bilateral ROIs were selected based on their established roles in speech processing: HG, PP, TP, SMG, MTG, and IFG. These regions were chosen to capture both early auditory processing and higher-level language areas.</p>
</sec>
</sec>
<sec sec-type="results" id="sec10">
<label>3</label>
<title>Results</title>
<sec id="sec11">
<label>3.1</label>
<title>TRF at sensor level</title>
<p><xref ref-type="fig" rid="fig1">Figure 1A</xref> displays the grand average TRFs at central electrodes (FCz, Cz, CPz) for both conditions. <xref ref-type="fig" rid="fig1">Figure 1C</xref> shows the corresponding GFP, which quantifies the overall neural response strength across all electrodes. The mTRF analysis revealed distinct components: N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub> (<xref ref-type="fig" rid="fig1">Figure 1A</xref>). GFP analysis identified three time periods with statistically significant differences between conditions (FDR-corrected paired <italic>t</italic>-tests, <italic>t</italic>(49), <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05): 80&#x2013;103&#x202F;ms (N1<sub>TRF</sub>), 131&#x2013;185&#x202F;ms (P2<sub>TRF</sub>), and 376&#x2013;408&#x202F;ms (P400<sub>TRF</sub>), respectively. Notably, significant differences were observed between the natural and vocoded conditions at the central electrodes for N1<sub>TRF</sub> (80&#x2013;103&#x202F;ms) and P2<sub>TRF</sub> (131&#x2013;185&#x202F;ms), with larger amplitudes in the natural condition (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, paired t-test, FDR-corrected) (<xref ref-type="fig" rid="fig1">Figure 1B</xref>). Additionally, the P400<sub>TRF</sub> within 376,408 ms showed more extensive responses in the vocoded condition than in the natural condition (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, paired t-test, FDR-corrected) (<xref ref-type="fig" rid="fig1">Figure 1B</xref>).</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p><bold>(A)</bold> Grand averaged TRF at the central electrodes (i.e., FCz, Cz, CPz). The black boxes represent significantly different periods between the natural (red line) and vocoded (blue line) conditions. The solid line denotes the grand average TRF, while the shaded area denotes the standard error of TRFs in each condition. <bold>(B)</bold> The box plot represents TRF amplitude at the N1<sub>TRF</sub> (80&#x2013;103&#x202F;ms), P2<sub>TRF</sub> (131&#x2013;185&#x202F;ms), and P400<sub>TRF</sub> (376&#x2013;408&#x202F;ms). <bold>(C)</bold> The global field power (GFP) of TRF is in the natural (red) and vocoded (blue) condition. <bold>(D)</bold> Topography of TRF components in the natural (top), in the vocoded (middle), and dominance (bottom) at the N1TRF, P2TRF, and P400TRF time lags. Topographic dominance maps show the results of paired <italic>t</italic>-tests comparing natural vs. vocoded conditions at each electrode (<italic>t</italic><sub>49</sub>, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected). Red indicates significantly larger responses to natural speech, blue indicates significantly larger responses to vocoded speech. The <italic>t</italic>-value scale ranges from &#x2212;10 to +10 to accommodate the full range of statistical differences observed across electrodes. The central area significantly differs in the N1TRF, P2<sub>TRF</sub>, and P400<sub>TRF</sub> components.</p>
</caption>
<graphic xlink:href="fnins-20-1751421-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Graphs and visuals illustrate the impact of natural and vocoded conditions on TRF units and global field power. (A) Line graph showing TRF unit (&#x03B2;) over lag (ms) with significant intervals labeled a, b, c. (B) Box plots of TRF unit across different lag intervals exhibiting statistically significant differences. (C) Line graph of global field power of TRF over lag, highlighting significant intervals. (D) Topographical maps showing distribution of TRF (&#x03B2;) for both natural and vocoded conditions, alongside dominance maps. Red and blue colors indicate TRF levels and dominance t-values. Significant differences marked by asterisks.</alt-text>
</graphic>
</fig>
<p>The topographic maps of grand average TRF for the natural and vocoded conditions (first and second row) and their <italic>t</italic>-value (third row) at the N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub> time lags are shown in <xref ref-type="fig" rid="fig1">Figure 1D</xref>. In the natural condition, N1<sub>TRF</sub> and P2<sub>TRF</sub> components dominated the central region, peaking at electrode FC4 [t(49)&#x202F;=&#x202F;+7.49, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected] and electrode C3 [t(49)&#x202F;=&#x202F;+12.10, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected], respectively. In contrast, the vocoded condition exhibited dominance in the P400<sub>TRF</sub> component, peaking at electrode Cz [t(49)&#x202F;=&#x202F;&#x2212;7.49, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected]. Across all 64 channels, <italic>t</italic>-values ranged from &#x2212;7.49 to +12.10 (<xref ref-type="fig" rid="fig1">Figure 1D</xref>).</p>
</sec>
<sec id="sec12">
<label>3.2</label>
<title>Control analysis: alpha power as a marker of alertness</title>
<p>To address the potential confound of fixed block order, we conducted control analyses examining alpha power (8&#x2013;12&#x202F;Hz) of the TRF model across conditions (<xref ref-type="fig" rid="fig2">Figure 2A</xref>). Alpha power did not differ significantly between natural (mean: 9.69&#x202F;&#x03BC;V<sup>2</sup>/Hz, standard deviation (SD): 9.45) and vocoded (mean: 9.50&#x202F;&#x03BC;V<sup>2</sup>/Hz, SD: 7.94) condition [t(49)&#x202F;=&#x202F;0.32, <italic>p</italic>&#x202F;=&#x202F;0.75]. The absence of significant differences indicates that condition effects are not confounded by fatigue or alertness changes.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p><bold>(A)</bold> Alpha power (8&#x2013;12&#x202F;Hz) across conditions. <bold>(B)</bold> Mean squared error (MSE) of TRF model prediction. <bold>(C)</bold> Pearson correlation between stimulus envelope and neural response. Box plots display median (center line), interquartile range (box), and full range (whiskers) for natural (red) and vocoded (blue) conditions. Asterisks denote statistical significance: &#x002A;<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, &#x002A;&#x002A;&#x002A;<italic>p</italic>&#x202F;&#x003C;&#x202F;0.001; n.s., not significant (<italic>p</italic>&#x202F;&#x003E;&#x202F;0.05).</p>
</caption>
<graphic xlink:href="fnins-20-1751421-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Three box plots comparing "Natural" and "Vocoded" conditions in panels A, B, and C. Panel A shows no significant difference in microvolts squared per hertz. Panel B shows a significant difference in mean squared error with an asterisk, favoring Natural. Panel C shows a highly significant difference in Pearson's r with three asterisks, favoring Vocoded.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec13">
<label>3.3</label>
<title>TRF model performance</title>
<p>To characterize acoustic-neural coupling, we examined measurements of TRF model prediction performance: MSE and Pearson correlation (<xref ref-type="fig" rid="fig2">Figures 2B</xref>,<xref ref-type="fig" rid="fig2">C</xref>). MSE of the TRF model differed significantly between conditions [Natural: 1.297&#x202F;&#x00B1;&#x202F;0.121, Vocoded: 1.227&#x202F;&#x00B1;&#x202F;0.095, t(49)&#x202F;=&#x202F;2.51, <italic>p</italic>&#x202F;=&#x202F;0.016]. Pearson correlation between stimulus envelope and neural response was significantly higher in the vocoded condition compared to the natural condition [Natural: 0.030&#x202F;&#x00B1;&#x202F;0.009, Vocoded: 0.037&#x202F;&#x00B1;&#x202F;0.011, t(49)&#x202F;=&#x202F;&#x2212;6.43, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001]. Notably, across both natural and vocoded conditions, valid TRFs showed markedly better prediction performance than shuffled TRFs at the subject level for both mean squared error and Pearson correlation (all <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001; <xref ref-type="table" rid="tab2">Table 2</xref>), indicating that the reported model performance cannot be explained by chance-level correlations.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>TRF model prediction performance for valid and shuffled models in natural and vocoded speech conditions.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Condition / metric</th>
<th align="center" valign="top">Valid</th>
<th align="center" valign="top">Shuffled</th>
<th align="center" valign="top">t(49)</th>
<th align="center" valign="top"><italic>p</italic>-value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Natural / r</td>
<td align="center" valign="middle">0.030&#x202F;&#x00B1;&#x202F;0.009</td>
<td align="center" valign="middle">0.007&#x202F;&#x00B1;&#x202F;0.006</td>
<td align="center" valign="middle">21.75</td>
<td align="center" valign="middle">&#x003C;0.001</td>
</tr>
<tr>
<td align="left" valign="middle">Natural / MSE</td>
<td align="center" valign="middle">1.297&#x202F;&#x00B1;&#x202F;0.121</td>
<td align="center" valign="middle">1.299&#x202F;&#x00B1;&#x202F;0.122</td>
<td align="center" valign="middle">&#x2212;12.95</td>
<td align="center" valign="middle">&#x003C;0.001</td>
</tr>
<tr>
<td align="left" valign="middle">Vocoded / r</td>
<td align="center" valign="middle">0.037&#x202F;&#x00B1;&#x202F;0.011</td>
<td align="center" valign="middle">0.008&#x202F;&#x00B1;&#x202F;0.006</td>
<td align="center" valign="middle">22.12</td>
<td align="center" valign="middle">&#x003C;0.001</td>
</tr>
<tr>
<td align="left" valign="middle">Vocoded / MSE</td>
<td align="center" valign="middle">1.227&#x202F;&#x00B1;&#x202F;0.095</td>
<td align="center" valign="middle">1.229&#x202F;&#x00B1;&#x202F;0.095</td>
<td align="center" valign="middle">&#x2212;12.60</td>
<td align="center" valign="middle">&#x003C;0.001</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Entries show subject-level Pearson correlation and mean squared error (mean &#x00B1; SD).</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="sec14">
<label>3.4</label>
<title>TRF components validation against chance level</title>
<p>Valid TRFs showed significantly higher amplitudes than shuffled TRFs at all three identified components in both conditions (all <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001) (<xref ref-type="fig" rid="fig3">Figure 3</xref>). For the N1<sub>TRF</sub> component, valid TRF amplitudes significantly exceeded shuffled controls in both the vocoded condition [Valid: 0.29&#x202F;&#x00B1;&#x202F;0.11, Shuffled: 0.18&#x202F;&#x00B1;&#x202F;0.05, t(49)&#x202F;=&#x202F;6.29, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001] and the natural condition [Valid: 0.43&#x202F;&#x00B1;&#x202F;0.20, Shuffled: 0.18&#x202F;&#x00B1;&#x202F;0.06, t(49)&#x202F;=&#x202F;8.44, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001]. Similarly, P2<sub>TRF</sub> component amplitudes were significantly higher for valid TRFs compared to shuffled TRFs in the vocoded condition [Valid: 0.31&#x202F;&#x00B1;&#x202F;0.11, Shuffled: 0.19&#x202F;&#x00B1;&#x202F;0.05, t(49)&#x202F;=&#x202F;7.71, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001] and the natural condition [Valid: 0.55&#x202F;&#x00B1;&#x202F;0.21, Shuffled: 0.19&#x202F;&#x00B1;&#x202F;0.05, t(49)&#x202F;=&#x202F;12.01, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001]. Most notably, the P400<sub>TRF</sub> component&#x2014;the late response hypothesized to reflect compensatory processing for degraded speech&#x2014;showed a robust validation pattern. P400<sub>TRF</sub> amplitudes significantly exceeded shuffled controls in the vocoded condition [Valid: 0.40&#x202F;&#x00B1;&#x202F;0.13, Shuffled: 0.17&#x202F;&#x00B1;&#x202F;0.05, t(49)&#x202F;=&#x202F;13.30, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001] and in the natural condition [Valid: 0.28&#x202F;&#x00B1;&#x202F;0.10, Shuffled: 0.17&#x202F;&#x00B1;&#x202F;0.05, t(49)&#x202F;=&#x202F;6.92, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001]. This robust validation across all three components&#x2014;particularly the pronounced P400<sub>TRF</sub> in the vocoded condition&#x2014;demonstrates that observed TRF amplitudes reflect genuine stimulus&#x2013;response coupling rather than noise or chance-level correlations. The significantly higher valid TRF amplitudes across all time windows provide definitive evidence that the identified neural responses represent robust, stimulus-locked neural processes underlying intelligibility-driven speech processing.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p><bold>(A,B)</bold> Grand-averaged TRF waveforms for natural (<bold>A</bold>, red) and vocoded (<bold>B</bold>, blue) conditions, comparing actual stimulus&#x2013;response pairs (colored lines with shaded 95% confidence interval) against shuffled pairs (black lines). Horizontal gray bars denote the three time windows analyzed (a: 80&#x2013;103&#x202F;ms, b: 131&#x2013;185&#x202F;ms, c: 376&#x2013;408&#x202F;ms). <bold>(C,D)</bold> Global field power (GFP) of valid TRFs (colored) versus shuffled TRFs (gray) for natural <bold>(C)</bold> and vocoded <bold>(D)</bold> conditions within each time window. Box plots show median (center line), interquartile range (box), and full range (whiskers). Asterisks denote statistical significance: &#x002A;&#x002A;&#x002A; <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001.</p>
</caption>
<graphic xlink:href="fnins-20-1751421-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Graphs showing global field power of TRF for actual and shuffled pairs. (A) Red line for actual pairs peaks around 200 ms. (B) Blue line for actual pairs shows consistent peaks. (C) Box plots for 80-103, 131-185, and 376-408 ms intervals showing significant differences marked with asterisks for red plots. (D) Similar significant differences for blue plots. Labels a, b, and c mark intervals.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec15">
<label>3.5</label>
<title>TRF at source level</title>
<p>At the source level, the mTRF analysis was performed for the N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub> periods, investigating source localization related to speech intelligibility (<xref ref-type="fig" rid="fig4">Figure 4</xref>). The natural condition exhibited significant early dominance for N1<sub>TRF</sub>, with the peak voxel located in the left MTG [t(49)&#x202F;=&#x202F;8.67, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05], and for P2<sub>TRF</sub>, peaking in the left SMG [t(49)&#x202F;=&#x202F;13.67, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05]. In contrast, the vocoded condition showed dominant P400<sub>TRF</sub> responses, with the peak voxel located in the right IFG [t(49)&#x202F;=&#x202F;&#x2212;7.24, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05]. <xref ref-type="fig" rid="fig4">Figure 4</xref> also shows the dominance of both natural and vocoded conditions, with the color bar representing <italic>t</italic>-values from &#x2212;10 to 10 (paired t-test, FDR-corrected). As observed in <xref ref-type="fig" rid="fig4">Figure 4</xref>, the natural condition exhibited dominance in the early components (N1<sub>TRF</sub> and P2<sub>TRF</sub>), while the vocoded condition showed dominance in the late component (P400<sub>TRF</sub>).</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Source localization of the continuous speech-evoked potential at N1TRF (80&#x2013;103&#x202F;ms), P2TRF (131&#x2013;185&#x202F;ms), and P400TRF (376&#x2013;408&#x202F;ms) time lags. The dominance row denotes the <italic>T</italic>-value from paired <italic>t</italic>-tests at each of the 20,484 cortical voxels (blue: vocoded condition &#x003E; natural condition, red: natural condition &#x003E; vocoded condition), comparing natural vs. vocoded conditions (<italic>t</italic><sub>49</sub>, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected across all voxels). The mask was applied to display only voxels reaching statistical significance after multiple comparison correction. A mask was applied (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, FDR-corrected). TRF, temporal response function; FDR, false discovery rate.</p>
</caption>
<graphic xlink:href="fnins-20-1751421-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Brain activation maps displaying TRF beta and dominance t-values during auditory processing at three time frames: N1_TRF (80&#x2013;103 ms), P2_TRF (131&#x2013;185 ms), and P400_TRF (376&#x2013;408 ms). Natural and vocoded stimuli are compared, showing varying levels of activity visualized with a color scale from blue to red for TRF beta and an additional scale from blue to yellow for dominance t-value.</alt-text>
</graphic>
</fig>
<p>The ratios of significant voxels within 6 ROIs to the condition were calculated (<xref ref-type="table" rid="tab3">Table 3</xref>). For all ROIs, the early components (N1<sub>TRF</sub> and P2<sub>TRF</sub>) were dominant to the natural condition, while the late component (P400<sub>TRF</sub>) was dominant to the vocoded condition.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>The percentage of significantly dominant voxels in 6 regions of interest (ROIs) related to language processing for N1<sub>TRF</sub>, P2<sub>TRF</sub>, and P400<sub>TRF</sub> (paired <italic>t</italic>-test, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05, false discovery rate-corrected).</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center" valign="top" rowspan="3">Anatomical region</th>
<th align="center" valign="top" colspan="6">Dominance (voxel, %)</th>
</tr>
<tr>
<th align="center" valign="top" colspan="2">N1<sub>TRF</sub></th>
<th align="center" valign="top" colspan="2">P2<sub>TRF</sub></th>
<th align="center" valign="top" colspan="2">P400<sub>TRF</sub></th>
</tr>
<tr>
<th align="center" valign="top">Natural</th>
<th align="center" valign="top">Vocoded</th>
<th align="center" valign="top">Natural</th>
<th align="center" valign="top">Vocoded</th>
<th align="center" valign="top">Natural</th>
<th align="center" valign="top">Vocoded</th>
</tr>
</thead>
<tbody>
<tr>
<td align="center" valign="middle">IFG</td>
<td align="center" valign="middle">20.42</td>
<td align="center" valign="middle">&#x2014;</td>
<td align="center" valign="middle">50.00</td>
<td align="center" valign="middle">0.52</td>
<td align="center" valign="middle">8.90</td>
<td align="center" valign="middle">19.63</td>
</tr>
<tr>
<td align="center" valign="middle">SMG</td>
<td align="center" valign="middle">36.06</td>
<td align="center" valign="middle">0.39</td>
<td align="center" valign="middle">49.71</td>
<td align="center" valign="middle">&#x2014;</td>
<td align="center" valign="middle">&#x2014;</td>
<td align="center" valign="middle">14.23</td>
</tr>
<tr>
<td align="center" valign="middle">HG</td>
<td align="center" valign="middle">62.03</td>
<td align="center" valign="middle">5.06</td>
<td align="center" valign="middle">79.75</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">70.89</td>
</tr>
<tr>
<td align="center" valign="middle">PP of STG</td>
<td align="center" valign="middle">33.01</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">70.87</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">51.46</td>
</tr>
<tr>
<td align="center" valign="middle">TP of STG</td>
<td align="center" valign="middle">44.87</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">73.72</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">54.49</td>
</tr>
<tr>
<td align="center" valign="middle">MTG</td>
<td align="center" valign="middle">35.94</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="middle">83.85</td>
<td align="center" valign="middle">&#x2014;</td>
<td align="center" valign="middle">0.26</td>
<td align="center" valign="middle">61.72</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>HG, Heschl&#x2019;s gyrus; IFG, inferior frontal gyrus; PP, planum polare; MTG, middle temporal gyrus; TP, temporal plane; SMG, supramarginal gyrus; STG, superior temporal gyrus.</p>
<p>The proportion of voxels within each ROI showed significant dominance either in the natural or vocoded condition for each TRF component. The early components (N1<sub>TRF</sub> and P2<sub>TRF</sub>) exhibited higher percentages of significantly dominant voxels in the natural condition, while the late component (P400<sub>TRF</sub>) showed higher percentages in the vocoded condition.</p>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec sec-type="discussion" id="sec16">
<label>4</label>
<title>Discussion</title>
<p>The present study sought to elucidate the time course of degraded speech perception by examining the TRF in response to speech with varying intelligibility. By comparing clear and vocoded speech conditions, we identified distinct neural patterns associated with each condition, offering insights into how the brain manages lexical competition in diverse auditory environments.</p>
<p>Numerous studies have demonstrated that language processing can occur automatically, even when listeners are not actively attending to speech. Neural responses associated with lexical and semantic processing&#x2014;such as the N400&#x2014;have been observed in passive listening paradigms, with recent evidence demonstrating that semantic-level markers including N400 and semantic TRFs can be elicited in such contexts (<xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref108">Yang et al., 2024</xref>). Recent work using TRF and continuous speech paradigms further shows that robust neural tracking of linguistic features is present during passive listening, suggesting that core aspects of language comprehension operate automatically and do not require focused attention (<xref ref-type="bibr" rid="ref13">Crosse et al., 2016</xref>; <xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>). Importantly, clinical and neuroimaging research confirms that passive auditory paradigms can reliably elicit language lateralization and language-related brain activity (<xref ref-type="bibr" rid="ref75">Okahara et al., 2024</xref>). In addition, electrophysiological studies have shown that native language advantages in speech processing are evident even in passive listening, supporting the notion that language processing can proceed without active engagement (<xref ref-type="bibr" rid="ref108">Yang et al., 2024</xref>). Collectively, these findings support the view that passive listening is sufficient to engage the neural mechanisms underlying language processing, allowing researchers to investigate speech comprehension in more naturalistic and ecologically valid settings.</p>
<p>Within a predictive coding framework (<xref ref-type="bibr" rid="ref35">Friston, 2005</xref>), our findings demonstrate that accurate sensory predictions in clear speech reduce early TRF amplitudes, while increased prediction errors under degraded conditions lead to delayed, amplified late TRF components (P400<sub>TRF</sub>). Our findings revealed that early TRF components, specifically N1<sub>TRF</sub> and P2<sub>TRF</sub>, exhibited stronger responses in the clear speech condition, indicating more effective initial auditory processing when the speech signal is intelligible (<xref ref-type="bibr" rid="ref69">N&#x00E4;&#x00E4;t&#x00E4;nen, 2001</xref>; <xref ref-type="bibr" rid="ref56">Krumbholz et al., 2003</xref>). This aligns with the predictive coding framework, where accurate top-down predictions facilitate rapid and efficient early sensory processing, minimizing prediction error and supporting fluent speech comprehension (<xref ref-type="bibr" rid="ref35">Friston, 2005</xref>; <xref ref-type="bibr" rid="ref2">Auksztulewicz and Friston, 2016</xref>; <xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref43">Heilbron et al., 2022</xref>). In contrast, the P400<sub>TRF</sub> component was delayed and more pronounced in response to degraded speech. Within the predictive coding framework, this late component is interpreted as a neural signature of increased prediction error and the recruitment of higher-order cortical regions to resolve uncertainty (<xref ref-type="bibr" rid="ref37">Garrido et al., 2009b</xref>; <xref ref-type="bibr" rid="ref5">Blank and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref43">Heilbron et al., 2022</xref>). Although this response showed a positive deflection&#x2014;unlike the canonical N400 observed in traditional ERP studies&#x2014;such polarity differences are less critical in the context of TRF analyses. TRF analysis estimates the linear stimulus&#x2013;response mapping (<xref ref-type="bibr" rid="ref13">Crosse et al., 2016</xref>). As highlighted in previous TRF studies, the timing and functional context of neural responses are more informative than their polarity, which can vary based on analysis methods and referencing schemes (<xref ref-type="bibr" rid="ref19">Di Liberto et al., 2015</xref>; <xref ref-type="bibr" rid="ref13">Crosse et al., 2016</xref>; <xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>). Thus, our key finding is the delayed temporal dynamics of the response, likely reflecting compensatory or additional semantic integration processes as the brain attempts to maintain comprehension when early acoustic features are degraded.</p>
<p>While our paradigm involved passive listening, converging evidence from EEG, MEG, and fMRI studies demonstrates that key markers of linguistic processing&#x2014;such as the N400 and semantic TRFs&#x2014;can be elicited even in the absence of explicit tasks (<xref ref-type="bibr" rid="ref104">Wacongne et al., 2011</xref>; <xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref44">Heilbron and Chait, 2018</xref>; <xref ref-type="bibr" rid="ref108">Yang et al., 2024</xref>). The late TRF component (P400<sub>TRF</sub>) observed in our study is consistent in timing and cortical localization with these established markers of semantic and lexical processing. Although passive listening may not engage all aspects of conscious language comprehension, these findings support the interpretation that our results reflect genuine linguistic processing, not merely low-level auditory prediction.</p>
<p>An interesting dissociation emerged in measurements of TRF model performance between vocoded and natural conditions. MSE of the TRF model was significantly lower for vocoded conditions (vocoded: 1.227&#x202F;&#x00B1;&#x202F;0.095 vs. natural: 1.297&#x202F;&#x00B1;&#x202F;0.121, <italic>p</italic> =&#x202F;0.016). Pearson correlation was significantly higher for vocoded speech (vocoded: 0.037&#x202F;&#x00B1;&#x202F;0.011 vs. natural: 0.030&#x202F;&#x00B1;&#x202F;0.009, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001). This dissociation likely reflects differences in neural processing complexity, consistent with recent findings on speech intelligibility (<xref ref-type="bibr" rid="ref49">Karunathilake et al., 2023b</xref>). Vocoded speech with reduced acoustic variability and more consistent spectral structure constrains neural processing to envelope-driven mechanisms, resulting in more stimulus-locked and thus more predictable neural responses (lower MSE and higher correlation), while natural speech enables complex linguistic and semantic processing beyond envelope tracking, introducing variability that reduces correlation and increases MSE. Within a predictive coding framework (<xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>), degraded speech simplifies processing to lower-level sensory predictions relying on prediction error minimization (<xref ref-type="bibr" rid="ref5">Blank and Davis, 2016</xref>), whereas intact speech engages richer, higher-order mechanisms that support the observed P400<sub>TRF</sub> response as a compensatory error-correction process. An important consideration is whether the reduced P400<sub>TRF</sub> in the natural condition reflects intelligibility-driven processing or instead arises from repetition suppression and stimulus familiarity effects, given that identical sentence content was presented in both vocoded and natural condition blocks. The component-specific pattern observed in our data argues against a pure familiarity account. Repetition suppression typically manifests as global neural suppression across all components; our findings instead reveal selective enhancement of early components (N1<sub>TRF</sub>, P2<sub>TRF</sub>) coupled with suppression only of the late P400<sub>TRF</sub> component (<xref ref-type="bibr" rid="ref40">Grill-Spector et al., 2006</xref>; <xref ref-type="bibr" rid="ref36">Garrido et al., 2009a</xref>). This selective dissociation is inconsistent with global suppression predicted by repetition suppression or familiarity-driven accounts. Furthermore, alpha power did not differ significantly between blocks [t(49)&#x202F;=&#x202F;0.32, <italic>p</italic>&#x202F;=&#x202F;0.75], indicating that systematic changes in neural vigilance or fatigue do not confound the observed condition differences (<xref ref-type="bibr" rid="ref27">Engell and Mccarthy, 2014</xref>). The source localization of the P400<sub>TRF</sub> to the inferior frontal gyrus (IFG)&#x2014;a region consistently associated with semantic processing, error correction, and prediction error detection during language comprehension rather than novelty detection&#x2014;further supports an intelligibility-driven interpretation (<xref ref-type="bibr" rid="ref99">Thompson-Schill et al., 1997</xref>; <xref ref-type="bibr" rid="ref41">Hartwigsen et al., 2013</xref>; <xref ref-type="bibr" rid="ref98">Teghipco et al., 2023</xref>). Within the predictive coding framework, the vocoded condition necessitates increased reliance on top-down predictions and error correction, manifested as enhanced P400<sub>TRF</sub> activity, while the natural condition provides sufficient acoustic cues for rapid word recognition with minimal prediction error, reducing the need for late error-correction mechanisms (<xref ref-type="bibr" rid="ref5">Blank and Davis, 2016</xref>; <xref ref-type="bibr" rid="ref94">Sohoglu and Davis, 2016</xref>). The observed dissociation between component enhancement (N1<sub>TRF</sub>, P2<sub>TRF</sub>) and suppression (P400<sub>TRF</sub>), combined with stable neural state and appropriate source localization, is most parsimoniously explained by intelligibility-driven neural processing rather than repetition suppression or familiarity effects.</p>
<p>The observed patterns at both the sensor and source levels highlight the brain&#x2019;s dynamic adaptation to varying auditory environments. In clear speech conditions, TRF components likely reflect successful top-down predictions that minimize sensory surprise, supporting rapid word recognition. However, under vocoded conditions, the brain appears to compensate by prolonging processing, recruiting additional cortical resources to support comprehension. This adaptive mechanism is particularly relevant for individuals with hearing impairments, such as CI users, who often experience degraded auditory signals similar to those simulated by vocoded speech in this study (<xref ref-type="bibr" rid="ref106">Wilson and Dorman, 2008</xref>). The neural adaptability observed here suggests that even under less-than-optimal listening conditions, the brain may reallocate resources to maintain comprehension, albeit with increased cognitive effort (<xref ref-type="bibr" rid="ref39">Giraud and Truy, 2002</xref>; <xref ref-type="bibr" rid="ref1">Alain et al., 2018</xref>).</p>
<p>The increased P400<sub>TRF</sub> in the vocoded condition may reflect a broader engagement of cognitive resources that supports semantic processing when acoustic cues are degraded. Evidence from continuous speech TRF analyses indicates that late neural components originating from prefrontal and temporal language regions encode semantic and linguistic information, with enhanced responses during high lexical competition scenarios or when acoustic input is compromised (<xref ref-type="bibr" rid="ref51">Kegler et al., 2022</xref>). This finding is consistent with studies on aging, which show that older adults often exhibit stronger neural tracking of speech envelopes despite poorer signal-to-noise ratios, likely due to compensatory mechanisms aimed at preserving speech comprehension (<xref ref-type="bibr" rid="ref83">Presacco et al., 2016</xref>; <xref ref-type="bibr" rid="ref26">Duta and Plunkett, 2021</xref>; <xref ref-type="bibr" rid="ref109">Yip et al., 2021</xref>). These results suggest that under degraded conditions, the brain increasingly depends on higher-level predictive and semantic processes to support understanding (<xref ref-type="bibr" rid="ref24">Ding and Simon, 2013</xref>; <xref ref-type="bibr" rid="ref78">Peelle and Wingfield, 2016</xref>).</p>
<p>It is noteworthy that we used a passive listening condition where subjects watched a silent movie during auditory stimuli were presented. Such conditions are commonly used to divert participants&#x2019; attention while still enabling the study of automatic neural responses related to lexical competition and syntactic processing (<xref ref-type="bibr" rid="ref85">Pulverm&#x00FC;ller et al., 2008</xref>; <xref ref-type="bibr" rid="ref53">Kong et al., 2014</xref>). These previous findings indicate that speech processing can occur without explicit attention, highlighting the robustness of automatic linguistic and lexical processing mechanisms. Our finding is consistent; both short and long-latency TRF components appeared during passive listening, which may suggest speech processing occurs even in the absence of an explicit task. This finding suggests that passive listening paradigms may be a reliable method for studying degraded speech perception, particularly in clinical populations who may have difficulty sustaining attention during active tasks.</p>
<p>Understanding these neural mechanisms is crucial for improving auditory prostheses and rehabilitation strategies for CI users. Enhanced early neural responses in clear conditions could inform the development of auditory training programs to aim at strengthening early prediction-based processing. Conversely, heightened late responses in degraded conditions suggest the importance of interventions that enhance error monitoring and semantic integration. This could include auditory training focused on top-down prediction, or the use of visual or contextual cues to support comprehension under challenging conditions (<xref ref-type="bibr" rid="ref91">Sharma et al., 2002a</xref>; <xref ref-type="bibr" rid="ref38">Giraud and Lee, 2007</xref>).</p>
<p>While the present study provides insights into the neural processing of degraded speech, one notable limitation warrants discussion. The behavioral SI test revealed near-ceiling performance in the natural speech condition (mean&#x202F;=&#x202F;99.65%, SD&#x202F;=&#x202F;1.01%), precluding examination of correlations between behavioral SI and neural response measures. Future studies employing populations with greater SI variability&#x2014;such as older adults, individuals with hearing loss, or cochlear implant users&#x2014;could reveal informative between-subject associations. Additionally, the present study employed normal-hearing young adults, limiting generalizability to other populations with degraded auditory input. Future research should investigate how these neural mechanisms evolve with experience and training, particularly in auditory-impaired populations. Longitudinal studies could provide valuable insights into the neural plasticity associated with degraded speech processing and word recognition strategies. Additionally, exploring the interplay between cognitive factors such as attention, working memory, and neural responses could further elucidate how the brain manages comprehension in challenging listening conditions (<xref ref-type="bibr" rid="ref17">Davis and Johnsrude, 2007</xref>; <xref ref-type="bibr" rid="ref64">McGettigan et al., 2010</xref>). Future studies should include explicit word-onset and lexical surprisal predictors to confirm the semantic interpretation of the P400<sub>TRF</sub> component (<xref ref-type="bibr" rid="ref13">Crosse et al., 2016</xref>; <xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>).</p>
<p>While our findings are well explained by the predictive coding framework, an alternative interpretation is the &#x201C;wait-and-see&#x201D; strategy (<xref ref-type="bibr" rid="ref70">Norris et al., 2016</xref>; <xref ref-type="bibr" rid="ref65">McMurray et al., 2017</xref>). This approach suggests that listeners may delay lexical decisions and accumulate more acoustic information before committing to a word, especially under degraded conditions (<xref ref-type="bibr" rid="ref7">Brodbeck et al., 2018</xref>; <xref ref-type="bibr" rid="ref8">Broderick et al., 2018</xref>). The delayed P400<sub>TRF</sub> component observed in our study could reflect this adaptive, conservative processing style. Although our passive listening paradigm does not directly test this hypothesis, the presence of a delayed neural response in degraded conditions is consistent with this interpretation. Future studies employing active tasks or explicit decision-making paradigms could further clarify the contribution of &#x201C;wait-and-see&#x201D; mechanisms to speech comprehension under degraded listening conditions.</p>
<p>In conclusion, this study advances our understanding of the neural dynamics involved in predictive processing under varying speech intelligibility conditions. The differential engagement of early and late TRF components across clear and vocoded speech conditions underscores the brain&#x2019;s flexibility in processing spoken language. These findings have significant implications for developing auditory prostheses and highlight the importance of tailored rehabilitation strategies that account for the cognitive demands of different listening environments.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec17">
<title>Data availability statement</title>
<p>De-identified raw data supporting the conclusions of this article will be made available from the authors upon reasonable request, in accordance with ethics approval and participant consent.</p>
</sec>
<sec sec-type="ethics-statement" id="sec18">
<title>Ethics statement</title>
<p>The studies involving humans were approved by the Institutional Review Board of the University of Ulsan. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.</p>
</sec>
<sec sec-type="author-contributions" id="sec19">
<title>Author contributions</title>
<p>YN: Formal analysis, Writing &#x2013; review &#x0026; editing, Data curation, Visualization, Methodology, Validation, Software, Investigation, Writing &#x2013; original draft. LD: Writing &#x2013; original draft, Software, Investigation, Visualization, Writing &#x2013; review &#x0026; editing, Methodology, Formal analysis, Data curation. HJ: Data curation, Writing &#x2013; review &#x0026; editing, Investigation, Writing &#x2013; original draft. IC: Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing, Validation, Supervision, Conceptualization. JW: Writing &#x2013; original draft, Resources, Writing &#x2013; review &#x0026; editing, Project administration, Validation, Funding acquisition, Supervision, Conceptualization.</p>
</sec>
<sec sec-type="COI-statement" id="sec20">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec21">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec22">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alain</surname><given-names>C.</given-names></name> <name><surname>Du</surname><given-names>Y.</given-names></name> <name><surname>Bernstein</surname><given-names>L. J.</given-names></name> <name><surname>Barten</surname><given-names>T.</given-names></name> <name><surname>Banai</surname><given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Listening under difficult conditions: an activation likelihood estimation meta-analysis</article-title>. <source>Hum. Brain Mapp.</source> <volume>39</volume>, <fpage>2695</fpage>&#x2013;<lpage>2709</lpage>. doi: <pub-id pub-id-type="doi">10.1002/HBM.24031</pub-id>, <pub-id pub-id-type="pmid">29536592</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Auksztulewicz</surname><given-names>R.</given-names></name> <name><surname>Friston</surname><given-names>K.</given-names></name></person-group> (<year>2016</year>). <article-title>Repetition suppression and its contextual determinants in predictive coding</article-title>. <source>Cortex</source> <volume>80</volume>, <fpage>125</fpage>&#x2013;<lpage>140</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.CORTEX.2015.11.024</pub-id>, <pub-id pub-id-type="pmid">26861557</pub-id></mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Baltzell</surname><given-names>L. S.</given-names></name> <name><surname>Srinivasan</surname><given-names>R.</given-names></name> <name><surname>Richards</surname><given-names>V. M.</given-names></name></person-group> (<year>2017</year>). <article-title>The effect of prior knowledge and intelligibility on the cortical entrainment response to speech</article-title>. <source>J. Neurophysiol.</source> <volume>118</volume>, <fpage>3144</fpage>&#x2013;<lpage>3151</lpage>. doi: <pub-id pub-id-type="doi">10.1152/jn.00023.2017</pub-id>, <pub-id pub-id-type="pmid">28877963</pub-id></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Benjamini</surname><given-names>Y.</given-names></name> <name><surname>Hochberg</surname><given-names>Y.</given-names></name></person-group> (<year>1995</year>). <article-title>Controlling the false discovery rate: a practical and powerful approach to multiple testing</article-title>. <source>J. R. Stat. Soc. Ser. B Stat Methodol.</source> <volume>57</volume>, <fpage>289</fpage>&#x2013;<lpage>300</lpage>. doi: <pub-id pub-id-type="doi">10.1111/j.2517-6161.1995.tb02031.x</pub-id></mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Blank</surname><given-names>H.</given-names></name> <name><surname>Davis</surname><given-names>M. H.</given-names></name></person-group> (<year>2016</year>). <article-title>Prediction errors but not sharpened signals simulate multivoxel fMRI patterns during speech perception</article-title>. <source>PLoS Biol.</source> <volume>14</volume>:<fpage>e1002577</fpage>. doi: <pub-id pub-id-type="doi">10.1371/JOURNAL.PBIO.1002577</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brandmeyer</surname><given-names>A.</given-names></name> <name><surname>Farquhar</surname><given-names>J. D. R.</given-names></name> <name><surname>McQueen</surname><given-names>J. M.</given-names></name> <name><surname>Desain</surname><given-names>P. W. M.</given-names></name></person-group> (<year>2013</year>). <article-title>Decoding speech perception by native and non-native speakers using single-trial electrophysiological data</article-title>. <source>PLoS One</source> <volume>8</volume>:<fpage>e68261</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0068261</pub-id>, <pub-id pub-id-type="pmid">23874567</pub-id></mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brodbeck</surname><given-names>C.</given-names></name> <name><surname>Hong</surname><given-names>L. E.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2018</year>). <article-title>Rapid transformation from auditory to linguistic representations of continuous speech</article-title>. <source>Curr. Biol.</source> <volume>28</volume>, <fpage>3976</fpage>&#x2013;<lpage>3983.e5</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cub.2018.10.042</pub-id>, <pub-id pub-id-type="pmid">30503620</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Broderick</surname><given-names>M. P.</given-names></name> <name><surname>Anderson</surname><given-names>A. J.</given-names></name> <name><surname>Di Liberto</surname><given-names>G. M.</given-names></name> <name><surname>Crosse</surname><given-names>M. J.</given-names></name> <name><surname>Lalor</surname><given-names>E. C.</given-names></name></person-group> (<year>2018</year>). <article-title>Electrophysiological correlates of semantic dissimilarity reflect the comprehension of natural, narrative speech</article-title>. <source>Curr. Biol.</source> <volume>28</volume>, <fpage>803</fpage>&#x2013;<lpage>809.e3</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cub.2018.01.080</pub-id></mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Browne</surname><given-names>M. W.</given-names></name></person-group> (<year>2000</year>). <article-title>Cross-validation methods</article-title>. <source>J. Math. Psychol.</source> <volume>44</volume>, <fpage>108</fpage>&#x2013;<lpage>132</lpage>. doi: <pub-id pub-id-type="doi">10.1006/jmps.1999.1279</pub-id></mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>&#x010C;eponien</surname><given-names>R.</given-names></name> <name><surname>Rinne</surname><given-names>T.</given-names></name> <name><surname>N&#x00E4;&#x00E4;t&#x00E4;nen</surname><given-names>R.</given-names></name></person-group> (<year>2002</year>). <article-title>Maturation of cortical sound processing as indexed by event-related potentials</article-title>. <source>Clin. Neurophysiol.</source> <volume>113</volume>, <fpage>870</fpage>&#x2013;<lpage>882</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S1388-2457(02)00078-0</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chi</surname><given-names>T.</given-names></name> <name><surname>Ru</surname><given-names>P.</given-names></name> <name><surname>Shamma</surname><given-names>S. A.</given-names></name></person-group> (<year>2005</year>). <article-title>Multiresolution spectrotemporal analysis of complex sounds</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>118</volume>, <fpage>887</fpage>&#x2013;<lpage>906</lpage>. doi: <pub-id pub-id-type="doi">10.1121/1.1945807</pub-id>, <pub-id pub-id-type="pmid">16158645</pub-id></mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Commuri</surname><given-names>V.</given-names></name> <name><surname>Kulasingham</surname><given-names>J. P.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2023</year>). <article-title>Cortical responses time-locked to continuous speech in the high-gamma band depend on selective attention</article-title>. <source>Front. Neurosci.</source> <volume>17</volume>:<fpage>1264453</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNINS.2023.1264453/BIBTEX</pub-id></mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Crosse</surname><given-names>M. J.</given-names></name> <name><surname>Di Liberto</surname><given-names>G. M.</given-names></name> <name><surname>Bednar</surname><given-names>A.</given-names></name> <name><surname>Lalor</surname><given-names>E. C.</given-names></name></person-group> (<year>2016</year>). <article-title>The multivariate temporal response function (mTRF) toolbox: a MATLAB toolbox for relating neural signals to continuous stimuli</article-title>. <source>Front. Hum. Neurosci.</source> <volume>10</volume>:<fpage>604</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnhum.2016.00604</pub-id>, <pub-id pub-id-type="pmid">27965557</pub-id></mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Dahan</surname><given-names>D.</given-names></name> <name><surname>Magnuson</surname><given-names>J. S.</given-names></name></person-group> (<year>2006</year>). &#x201C;<article-title>Spoken word recognition</article-title>&#x201D; in <source>Handbook of psycholinguistics</source>. eds. <person-group person-group-type="editor"><name><surname>Traxler</surname><given-names>M. J.</given-names></name> <name><surname>Gernsbacher</surname><given-names>M. A.</given-names></name></person-group> (<publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Elsevier</publisher-name>), <fpage>249</fpage>&#x2013;<lpage>283</lpage>.</mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dale</surname><given-names>A. M.</given-names></name> <name><surname>Liu</surname><given-names>A. K.</given-names></name> <name><surname>Fischl</surname><given-names>B. R.</given-names></name> <name><surname>Buckner</surname><given-names>R. L.</given-names></name> <name><surname>Belliveau</surname><given-names>J. W.</given-names></name> <name><surname>Lewine</surname><given-names>J. D.</given-names></name> <etal/></person-group>. (<year>2000</year>). <article-title>Dynamic statistical parametric mapping: combining fMRI and MEG for high-resolution imaging of cortical activity</article-title>. <source>Neuron</source> <volume>26</volume>, <fpage>55</fpage>&#x2013;<lpage>67</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0896-6273(00)81138-1</pub-id>, <pub-id pub-id-type="pmid">10798392</pub-id></mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Das</surname><given-names>N.</given-names></name> <name><surname>Vanthornhout</surname><given-names>J.</given-names></name> <name><surname>Francart</surname><given-names>T.</given-names></name> <name><surname>Bertrand</surname><given-names>A.</given-names></name></person-group> (<year>2020</year>). <article-title>Stimulus-aware spatial filtering for single-trial neural response and temporal response function estimation in high-density EEG with applications in auditory research</article-title>. <source>NeuroImage</source> <volume>204</volume>:<fpage>116211</fpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROIMAGE.2019.116211</pub-id>, <pub-id pub-id-type="pmid">31546052</pub-id></mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Davis</surname><given-names>M. H.</given-names></name> <name><surname>Johnsrude</surname><given-names>I. S.</given-names></name></person-group> (<year>2007</year>). <article-title>Hearing speech sounds: top-down influences on the interface between audition and speech perception</article-title>. <source>Hear. Res.</source> <volume>229</volume>, <fpage>132</fpage>&#x2013;<lpage>147</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.heares.2007.01.014</pub-id>, <pub-id pub-id-type="pmid">17317056</pub-id></mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Decruy</surname><given-names>L.</given-names></name> <name><surname>Vanthornhout</surname><given-names>J.</given-names></name> <name><surname>Kuchinsky</surname><given-names>S. E.</given-names></name> <name><surname>Anderson</surname><given-names>S.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name> <name><surname>Francart</surname><given-names>T.</given-names></name></person-group> (<year>2021</year>). &#x201C;<article-title>Neural tracking of continuous speech is exaggerated in healthy aging and hearing impaired adults</article-title>&#x201D; in <source>Abstract book of 2021 midwinter meeting of Association for Research in otolaryngology</source>. Orlando, Florida, United states.</mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Di Liberto</surname><given-names>G. M.</given-names></name> <name><surname>O&#x2019;Sullivan</surname><given-names>J. A.</given-names></name> <name><surname>Lalor</surname><given-names>E. C.</given-names></name></person-group> (<year>2015</year>). <article-title>Low-frequency cortical entrainment to speech reflects phoneme-level processing</article-title>. <source>Curr. Biol.</source> <volume>25</volume>, <fpage>2457</fpage>&#x2013;<lpage>2465</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cub.2015.08.030</pub-id></mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dimitrijevic</surname><given-names>A.</given-names></name> <name><surname>Smith</surname><given-names>M. L.</given-names></name> <name><surname>Kadis</surname><given-names>D. S.</given-names></name> <name><surname>Moore</surname><given-names>D. R.</given-names></name></person-group> (<year>2019</year>). <article-title>Neural indices of listening effort in noisy environments</article-title>. <source>Sci. Rep.</source> <volume>9</volume>:<fpage>11278</fpage>. doi: <pub-id pub-id-type="doi">10.1038/S41598-019-47643-1</pub-id>, <pub-id pub-id-type="pmid">31375712</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname><given-names>N.</given-names></name> <name><surname>Chatterjee</surname><given-names>M.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2014</year>). <article-title>Robust cortical entrainment to the speech envelope relies on the spectro-temporal fine structure</article-title>. <source>NeuroImage</source> <volume>88</volume>, <fpage>41</fpage>&#x2013;<lpage>46</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROIMAGE.2013.10.054</pub-id>, <pub-id pub-id-type="pmid">24188816</pub-id></mixed-citation></ref>
<ref id="ref22"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname><given-names>N.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2012a</year>). <article-title>Emergence of neural encoding of auditory objects while listening to competing speakers</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>109</volume>, <fpage>11854</fpage>&#x2013;<lpage>11859</lpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.1205381109</pub-id>, <pub-id pub-id-type="pmid">22753470</pub-id></mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname><given-names>N.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2012b</year>). <article-title>Neural coding of continuous speech in auditory cortex during monaural and dichotic listening</article-title>. <source>J. Neurophysiol.</source> <volume>107</volume>, <fpage>78</fpage>&#x2013;<lpage>89</lpage>. doi: <pub-id pub-id-type="doi">10.1152/jn.00297.2011</pub-id>, <pub-id pub-id-type="pmid">21975452</pub-id></mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname><given-names>N.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2013</year>). <article-title>Adaptive temporal encoding leads to a background-insensitive cortical representation of speech</article-title>. <source>J. Neurosci.</source> <volume>33</volume>, <fpage>5728</fpage>&#x2013;<lpage>5735</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.5297-12.2013</pub-id>, <pub-id pub-id-type="pmid">23536086</pub-id></mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Drennan</surname><given-names>D. P.</given-names></name> <name><surname>Lalor</surname><given-names>E. C.</given-names></name></person-group> (<year>2019</year>). <article-title>Cortical tracking of complex sound envelopes: modeling the changes in response with intensity</article-title>. <source>eNeuro</source> <volume>6</volume>:<fpage>eneuro 0082</fpage>&#x2013;<lpage>19-2019</lpage>. doi: <pub-id pub-id-type="doi">10.1523/eneuro.0082-19.2019</pub-id></mixed-citation></ref>
<ref id="ref26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Duta</surname><given-names>M.</given-names></name> <name><surname>Plunkett</surname><given-names>K.</given-names></name></person-group> (<year>2021</year>). <article-title>A neural network model of lexical-semantic competition during spoken word recognition</article-title>. <source>Front. Hum. Neurosci.</source> <volume>15</volume>:<fpage>700281</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNHUM.2021.700281/BIBTEX</pub-id></mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Engell</surname><given-names>A. D.</given-names></name> <name><surname>Mccarthy</surname><given-names>G.</given-names></name></person-group> (<year>2014</year>). <article-title>Repetition suppression of face-selective evoked and induced EEG recorded from human cortex</article-title>. <source>Hum. Brain Mapp.</source> <volume>35</volume>, <fpage>4155</fpage>&#x2013;<lpage>4162</lpage>. doi: <pub-id pub-id-type="doi">10.1002/HBM.22467</pub-id>, <pub-id pub-id-type="pmid">24677530</pub-id></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Etard</surname><given-names>O.</given-names></name> <name><surname>Reichenbach</surname><given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Neural speech tracking in the theta and in the delta frequency band differentially encode clarity and comprehension of speech in noise</article-title>. <source>J. Neurosci.</source> <volume>39</volume>, <fpage>5750</fpage>&#x2013;<lpage>5759</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.1828-18.2019</pub-id>, <pub-id pub-id-type="pmid">31109963</pub-id></mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Farris-Trimble</surname><given-names>A.</given-names></name> <name><surname>McMurray</surname><given-names>B.</given-names></name> <name><surname>Cigrand</surname><given-names>N.</given-names></name> <name><surname>Bruce Tomblin</surname><given-names>J.</given-names></name></person-group> (<year>2014</year>). <article-title>The process of spoken word recognition in the face of signal degradation</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>40</volume>, <fpage>308</fpage>&#x2013;<lpage>327</lpage>. doi: <pub-id pub-id-type="doi">10.1037/A0034353</pub-id>, <pub-id pub-id-type="pmid">24041330</pub-id></mixed-citation></ref>
<ref id="ref30"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fernald</surname><given-names>A.</given-names></name> <name><surname>Swingley</surname><given-names>D.</given-names></name> <name><surname>Pinto</surname><given-names>J. P.</given-names></name></person-group> (<year>2001</year>). <article-title>When half a word is enough: infants can recognize spoken words using partial phonetic information</article-title>. <source>Child Dev.</source> <volume>72</volume>, <fpage>1003</fpage>&#x2013;<lpage>1015</lpage>. doi: <pub-id pub-id-type="doi">10.1111/1467-8624.00331</pub-id>, <pub-id pub-id-type="pmid">11480931</pub-id></mixed-citation></ref>
<ref id="ref31"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fischl</surname><given-names>B.</given-names></name></person-group> (<year>2012</year>). <article-title>FreeSurfer</article-title>. <source>NeuroImage</source> <volume>62</volume>, <fpage>774</fpage>&#x2013;<lpage>781</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROIMAGE.2012.01.021</pub-id>, <pub-id pub-id-type="pmid">22248573</pub-id></mixed-citation></ref>
<ref id="ref32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Frank</surname><given-names>S. L.</given-names></name> <name><surname>Otten</surname><given-names>L. J.</given-names></name> <name><surname>Galli</surname><given-names>G.</given-names></name> <name><surname>Vigliocco</surname><given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>The ERP response to the amount of information conveyed by words in sentences</article-title>. <source>Brain Lang.</source> <volume>140</volume>, <fpage>1</fpage>&#x2013;<lpage>11</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.bandl.2014.10.006</pub-id>, <pub-id pub-id-type="pmid">25461915</pub-id></mixed-citation></ref>
<ref id="ref33"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Frank</surname><given-names>S. L.</given-names></name> <name><surname>Willems</surname><given-names>R. M.</given-names></name></person-group> (<year>2017</year>). <article-title>Word predictability and semantic similarity show distinct patterns of brain activity during language comprehension</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>32</volume>, <fpage>1192</fpage>&#x2013;<lpage>1203</lpage>. doi: <pub-id pub-id-type="doi">10.1080/23273798.2017.1323109</pub-id></mixed-citation></ref>
<ref id="ref34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Friesen</surname><given-names>L. M.</given-names></name> <name><surname>Shannon</surname><given-names>R. V.</given-names></name> <name><surname>Baskent</surname><given-names>D.</given-names></name> <name><surname>Wang</surname><given-names>X.</given-names></name></person-group> (<year>2001</year>). <article-title>Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>110</volume>, <fpage>1150</fpage>&#x2013;<lpage>1163</lpage>. doi: <pub-id pub-id-type="doi">10.1121/1.1381538</pub-id>, <pub-id pub-id-type="pmid">11519582</pub-id></mixed-citation></ref>
<ref id="ref35"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Friston</surname><given-names>K.</given-names></name></person-group> (<year>2005</year>). <article-title>A theory of cortical responses</article-title>. <source>Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci.</source> <volume>360</volume>, <fpage>815</fpage>&#x2013;<lpage>836</lpage>. doi: <pub-id pub-id-type="doi">10.1098/RSTB.2005.1622</pub-id>, <pub-id pub-id-type="pmid">15937014</pub-id></mixed-citation></ref>
<ref id="ref36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garrido</surname><given-names>M. I.</given-names></name> <name><surname>Kilner</surname><given-names>J. M.</given-names></name> <name><surname>Kiebel</surname><given-names>S. J.</given-names></name> <name><surname>Stephan</surname><given-names>K. E.</given-names></name> <name><surname>Baldeweg</surname><given-names>T.</given-names></name> <name><surname>Friston</surname><given-names>K. J.</given-names></name></person-group> (<year>2009a</year>). <article-title>Repetition suppression and plasticity in the human brain</article-title>. <source>NeuroImage</source> <volume>48</volume>, <fpage>269</fpage>&#x2013;<lpage>279</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROIMAGE.2009.06.034</pub-id>, <pub-id pub-id-type="pmid">19540921</pub-id></mixed-citation></ref>
<ref id="ref37"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garrido</surname><given-names>M. I.</given-names></name> <name><surname>Kilner</surname><given-names>J. M.</given-names></name> <name><surname>Stephan</surname><given-names>K. E.</given-names></name> <name><surname>Friston</surname><given-names>K. J.</given-names></name></person-group> (<year>2009b</year>). <article-title>The mismatch negativity: a review of underlying mechanisms</article-title>. <source>Clin. Neurophysiol.</source> <volume>120</volume>, <fpage>453</fpage>&#x2013;<lpage>463</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.CLINPH.2008.11.029</pub-id></mixed-citation></ref>
<ref id="ref38"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Giraud</surname><given-names>A.-L.</given-names></name> <name><surname>Lee</surname><given-names>H.-J.</given-names></name></person-group> (<year>2007</year>). <article-title>Predicting cochlear implant outcome from brain organisation in the deaf</article-title>. <source>Restor. Neurol. Neurosci.</source> <volume>25</volume>, <fpage>381</fpage>&#x2013;<lpage>390</lpage>. doi: <pub-id pub-id-type="doi">10.3233/RNN-2007-253420</pub-id>, <pub-id pub-id-type="pmid">17943013</pub-id></mixed-citation></ref>
<ref id="ref39"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Giraud</surname><given-names>A. L.</given-names></name> <name><surname>Truy</surname><given-names>E.</given-names></name></person-group> (<year>2002</year>). <article-title>The contribution of visual areas to speech comprehension: a PET study in cochlear implants patients and normal-hearing subjects</article-title>. <source>Neuropsychologia</source> <volume>40</volume>, <fpage>1562</fpage>&#x2013;<lpage>1569</lpage>. doi: <pub-id pub-id-type="doi">10.1016/s0028-3932(02)00023-4</pub-id>, <pub-id pub-id-type="pmid">11985837</pub-id></mixed-citation></ref>
<ref id="ref40"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Grill-Spector</surname><given-names>K.</given-names></name> <name><surname>Henson</surname><given-names>R.</given-names></name> <name><surname>Martin</surname><given-names>A.</given-names></name></person-group> (<year>2006</year>). <article-title>Repetition and the brain: neural models of stimulus-specific effects</article-title>. <source>Trends Cogn. Sci.</source> <volume>10</volume>, <fpage>14</fpage>&#x2013;<lpage>23</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.tics.2005.11.006</pub-id>, <pub-id pub-id-type="pmid">16321563</pub-id></mixed-citation></ref>
<ref id="ref41"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hartwigsen</surname><given-names>G.</given-names></name> <name><surname>Saur</surname><given-names>D.</given-names></name> <name><surname>Price</surname><given-names>C. J.</given-names></name> <name><surname>Ulmer</surname><given-names>S.</given-names></name> <name><surname>Baumgaertner</surname><given-names>A.</given-names></name> <name><surname>Siebner</surname><given-names>H. R.</given-names></name></person-group> (<year>2013</year>). <article-title>Perturbation of the left inferior frontal gyrus triggers adaptive plasticity in the right homologous area during speech production</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>110</volume>, <fpage>16402</fpage>&#x2013;<lpage>16407</lpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.1310190110</pub-id>, <pub-id pub-id-type="pmid">24062469</pub-id></mixed-citation></ref>
<ref id="ref42"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hauswald</surname><given-names>A.</given-names></name> <name><surname>Keitel</surname><given-names>A.</given-names></name> <name><surname>Chen</surname><given-names>Y.-P.</given-names></name> <name><surname>R&#x00F6;sch</surname><given-names>S.</given-names></name> <name><surname>Weisz</surname><given-names>N.</given-names></name></person-group> (<year>2022</year>). <article-title>Degradation levels of continuous speech affect neural speech tracking and alpha power differently</article-title>. <source>Eur. J. Neurosci.</source> <volume>55</volume>, <fpage>3288</fpage>&#x2013;<lpage>3302</lpage>. doi: <pub-id pub-id-type="doi">10.1111/ejn.14912</pub-id>, <pub-id pub-id-type="pmid">32687616</pub-id></mixed-citation></ref>
<ref id="ref43"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Heilbron</surname><given-names>M.</given-names></name> <name><surname>Armeni</surname><given-names>K.</given-names></name> <name><surname>Schoffelen</surname><given-names>J. M.</given-names></name> <name><surname>Hagoort</surname><given-names>P.</given-names></name> <name><surname>De Lange</surname><given-names>F. P.</given-names></name></person-group> (<year>2022</year>). <article-title>A hierarchy of linguistic predictions during natural language comprehension</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>119</volume>:<fpage>e2201968119</fpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.2201968119</pub-id>, <pub-id pub-id-type="pmid">35921434</pub-id></mixed-citation></ref>
<ref id="ref44"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Heilbron</surname><given-names>M.</given-names></name> <name><surname>Chait</surname><given-names>M.</given-names></name></person-group> (<year>2018</year>). <article-title>Great expectations: is there evidence for predictive coding in auditory cortex?</article-title> <source>Neuroscience</source> <volume>389</volume>, <fpage>54</fpage>&#x2013;<lpage>73</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROSCIENCE.2017.07.061</pub-id>, <pub-id pub-id-type="pmid">28782642</pub-id></mixed-citation></ref>
<ref id="ref45"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huettig</surname><given-names>F.</given-names></name> <name><surname>Rommers</surname><given-names>J.</given-names></name> <name><surname>Meyer</surname><given-names>A. S.</given-names></name></person-group> (<year>2011</year>). <article-title>Using the visual world paradigm to study language processing: a review and critical evaluation</article-title>. <source>Acta Psychol.</source> <volume>137</volume>, <fpage>151</fpage>&#x2013;<lpage>171</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.ACTPSY.2010.11.003</pub-id>, <pub-id pub-id-type="pmid">21288498</pub-id></mixed-citation></ref>
<ref id="ref46"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jafarpisheh</surname><given-names>A. S.</given-names></name> <name><surname>Jafari</surname><given-names>A. H.</given-names></name> <name><surname>Abolhassani</surname><given-names>M.</given-names></name> <name><surname>Farhadi</surname><given-names>M.</given-names></name> <name><surname>Sadjedi</surname><given-names>H.</given-names></name> <name><surname>Pourbakht</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Nonlinear feature extraction for objective classification of complex auditory brainstem responses to diotic perceptually critical consonant-vowel syllables</article-title>. <source>Auris Nasus Larynx</source> <volume>43</volume>, <fpage>37</fpage>&#x2013;<lpage>44</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.anl.2015.06.003</pub-id>, <pub-id pub-id-type="pmid">26304760</pub-id></mixed-citation></ref>
<ref id="ref47"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jang</surname><given-names>H.</given-names></name> <name><surname>Lee</surname><given-names>J.</given-names></name> <name><surname>Lim</surname><given-names>D.</given-names></name> <name><surname>Lee</surname><given-names>K.</given-names></name> <name><surname>Jeon</surname><given-names>A.</given-names></name> <name><surname>Jung</surname><given-names>E.</given-names></name></person-group> (<year>2008</year>). <article-title>Development of Korean standard sentence lists for sentence recognition tests</article-title>. <source>Audiol</source> <volume>4</volume>, <fpage>161</fpage>&#x2013;<lpage>177</lpage>. doi: <pub-id pub-id-type="doi">10.21848/audiol.2008.4.2.161</pub-id></mixed-citation></ref>
<ref id="ref48"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Karunathilake</surname><given-names>I. M. D.</given-names></name> <name><surname>Dunlap</surname><given-names>J. L.</given-names></name> <name><surname>Perera</surname><given-names>J.</given-names></name> <name><surname>Presacco</surname><given-names>A.</given-names></name> <name><surname>Decruy</surname><given-names>L.</given-names></name> <name><surname>Anderson</surname><given-names>S.</given-names></name> <etal/></person-group>. (<year>2023a</year>). <article-title>Effects of aging on cortical representations of continuous speech</article-title>. <source>J. Neurophysiol.</source> <volume>129</volume>, <fpage>1359</fpage>&#x2013;<lpage>1377</lpage>. doi: <pub-id pub-id-type="doi">10.1152/JN.00356.2022</pub-id></mixed-citation></ref>
<ref id="ref49"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Karunathilake</surname><given-names>I. M. D.</given-names></name> <name><surname>Kulasingham</surname><given-names>J. P.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name></person-group> (<year>2023b</year>). <article-title>Neural tracking measures of speech intelligibility: manipulating intelligibility while keeping acoustics unchanged</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>120</volume>:<fpage>e2309166120</fpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.2309166120</pub-id>, <pub-id pub-id-type="pmid">38032934</pub-id></mixed-citation></ref>
<ref id="ref50"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kayser</surname><given-names>S. J.</given-names></name> <name><surname>Ince</surname><given-names>R. A. A.</given-names></name> <name><surname>Gross</surname><given-names>J.</given-names></name> <name><surname>Kayser</surname><given-names>C.</given-names></name></person-group> (<year>2015</year>). <article-title>Irregular speech rate dissociates auditory cortical entrainment, evoked responses, and frontal alpha</article-title>. <source>J. Neurosci.</source> <volume>35</volume>, <fpage>14691</fpage>&#x2013;<lpage>14701</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2243-15.2015</pub-id>, <pub-id pub-id-type="pmid">26538641</pub-id></mixed-citation></ref>
<ref id="ref51"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kegler</surname><given-names>M.</given-names></name> <name><surname>Weissbart</surname><given-names>H.</given-names></name> <name><surname>Reichenbach</surname><given-names>T.</given-names></name></person-group> (<year>2022</year>). <article-title>The neural response at the fundamental frequency of speech is modulated by word-level acoustic and linguistic information</article-title>. <source>Front. Neurosci.</source> <volume>16</volume>:<fpage>915744</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNINS.2022.915744/BIBTEX</pub-id></mixed-citation></ref>
<ref id="ref52"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khalighinejad</surname><given-names>B.</given-names></name> <name><surname>Cruzatto da Silva</surname><given-names>G.</given-names></name> <name><surname>Mesgarani</surname><given-names>N.</given-names></name></person-group> (<year>2017</year>). <article-title>Dynamic encoding of acoustic features in neural responses to continuous speech</article-title>. <source>J. Neurosci.</source> <volume>37</volume>, <fpage>2176</fpage>&#x2013;<lpage>2185</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2383-16.2017</pub-id>, <pub-id pub-id-type="pmid">28119400</pub-id></mixed-citation></ref>
<ref id="ref53"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kong</surname><given-names>Y. Y.</given-names></name> <name><surname>Mullangi</surname><given-names>A.</given-names></name> <name><surname>Ding</surname><given-names>N.</given-names></name></person-group> (<year>2014</year>). <article-title>Differential modulation of auditory responses to attended and unattended speech in different listening conditions</article-title>. <source>Hear. Res.</source> <volume>316</volume>, <fpage>73</fpage>&#x2013;<lpage>81</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.HEARES.2014.07.009</pub-id></mixed-citation></ref>
<ref id="ref54"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>K&#x00F6;sem</surname><given-names>A.</given-names></name> <name><surname>Dai</surname><given-names>B.</given-names></name> <name><surname>McQueen</surname><given-names>J. M.</given-names></name> <name><surname>Hagoort</surname><given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>Neural tracking of speech envelope does not unequivocally reflect intelligibility</article-title>. <source>NeuroImage</source> <volume>272</volume>:<fpage>120040</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2023.120040</pub-id></mixed-citation></ref>
<ref id="ref55"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Koskinen</surname><given-names>M.</given-names></name> <name><surname>Kurimo</surname><given-names>M.</given-names></name> <name><surname>Gross</surname><given-names>J.</given-names></name> <name><surname>Hyv&#x00E4;rinen</surname><given-names>A.</given-names></name> <name><surname>Hari</surname><given-names>R.</given-names></name></person-group> (<year>2020</year>). <article-title>Brain activity reflects the predictability of word sequences in listened continuous speech</article-title>. <source>NeuroImage</source> <volume>219</volume>:<fpage>116936</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2020.116936</pub-id></mixed-citation></ref>
<ref id="ref56"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Krumbholz</surname><given-names>K.</given-names></name> <name><surname>Patterson</surname><given-names>R. D.</given-names></name> <name><surname>Seither-Preisler</surname><given-names>A.</given-names></name> <name><surname>Lammertmann</surname><given-names>C.</given-names></name> <name><surname>L&#x00FC;tkenh&#x00F6;ner</surname><given-names>B.</given-names></name></person-group> (<year>2003</year>). <article-title>Neuromagnetic evidence for a pitch processing center in Heschl&#x2019;s gyrus</article-title>. <source>Cereb. Cortex</source> <volume>13</volume>, <fpage>765</fpage>&#x2013;<lpage>772</lpage>. doi: <pub-id pub-id-type="doi">10.1093/cercor/13.7.765</pub-id>, <pub-id pub-id-type="pmid">12816892</pub-id></mixed-citation></ref>
<ref id="ref57"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kutas</surname><given-names>M.</given-names></name> <name><surname>Federmeier</surname><given-names>K. D.</given-names></name></person-group> (<year>2011</year>). <article-title>Thirty years and counting: finding meaning in the N400 component of the event-related brain potential (ERP)</article-title>. <source>Annu. Rev. Psychol.</source> <volume>62</volume>, <fpage>621</fpage>&#x2013;<lpage>647</lpage>. doi: <pub-id pub-id-type="doi">10.1146/annurev.psych.093008.131123</pub-id>, <pub-id pub-id-type="pmid">20809790</pub-id></mixed-citation></ref>
<ref id="ref58"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lai</surname><given-names>J.</given-names></name> <name><surname>Price</surname><given-names>C. N.</given-names></name> <name><surname>Bidelman</surname><given-names>G. M.</given-names></name></person-group> (<year>2022</year>). <article-title>Brainstem speech encoding is dynamically shaped online by fluctuations in cortical &#x03B1; state</article-title>. <source>NeuroImage</source> <volume>263</volume>:<fpage>119627</fpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROIMAGE.2022.119627</pub-id>, <pub-id pub-id-type="pmid">36122686</pub-id></mixed-citation></ref>
<ref id="ref59"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lalor</surname><given-names>E. C.</given-names></name> <name><surname>Foxe</surname><given-names>J. J.</given-names></name></person-group> (<year>2010</year>). <article-title>Neural responses to uninterrupted natural speech can be extracted with precise temporal resolution</article-title>. <source>Eur. J. Neurosci.</source> <volume>31</volume>, <fpage>189</fpage>&#x2013;<lpage>193</lpage>. doi: <pub-id pub-id-type="doi">10.1111/j.1460-9568.2009.07055.x</pub-id>, <pub-id pub-id-type="pmid">20092565</pub-id></mixed-citation></ref>
<ref id="ref60"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Langner</surname><given-names>F.</given-names></name> <name><surname>Arenberg</surname><given-names>J. G.</given-names></name> <name><surname>B&#x00FC;chner</surname><given-names>A.</given-names></name> <name><surname>Nogueira</surname><given-names>W.</given-names></name></person-group> (<year>2021</year>). <article-title>Assessing the relationship between neural health measures and speech performance with simultaneous electric stimulation in cochlear implant listeners</article-title>. <source>PLoS One</source> <volume>16</volume>:<fpage>e0261295</fpage>. doi: <pub-id pub-id-type="doi">10.1371/JOURNAL.PONE.0261295</pub-id>, <pub-id pub-id-type="pmid">34898654</pub-id></mixed-citation></ref>
<ref id="ref61"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lau</surname><given-names>E. F.</given-names></name> <name><surname>Phillips</surname><given-names>C.</given-names></name> <name><surname>Poeppel</surname><given-names>D.</given-names></name></person-group> (<year>2008</year>). <article-title>A cortical network for semantics:(de) constructing the N400</article-title>. <source>Nat. Rev. Neurosci.</source> <volume>9</volume>, <fpage>920</fpage>&#x2013;<lpage>933</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nrn2532</pub-id>, <pub-id pub-id-type="pmid">19020511</pub-id></mixed-citation></ref>
<ref id="ref62"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mai</surname><given-names>G.</given-names></name> <name><surname>Wang</surname><given-names>W. S. Y.</given-names></name></person-group> (<year>2019</year>). <article-title>Delta and theta neural entrainment during phonological and semantic processing in speech perception</article-title>. <source>bioRxiv</source>:<fpage>556837</fpage>. doi: <pub-id pub-id-type="doi">10.1101/556837</pub-id></mixed-citation></ref>
<ref id="ref63"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Martin</surname><given-names>B. A.</given-names></name> <name><surname>Tremblay</surname><given-names>K. L.</given-names></name> <name><surname>Korczak</surname><given-names>P.</given-names></name></person-group> (<year>2008</year>). <article-title>Speech evoked potentials: from the laboratory to the clinic</article-title>. <source>Ear Hear.</source> <volume>29</volume>, <fpage>285</fpage>&#x2013;<lpage>313</lpage>. doi: <pub-id pub-id-type="doi">10.1097/AUD.0B013E3181662C0E</pub-id>, <pub-id pub-id-type="pmid">18453883</pub-id></mixed-citation></ref>
<ref id="ref64"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McGettigan</surname><given-names>C.</given-names></name> <name><surname>Agnew</surname><given-names>Z. K.</given-names></name> <name><surname>Scott</surname><given-names>S. K.</given-names></name></person-group> (<year>2010</year>). <article-title>Are articulatory commands automatically and involuntarily activated during speech perception?</article-title> <source>Proc. Natl. Acad. Sci.</source> <volume>107</volume>, <fpage>E42</fpage>&#x2013;<lpage>E42</lpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.1000186107</pub-id></mixed-citation></ref>
<ref id="ref65"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McMurray</surname><given-names>B.</given-names></name> <name><surname>Farris-Trimble</surname><given-names>A.</given-names></name> <name><surname>Rigler</surname><given-names>H.</given-names></name></person-group> (<year>2017</year>). <article-title>Waiting for lexical access: Cochlear implants or severely degraded input lead listeners to process speech less incrementally</article-title>. <source>Cognition</source> <volume>169</volume>, <fpage>147</fpage>&#x2013;<lpage>164</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.COGNITION.2017.08.013</pub-id>, <pub-id pub-id-type="pmid">28917133</pub-id></mixed-citation></ref>
<ref id="ref66"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McQueen</surname><given-names>J. M.</given-names></name> <name><surname>Norris</surname><given-names>D.</given-names></name> <name><surname>Cutler</surname><given-names>A.</given-names></name></person-group> (<year>1999</year>). <article-title>Lexical influence in phonetic decision making: evidence from subcategorical mismatches</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>25</volume>:<fpage>1363</fpage>. doi: <pub-id pub-id-type="doi">10.1037/0096-1523.25.5.1363</pub-id></mixed-citation></ref>
<ref id="ref67"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mehta</surname><given-names>A. H.</given-names></name> <name><surname>Oxenham</surname><given-names>A. J.</given-names></name></person-group> (<year>2017</year>). <article-title>Vocoder simulations explain complex pitch perception limitations experienced by cochlear implant users</article-title>. <source>J. Assoc. Res. Otolaryngol.</source> <volume>18</volume>, <fpage>789</fpage>&#x2013;<lpage>802</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10162-017-0632-x</pub-id>, <pub-id pub-id-type="pmid">28733803</pub-id></mixed-citation></ref>
<ref id="ref68"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Muncke</surname><given-names>J.</given-names></name> <name><surname>Kuruvila</surname><given-names>I.</given-names></name> <name><surname>Hoppe</surname><given-names>U.</given-names></name></person-group> (<year>2022</year>). <article-title>Prediction of speech intelligibility by means of EEG responses to sentences in noise</article-title>. <source>Front. Neurosci.</source> <volume>16</volume>:<fpage>876421</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNINS.2022.876421/BIBTEX</pub-id></mixed-citation></ref>
<ref id="ref69"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>N&#x00E4;&#x00E4;t&#x00E4;nen</surname><given-names>R.</given-names></name></person-group> (<year>2001</year>). <article-title>The perception of speech sounds by the human brain as reflected by the mismatch negativity (MMN) and its magnetic equivalent (MMNm)</article-title>. <source>Psychophysiology</source> <volume>38</volume>, <fpage>1</fpage>&#x2013;<lpage>21</lpage>. doi: <pub-id pub-id-type="doi">10.1111/1469-8986.3810001</pub-id>, <pub-id pub-id-type="pmid">11321610</pub-id></mixed-citation></ref>
<ref id="ref70"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Norris</surname><given-names>D.</given-names></name> <name><surname>McQueen</surname><given-names>J. M.</given-names></name> <name><surname>Cutler</surname><given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>Prediction, Bayesian inference and feedback in speech recognition</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>31</volume>, <fpage>4</fpage>&#x2013;<lpage>18</lpage>. doi: <pub-id pub-id-type="doi">10.1080/23273798.2015.1081703</pub-id>, <pub-id pub-id-type="pmid">26740960</pub-id></mixed-citation></ref>
<ref id="ref71"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>O&#x2019;Sullivan</surname><given-names>J. A.</given-names></name> <name><surname>Power</surname><given-names>A. J.</given-names></name> <name><surname>Mesgarani</surname><given-names>N.</given-names></name> <name><surname>Rajaram</surname><given-names>S.</given-names></name> <name><surname>Foxe</surname><given-names>J. J.</given-names></name> <name><surname>Shinn-Cunningham</surname><given-names>B. G.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Attentional selection in a cocktail party environment can be decoded from single-trial EEG</article-title>. <source>Cereb. Cortex</source> <volume>25</volume>, <fpage>1697</fpage>&#x2013;<lpage>1706</lpage>. doi: <pub-id pub-id-type="doi">10.1093/cercor/bht355</pub-id></mixed-citation></ref>
<ref id="ref72"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Obleser</surname><given-names>J.</given-names></name> <name><surname>Kotz</surname><given-names>S. A.</given-names></name></person-group> (<year>2010</year>). <article-title>Expectancy constraints in degraded speech modulate the language comprehension network</article-title>. <source>Cereb. Cortex</source> <volume>20</volume>, <fpage>633</fpage>&#x2013;<lpage>640</lpage>. doi: <pub-id pub-id-type="doi">10.1093/CERCOR/BHP128</pub-id>, <pub-id pub-id-type="pmid">19561061</pub-id></mixed-citation></ref>
<ref id="ref73"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Obleser</surname><given-names>J.</given-names></name> <name><surname>Scott</surname><given-names>S. K.</given-names></name> <name><surname>Eulitz</surname><given-names>C.</given-names></name></person-group> (<year>2006</year>). <article-title>Now you hear it, now you don&#x2019;t: transient traces of consonants and their nonspeech analogues in the human brain</article-title>. <source>Cereb. Cortex</source> <volume>16</volume>, <fpage>1069</fpage>&#x2013;<lpage>1076</lpage>. doi: <pub-id pub-id-type="doi">10.1093/cercor/bhj047</pub-id>, <pub-id pub-id-type="pmid">16207930</pub-id></mixed-citation></ref>
<ref id="ref74"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Obleser</surname><given-names>J.</given-names></name> <name><surname>Wise</surname><given-names>R. J. S.</given-names></name> <name><surname>Dresner</surname><given-names>M. A.</given-names></name> <name><surname>Scott</surname><given-names>S. K.</given-names></name></person-group> (<year>2007</year>). <article-title>Functional integration across brain regions improves speech perception under adverse listening conditions</article-title>. <source>J. Neurosci.</source> <volume>27</volume>, <fpage>2283</fpage>&#x2013;<lpage>2289</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.4663-06.2007</pub-id>, <pub-id pub-id-type="pmid">17329425</pub-id></mixed-citation></ref>
<ref id="ref75"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Okahara</surname><given-names>Y.</given-names></name> <name><surname>Aoyagi</surname><given-names>K.</given-names></name> <name><surname>Iwasa</surname><given-names>H.</given-names></name> <name><surname>Higuchi</surname><given-names>Y.</given-names></name></person-group> (<year>2024</year>). <article-title>Language lateralization by passive auditory fMRI in presurgical assessment for temporal lobe epilepsy: a single-center retrospective study</article-title>. <source>J. Clin. Med.</source> <volume>13</volume>, <fpage>1706</fpage>&#x2013;<lpage>1706</lpage>. doi: <pub-id pub-id-type="doi">10.3390/JCM13061706</pub-id></mixed-citation></ref>
<ref id="ref76"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peelle</surname><given-names>J. E.</given-names></name> <name><surname>Gross</surname><given-names>J.</given-names></name> <name><surname>Davis</surname><given-names>M. H.</given-names></name></person-group> (<year>2013</year>). <article-title>Phase-locked responses to speech in human auditory cortex are enhanced during comprehension</article-title>. <source>Cereb. Cortex</source> <volume>23</volume>, <fpage>1378</fpage>&#x2013;<lpage>1387</lpage>. doi: <pub-id pub-id-type="doi">10.1093/CERCOR/BHS118</pub-id>, <pub-id pub-id-type="pmid">22610394</pub-id></mixed-citation></ref>
<ref id="ref77"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peelle</surname><given-names>J. E.</given-names></name> <name><surname>Johnsrude</surname><given-names>I. S.</given-names></name> <name><surname>Davis</surname><given-names>M. H.</given-names></name></person-group> (<year>2010</year>). <article-title>Hierarchical processing for speech in human auditory cortex and beyond</article-title>. <source>Front. Hum. Neurosci.</source> <volume>4</volume>:<fpage>51</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNHUM.2010.00051/FULL</pub-id></mixed-citation></ref>
<ref id="ref78"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peelle</surname><given-names>J. E.</given-names></name> <name><surname>Wingfield</surname><given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>The neural consequences of age-related hearing loss</article-title>. <source>Trends Neurosci.</source> <volume>39</volume>, <fpage>486</fpage>&#x2013;<lpage>497</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.tins.2016.05.001</pub-id>, <pub-id pub-id-type="pmid">27262177</pub-id></mixed-citation></ref>
<ref id="ref79"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peelle</surname><given-names>J.</given-names></name> <name><surname>Wingfield</surname><given-names>A.</given-names></name></person-group> (<year>2022</year>). <article-title>How our brains make sense of noisy speech</article-title>. <source>Acoust. Today</source> <volume>18</volume>, <fpage>40</fpage>&#x2013;<lpage>48</lpage>. doi: <pub-id pub-id-type="doi">10.1121/AT.2022.18.3.40</pub-id></mixed-citation></ref>
<ref id="ref80"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pendyala</surname><given-names>V.</given-names></name> <name><surname>Sethares</surname><given-names>W.</given-names></name> <name><surname>Easwar</surname><given-names>V.</given-names></name></person-group> (<year>2024</year>). <article-title>Assessing speech audibility via syllabic-rate neural responses in adults and children with and without hearing loss</article-title>. <source>Trends Hear.</source> <volume>28</volume>:<fpage>23312165241227815</fpage>. doi: <pub-id pub-id-type="doi">10.1177/23312165241227815</pub-id>, <pub-id pub-id-type="pmid">38545698</pub-id></mixed-citation></ref>
<ref id="ref81"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Phillips</surname><given-names>C.</given-names></name> <name><surname>Pellathy</surname><given-names>T.</given-names></name> <name><surname>Marantz</surname><given-names>A.</given-names></name> <name><surname>Yellin</surname><given-names>E.</given-names></name> <name><surname>Wexler</surname><given-names>K.</given-names></name> <name><surname>Poeppel</surname><given-names>D.</given-names></name> <etal/></person-group>. (<year>2000</year>). <article-title>Auditory cortex accesses phonological categories: an MEG mismatch study</article-title>. <source>J. Cogn. Neurosci.</source> <volume>12</volume>, <fpage>1038</fpage>&#x2013;<lpage>1055</lpage>. doi: <pub-id pub-id-type="doi">10.1162/08989290051137567</pub-id>, <pub-id pub-id-type="pmid">11177423</pub-id></mixed-citation></ref>
<ref id="ref82"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pivik</surname><given-names>R. T.</given-names></name> <name><surname>Harman</surname><given-names>K.</given-names></name></person-group> (<year>1995</year>). <article-title>A reconceptualization of EEG alpha activity as an index of arousal during sleep: all alpha activity is not equal</article-title>. <source>J. Sleep Res.</source> <volume>4</volume>, <fpage>131</fpage>&#x2013;<lpage>137</lpage>. doi: <pub-id pub-id-type="doi">10.1111/J.1365-2869.1995.TB00161.X</pub-id>, <pub-id pub-id-type="pmid">10607151</pub-id></mixed-citation></ref>
<ref id="ref83"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Presacco</surname><given-names>A.</given-names></name> <name><surname>Simon</surname><given-names>J. Z.</given-names></name> <name><surname>Anderson</surname><given-names>S.</given-names></name></person-group> (<year>2016</year>). <article-title>Evidence of degraded representation of speech in noise, in the aging midbrain and cortex</article-title>. <source>J. Neurophysiol.</source> <volume>116</volume>, <fpage>2346</fpage>&#x2013;<lpage>2355</lpage>. doi: <pub-id pub-id-type="doi">10.1152/jn.00372.2016</pub-id>, <pub-id pub-id-type="pmid">27535374</pub-id></mixed-citation></ref>
<ref id="ref84"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Prince</surname><given-names>P.</given-names></name> <name><surname>Paul</surname><given-names>B. T.</given-names></name> <name><surname>Chen</surname><given-names>J.</given-names></name> <name><surname>Le</surname><given-names>T.</given-names></name> <name><surname>Lin</surname><given-names>V.</given-names></name> <name><surname>Dimitrijevic</surname><given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>Neural correlates of visual stimulus encoding and verbal working memory differ between cochlear implant users and normal-hearing controls</article-title>. <source>Eur. J. Neurosci.</source> <volume>54</volume>, <fpage>5016</fpage>&#x2013;<lpage>5037</lpage>. doi: <pub-id pub-id-type="doi">10.1111/EJN.15365</pub-id>, <pub-id pub-id-type="pmid">34146363</pub-id></mixed-citation></ref>
<ref id="ref85"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pulverm&#x00FC;ller</surname><given-names>F.</given-names></name> <name><surname>Shtyrov</surname><given-names>Y.</given-names></name> <name><surname>Hasting</surname><given-names>A. S.</given-names></name> <name><surname>Carlyon</surname><given-names>R. P.</given-names></name></person-group> (<year>2008</year>). <article-title>Syntax as a reflex: neurophysiological evidence for early automaticity of grammatical processing</article-title>. <source>Brain Lang.</source> <volume>104</volume>, <fpage>244</fpage>&#x2013;<lpage>253</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.BANDL.2007.05.002</pub-id>, <pub-id pub-id-type="pmid">17624417</pub-id></mixed-citation></ref>
<ref id="ref86"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rao</surname><given-names>R. P. N.</given-names></name> <name><surname>Ballard</surname><given-names>D. H.</given-names></name></person-group> (<year>1999</year>). <article-title>Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects</article-title>. <source>Nat. Neurosci.</source> <volume>2</volume>, <fpage>79</fpage>&#x2013;<lpage>87</lpage>. doi: <pub-id pub-id-type="doi">10.1038/4580</pub-id>, <pub-id pub-id-type="pmid">10195184</pub-id></mixed-citation></ref>
<ref id="ref87"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rigler</surname><given-names>H.</given-names></name> <name><surname>Farris-Trimble</surname><given-names>A.</given-names></name> <name><surname>Greiner</surname><given-names>L.</given-names></name> <name><surname>Walker</surname><given-names>J.</given-names></name> <name><surname>Tomblin</surname><given-names>J. B.</given-names></name> <name><surname>McMurray</surname><given-names>B.</given-names></name></person-group> (<year>2015</year>). <article-title>The slow developmental time course of real-time spoken word recognition</article-title>. <source>Dev. Psychol.</source> <volume>51</volume>:<fpage>1690</fpage>. doi: <pub-id pub-id-type="doi">10.1037/dev0000044</pub-id>, <pub-id pub-id-type="pmid">26479544</pub-id></mixed-citation></ref>
<ref id="ref88"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rosen</surname><given-names>S.</given-names></name> <name><surname>Souza</surname><given-names>P.</given-names></name> <name><surname>Ekelund</surname><given-names>C.</given-names></name> <name><surname>Majeed</surname><given-names>A. A.</given-names></name></person-group> (<year>2013</year>). <article-title>Listening to speech in a background of other talkers: effects of talker number and noise vocoding</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>133</volume>, <fpage>2431</fpage>&#x2013;<lpage>2443</lpage>. doi: <pub-id pub-id-type="doi">10.1121/1.4794379</pub-id>, <pub-id pub-id-type="pmid">23556608</pub-id></mixed-citation></ref>
<ref id="ref89"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sedley</surname><given-names>W.</given-names></name> <name><surname>Gander</surname><given-names>P. E.</given-names></name> <name><surname>Kumar</surname><given-names>S.</given-names></name> <name><surname>Kovach</surname><given-names>C. K.</given-names></name> <name><surname>Oya</surname><given-names>H.</given-names></name> <name><surname>Kawasaki</surname><given-names>H.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Neural signatures of perceptual inference</article-title>. <source>eLife</source> <volume>5</volume>:<fpage>e11476</fpage>. doi: <pub-id pub-id-type="doi">10.7554/ELIFE.11476</pub-id>, <pub-id pub-id-type="pmid">26949254</pub-id></mixed-citation></ref>
<ref id="ref90"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sekerina</surname><given-names>I. A.</given-names></name> <name><surname>Brooks</surname><given-names>P. J.</given-names></name></person-group> (<year>2007</year>). <article-title>Eye movements during spoken word recognition in Russian children</article-title>. <source>J. Exp. Child Psychol.</source> <volume>98</volume>, <fpage>20</fpage>&#x2013;<lpage>45</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jecp.2007.04.005</pub-id>, <pub-id pub-id-type="pmid">17560596</pub-id></mixed-citation></ref>
<ref id="ref91"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname><given-names>A.</given-names></name> <name><surname>Dorman</surname><given-names>M. F.</given-names></name> <name><surname>Spahr</surname><given-names>A. J.</given-names></name></person-group> (<year>2002a</year>). <article-title>A sensitive period for the development of the central auditory system in children with cochlear implants: implications for age of implantation</article-title>. <source>Ear Hear.</source> <volume>23</volume>, <fpage>532</fpage>&#x2013;<lpage>539</lpage>. doi: <pub-id pub-id-type="doi">10.1097/00003446-200212000-00004</pub-id>, <pub-id pub-id-type="pmid">12476090</pub-id></mixed-citation></ref>
<ref id="ref92"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname><given-names>A.</given-names></name> <name><surname>Dorman</surname><given-names>M. F.</given-names></name> <name><surname>Spahr</surname><given-names>A. J.</given-names></name></person-group> (<year>2002b</year>). <article-title>Rapid development of cortical auditory evoked potentials after early cochlear implantation</article-title>. <source>Neuroreport</source> <volume>13</volume>, <fpage>1365</fpage>&#x2013;<lpage>1368</lpage>. doi: <pub-id pub-id-type="doi">10.1097/00001756-200207190-00030</pub-id>, <pub-id pub-id-type="pmid">12151804</pub-id></mixed-citation></ref>
<ref id="ref93"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Slaats</surname><given-names>S.</given-names></name> <name><surname>Weissbart</surname><given-names>H.</given-names></name> <name><surname>Schoffelen</surname><given-names>J. M.</given-names></name> <name><surname>Meyer</surname><given-names>A. S.</given-names></name> <name><surname>Martin</surname><given-names>A. E.</given-names></name></person-group> (<year>2023</year>). <article-title>Delta-band neural responses to individual words are modulated by sentence processing</article-title>. <source>J. Neurosci.</source> <volume>43</volume>, <fpage>4867</fpage>&#x2013;<lpage>4883</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0964-22.2023</pub-id>, <pub-id pub-id-type="pmid">37221093</pub-id></mixed-citation></ref>
<ref id="ref94"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sohoglu</surname><given-names>E.</given-names></name> <name><surname>Davis</surname><given-names>M. H.</given-names></name></person-group> (<year>2016</year>). <article-title>Perceptual learning of degraded speech by minimizing prediction error</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>113</volume>, <fpage>E1747</fpage>&#x2013;<lpage>E1756</lpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.1523266113</pub-id>, <pub-id pub-id-type="pmid">26957596</pub-id></mixed-citation></ref>
<ref id="ref95"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Steinmetzger</surname><given-names>K.</given-names></name> <name><surname>Rosen</surname><given-names>S.</given-names></name></person-group> (<year>2017</year>). <article-title>Effects of acoustic periodicity and intelligibility on the neural oscillations in response to speech</article-title>. <source>Neuropsychologia</source> <volume>95</volume>, <fpage>173</fpage>&#x2013;<lpage>181</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuropsychologia.2016.12.003</pub-id>, <pub-id pub-id-type="pmid">27939190</pub-id></mixed-citation></ref>
<ref id="ref96"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Swingley</surname><given-names>D.</given-names></name> <name><surname>Pinto</surname><given-names>J. P.</given-names></name> <name><surname>Fernald</surname><given-names>A.</given-names></name></person-group> (<year>1999</year>). <article-title>Continuous processing in word recognition at 24 months</article-title>. <source>Cognition</source> <volume>71</volume>, <fpage>73</fpage>&#x2013;<lpage>108</lpage>. doi: <pub-id pub-id-type="doi">10.1016/s0010-0277(99)00021-9</pub-id>, <pub-id pub-id-type="pmid">10444905</pub-id></mixed-citation></ref>
<ref id="ref97"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Synigal</surname><given-names>S. R.</given-names></name> <name><surname>Teoh</surname><given-names>E. S.</given-names></name> <name><surname>Lalor</surname><given-names>E. C.</given-names></name></person-group> (<year>2020</year>). <article-title>Including measures of high gamma power can improve the decoding of natural speech from EEG</article-title>. <source>Front. Hum. Neurosci.</source> <volume>14</volume>:<fpage>130</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnhum.2020.00130</pub-id>, <pub-id pub-id-type="pmid">32410969</pub-id></mixed-citation></ref>
<ref id="ref98"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Teghipco</surname><given-names>A.</given-names></name> <name><surname>Okada</surname><given-names>K.</given-names></name> <name><surname>Murphy</surname><given-names>E.</given-names></name> <name><surname>Hickok</surname><given-names>G.</given-names></name></person-group> (<year>2023</year>). <article-title>Predictive coding and internal error correction in speech production</article-title>. <source>Neurobiol. Lang.</source> <volume>4</volume>, <fpage>81</fpage>&#x2013;<lpage>119</lpage>. doi: <pub-id pub-id-type="doi">10.1162/NOL_A_00088</pub-id></mixed-citation></ref>
<ref id="ref99"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thompson-Schill</surname><given-names>S. L.</given-names></name> <name><surname>D&#x2019;Esposito</surname><given-names>M.</given-names></name> <name><surname>Aguirre</surname><given-names>G. K.</given-names></name> <name><surname>Farah</surname><given-names>M. J.</given-names></name></person-group> (<year>1997</year>). <article-title>Role of left inferior prefrontal cortex in retrieval of semantic knowledge: a&#x202F;reevaluation</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>94</volume>, <fpage>14792</fpage>&#x2013;<lpage>14797</lpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.94.26.14792</pub-id>, <pub-id pub-id-type="pmid">9405692</pub-id></mixed-citation></ref>
<ref id="ref100"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Todorovic</surname><given-names>A.</given-names></name> <name><surname>van Ede</surname><given-names>F.</given-names></name> <name><surname>Maris</surname><given-names>E.</given-names></name> <name><surname>de Lange</surname><given-names>F. P.</given-names></name></person-group> (<year>2011</year>). <article-title>Prior expectation mediates neural adaptation to repeated sounds in the auditory cortex: an MEG study</article-title>. <source>J. Neurosci.</source> <volume>31</volume>, <fpage>9118</fpage>&#x2013;<lpage>9123</lpage>. doi: <pub-id pub-id-type="doi">10.1523/JNEUROSCI.1425-11.2011</pub-id>, <pub-id pub-id-type="pmid">21697363</pub-id></mixed-citation></ref>
<ref id="ref101"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Toffolo</surname><given-names>K. K.</given-names></name> <name><surname>Freedman</surname><given-names>E. G.</given-names></name> <name><surname>Foxe</surname><given-names>J. J.</given-names></name></person-group> (<year>2022</year>). <article-title>Evoking the N400 event-related potential (ERP) component using a publicly available novel set of sentences with semantically incongruent or congruent eggplants (endings)</article-title>. <source>Neuroscience</source> <volume>501</volume>, <fpage>143</fpage>&#x2013;<lpage>158</lpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROSCIENCE.2022.07.030</pub-id></mixed-citation></ref>
<ref id="ref102"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tremblay</surname><given-names>K. L.</given-names></name> <name><surname>Friesen</surname><given-names>L.</given-names></name> <name><surname>Martin</surname><given-names>B. A.</given-names></name> <name><surname>Wright</surname><given-names>R.</given-names></name></person-group> (<year>2003</year>). <article-title>Test-retest reliability of cortical evoked potentials using naturally produced speech sounds</article-title>. <source>Ear Hear.</source> <volume>24</volume>, <fpage>225</fpage>&#x2013;<lpage>232</lpage>. doi: <pub-id pub-id-type="doi">10.1097/01.aud.0000069229.84883.03</pub-id>, <pub-id pub-id-type="pmid">12799544</pub-id></mixed-citation></ref>
<ref id="ref103"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vanthornhout</surname><given-names>J.</given-names></name> <name><surname>Decruy</surname><given-names>L.</given-names></name> <name><surname>Francart</surname><given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Effect of task and attention on neural tracking of speech</article-title>. <source>Front. Neurosci.</source> <volume>13</volume>:<fpage>977</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FNINS.2019.00977/BIBTEX</pub-id></mixed-citation></ref>
<ref id="ref104"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wacongne</surname><given-names>C.</given-names></name> <name><surname>Labyt</surname><given-names>E.</given-names></name> <name><surname>Van Wassenhove</surname><given-names>V.</given-names></name> <name><surname>Bekinschtein</surname><given-names>T.</given-names></name> <name><surname>Naccache</surname><given-names>L.</given-names></name> <name><surname>Dehaene</surname><given-names>S.</given-names></name></person-group> (<year>2011</year>). <article-title>Evidence for a hierarchy of predictions and prediction errors in human cortex</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>108</volume>, <fpage>20754</fpage>&#x2013;<lpage>20759</lpage>. doi: <pub-id pub-id-type="doi">10.1073/PNAS.1117807108</pub-id>, <pub-id pub-id-type="pmid">22147913</pub-id></mixed-citation></ref>
<ref id="ref105"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Weissbart</surname><given-names>H.</given-names></name> <name><surname>Kandylaki</surname><given-names>K. D.</given-names></name> <name><surname>Reichenbach</surname><given-names>T.</given-names></name></person-group> (<year>2020</year>). <article-title>Cortical tracking of surprisal during continuous speech comprehension</article-title>. <source>J. Cogn. Neurosci.</source> <volume>32</volume>, <fpage>155</fpage>&#x2013;<lpage>166</lpage>. doi: <pub-id pub-id-type="doi">10.1162/jocn_a_01467</pub-id>, <pub-id pub-id-type="pmid">31479349</pub-id></mixed-citation></ref>
<ref id="ref106"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wilson</surname><given-names>B. S.</given-names></name> <name><surname>Dorman</surname><given-names>M. F.</given-names></name></person-group> (<year>2008</year>). <article-title>Cochlear implants: a remarkable past and a brilliant future</article-title>. <source>Hear. Res.</source> <volume>242</volume>, <fpage>3</fpage>&#x2013;<lpage>21</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.heares.2008.06.005</pub-id>, <pub-id pub-id-type="pmid">18616994</pub-id></mixed-citation></ref>
<ref id="ref107"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wilson</surname><given-names>B. S.</given-names></name> <name><surname>Finley</surname><given-names>C. C.</given-names></name> <name><surname>Lawson</surname><given-names>D. T.</given-names></name> <name><surname>Wolford</surname><given-names>R. D.</given-names></name> <name><surname>Eddington</surname><given-names>D. K.</given-names></name> <name><surname>Rabinowitz</surname><given-names>W. M.</given-names></name></person-group> (<year>1991</year>). <article-title>Better speech recognition with cochlear implants</article-title>. <source>Nature</source> <volume>352</volume>, <fpage>236</fpage>&#x2013;<lpage>238</lpage>. doi: <pub-id pub-id-type="doi">10.1038/352236A0;KWRD=SCIENCE</pub-id></mixed-citation></ref>
<ref id="ref108"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>T.</given-names></name> <name><surname>Kurkela</surname><given-names>J. L. O.</given-names></name> <name><surname>Chen</surname><given-names>K.</given-names></name> <name><surname>Liu</surname><given-names>Y.</given-names></name> <name><surname>Shu</surname><given-names>H.</given-names></name> <name><surname>Cong</surname><given-names>F.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Native language advantage in electrical brain responses to speech sound changes in passive and active listening condition</article-title>. <source>Neuropsychologia</source> <volume>201</volume>:<fpage>108936</fpage>. doi: <pub-id pub-id-type="doi">10.1016/J.NEUROPSYCHOLOGIA.2024.108936</pub-id>, <pub-id pub-id-type="pmid">38851314</pub-id></mixed-citation></ref>
<ref id="ref109"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yip</surname><given-names>M. C. W.</given-names></name> <name><surname>Blumenfeld</surname><given-names>H. K.</given-names></name> <name><surname>Cie&#x015B;licka</surname><given-names>A. B.</given-names></name></person-group> (<year>2021</year>). <article-title>Bilingual and multilingual spoken-word recognition: empirical and theoretical perspectives</article-title>. <source>Front. Psychol.</source> <volume>12</volume>:<fpage>696354</fpage>. doi: <pub-id pub-id-type="doi">10.3389/FPSYG.2021.696354</pub-id>, <pub-id pub-id-type="pmid">34149579</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0002">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2708/overview">Claude Alain</ext-link>, Rotman Research Institute (RRI), Canada</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0003">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1625877/overview">Jesyin Lai</ext-link>, St. Jude Children&#x2019;s Research Hospital, United States</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/820821/overview">Mareike Daeglau</ext-link>, University of Oldenburg, Germany</p>
</fn>
</fn-group>
</back>
</article>