<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Integr. Neurosci.</journal-id>
<journal-title>Frontiers in Integrative Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Integr. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5145</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnint.2012.00064</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research Article</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Coding of multisensory temporal patterns in human superior temporal sulcus</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Noesselt</surname> <given-names>T&#x000F6;mme</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x0002A;</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Bergmann</surname> <given-names>Daniel</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x02020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Heinze</surname> <given-names>Hans-Jochen</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>M&#x000FC;nte</surname> <given-names>Thomas</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Spence</surname> <given-names>Charles</given-names></name>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Biological Psychology, Otto-von-Guericke-Universit&#x000E4;t Magdeburg</institution> <country>Magdeburg, Germany</country></aff>
<aff id="aff2"><sup>2</sup><institution>Center of Behavioral Brain Sciences, Otto-von-Guericke-Universit&#x000E4;t Magdeburg</institution> <country>Magdeburg, Germany</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Neurology, Otto-von-Guericke-Universit&#x000E4;t Magdeburg</institution> <country>Magdeburg, Germany</country></aff>
<aff id="aff4"><sup>4</sup><institution>Psychosomatic Medicine, Asklepios Westklinikum Hamburg</institution> <country>Hamburg, Germany</country></aff>
<aff id="aff5"><sup>5</sup><institution>Department of Neurology, Universit&#x000E4;tzuL&#x000FC;beck</institution> <country>L&#x000FC;beck, Germany</country></aff>
<aff id="aff6"><sup>6</sup><institution>Crossmodal Research Laboratory, Department of Experimental Psychology, University of Oxford</institution> <country>Oxford, UK</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Zhuanghua Shi, Ludwig-Maximilians-Universit&#x000E4;t M&#x000FC;nchen, Germany</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Emiliano Macaluso, Fondazione Santa Lucia, Italy; Mark W. Greenlee, University of Regensburg, Germany</p></fn>
<fn fn-type="corresp" id="fn001"><p>&#x0002A;Correspondence: T&#x000F6;mme Noesselt, Department of Biological Psychology, Otto-von-Guericke-Universit&#x000E4;t Magdeburg, Universit&#x000E4;tsplatz 2, 39106 Magdeburg, Germany. e-mail: <email>toemme&#x00040;med.ovgu.de</email></p></fn>
<fn fn-type="present-address" id="fn002"><p>&#x02020;These authors equally contributed to this work.</p></fn>
</author-notes>
<pub-date pub-type="epreprint">
<day>20</day>
<month>06</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="epub">
<day>28</day>
<month>08</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="collection">
<year>2012</year>
</pub-date>
<volume>6</volume>
<elocation-id>64</elocation-id>
<history>
<date date-type="received">
<day>06</day>
<month>06</month>
<year>2012</year>
</date>
<date date-type="accepted">
<day>07</day>
<month>08</month>
<year>2012</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2012 Noesselt, Bergmann, Heinze, M&#x000FC;nte and Spence.</copyright-statement>
<copyright-year>2012</copyright-year>
<license license-type="open-access" xlink:href="http://www.frontiersin.org/licenseagreement"><p>This is an open-access article distributed under the terms of the <uri xlink:href="http://creativecommons.org/licenses/by/3.0/">Creative Commons Attribution License</uri>, which permits use, distribution and reproduction in other forums, provided the original authors and source are credited and subject to any copyright notices concerning any third-party graphics etc.</p>
</license>
</permissions>
<abstract><p>Philosophers, psychologists, and neuroscientists have long been interested in how the temporal aspects of perception are represented in the brain. In the present study, we investigated the neural basis of the temporal perception of synchrony/asynchrony for audiovisual speech stimuli using functional magnetic resonance imaging (fMRI). Subjects judged the temporal relation of (a)synchronous audiovisual speech streams, and indicated any changes in their perception of the stimuli over time. Differential hemodynamic responses for synchronous versus asynchronous stimuli were observed in the multisensory superior temporal sulcus complex (mSTS-c) and prefrontal cortex. Within mSTS-c we found adjacent regions expressing an enhanced BOLD-response to the different physical (a)synchrony conditions. These regions were further modulated by the subjects&#x00027; perceptual state. By calculating the distances between the modulated regions within mSTS-c in single-subjects we demonstrate that the &#x0201C;auditory leading (A<sub>L</sub>)&#x0201D; and &#x0201C;visual leading (V<sub>L</sub>) areas&#x0201D; lie closer to &#x0201C;synchrony areas&#x0201D; than to each other. Moreover, analysis of interregional connectivity indicates a stronger functional connection between multisensory prefrontal cortex and mSTS-c during the perception of asynchrony. Taken together, these results therefore suggest the presence of distinct sub-regions within the human STS-c for the maintenance of temporal relations for audiovisual speech stimuli plus differential functional connectivity with prefrontal regions. The respective local activity in mSTS-c is dependent both upon the physical properties of the stimuli presented and upon the subjects&#x00027; perception of (a)synchrony.</p></abstract>
<kwd-group>
<kwd>audiovisual</kwd>
<kwd>temporal perception</kwd>
<kwd>fMRI</kwd>
<kwd>speech</kwd>
<kwd>human</kwd>
</kwd-group>
<counts>
<fig-count count="6"/>
<table-count count="4"/>
<equation-count count="0"/>
<ref-count count="52"/>
<page-count count="14"/>
<word-count count="8843"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="introduction" id="s1">
<title>Introduction</title>
<p>When observers are confronted with incongruent auditory and visual information, that information is often fused into a congruent multisensory percept. Spatial, semantic, and temporal factors have all been shown to contribute to this perceptual fusion (see e.g., Driver and Noesselt, <xref ref-type="bibr" rid="B19">2008</xref>, for a review). The temporal relationship between inputs from different senses plays a particularly important role in multisensory integration (K&#x000F6;hler, <xref ref-type="bibr" rid="B29">1947</xref>; Dennett, <xref ref-type="bibr" rid="B14">1991</xref>; Spence and Squire, <xref ref-type="bibr" rid="B41">2003</xref>; Kelly, <xref ref-type="bibr" rid="B27">2005</xref>) and the perceived synchrony declines when the audio-visual asynchrony exceeds a certain temporal delay. When simple auditory beeps and visual flashes are being judged, subjects&#x00027; temporal synchrony window spans approximately 100 ms (Slutsky and Recanzone, <xref ref-type="bibr" rid="B40">2001</xref>; Vatakis and Spence, <xref ref-type="bibr" rid="B48">2006a</xref>) becoming broader/wider when stimuli are more complex (consisting of semantic content; Dixon and Spitz, <xref ref-type="bibr" rid="B17">1980</xref>; McGrath and Summerfield, <xref ref-type="bibr" rid="B33">1985</xref>; Spence and Squire, <xref ref-type="bibr" rid="B41">2003</xref>; Miller and D&#x00027;Esposito, <xref ref-type="bibr" rid="B34">2005</xref>; Vatakis and Spence, <xref ref-type="bibr" rid="B49">2006b</xref>, see also Vroomen and Keetels, <xref ref-type="bibr" rid="B50">2010</xref> for review).</p>
<p>Several brain structures have been implicated in the multisensory integration of auditory and visual stimuli. Among them are the superior colliculi (Stein and Meredith, <xref ref-type="bibr" rid="B42">1993</xref>), the superior temporal sulcus complex (STS-c), the intraparietal sulcus (IPS), the insular cortex, the claustrum and prefrontal areas (e.g., Calvert et al., <xref ref-type="bibr" rid="B11">2000</xref>; Bushara et al., <xref ref-type="bibr" rid="B8">2001</xref>; Calvert, <xref ref-type="bibr" rid="B9">2001</xref>; Driver and Noesselt, <xref ref-type="bibr" rid="B19">2008</xref>). Within the STS-c, areas within or close to the upper bank have been identified as key regions governing multisensory integration in both humans (Wright et al., <xref ref-type="bibr" rid="B51">2003</xref>; Beauchamp, <xref ref-type="bibr" rid="B2">2005a</xref>; Noesselt et al., <xref ref-type="bibr" rid="B35">2007</xref>) and non-human primates (Benevento et al., <xref ref-type="bibr" rid="B6">1977</xref>; Desimone and Gross, <xref ref-type="bibr" rid="B15">1979</xref>; Bruce et al., <xref ref-type="bibr" rid="B7">1981</xref>; Hikosaka et al., <xref ref-type="bibr" rid="B25">1988</xref>; Barraclough et al., <xref ref-type="bibr" rid="B1">2005</xref>). Direct neuronal recordings from the superior temporal polysensory (STP) region in monkeys have revealed that neurons can respond to both visual and auditory stimuli in both the upper (Bruce et al., <xref ref-type="bibr" rid="B7">1981</xref>; Hikosaka et al., <xref ref-type="bibr" rid="B25">1988</xref>) and lower banks (Benevento et al., <xref ref-type="bibr" rid="B6">1977</xref>). Barraclough et al. (<xref ref-type="bibr" rid="B1">2005</xref>) reported neurons within the STS-c that respond to action-related congruent audiovisual stimulation. When focusing on complex, speech-related animal communication, results from studies in macaques suggest that temporal regions in the macaque brain (including in the STS-c) are activated by audiovisual species-specific vocalizations (Gil-da-Costa et al., <xref ref-type="bibr" rid="B23">2004</xref>; Ghazanfar et al., <xref ref-type="bibr" rid="B22">2008</xref>). In humans, using linguistic stimuli, van Atteveldt et al. (<xref ref-type="bibr" rid="B46">2004</xref>) found regions in the STS-c that responded to visually presented letters, spoken single letters, or their combination. As in the study by Wright et al. (<xref ref-type="bibr" rid="B51">2003</xref>) employing lip-movements plus spoken syllables, the STS-c response was greatest for audiovisual stimuli. van Atteveldt and colleagues (<xref ref-type="bibr" rid="B46">2004</xref>) reported that multisensory enhancement was seen for congruent but not for incongruent stimuli. However, other studies reported enhancements in functional magnetic resonance imaging (fMRI)-responses for incongruent stimuli within STS-c (e.g., van Atteveldt et al., <xref ref-type="bibr" rid="B47">2007</xref>). These findings suggest that the STS-c is involved in the temporal binding of audiovisual stimuli. However, it still needs to be established whether congruent or incongruent audiovisual stimuli elicit a higher fMRI-signal in STS-c, or whether different subregions within the STS-c may differentially code multisensory temporal relations.</p>
<p>Hence, the aim of the present study was to investigate the functional neuroanatomy of the multisensory regions including STS-c and prefrontal cortex when perceiving a temporal (mis-)alignment of ecologically-valid long speech sequences; and to examine whether audiovisual temporal relationships may subdivide multisensory regions functionally. Subjects were shown videos of temporally aligned and misaligned video streams [either auditory leading (A<sub>L</sub>) or visual leading (V<sub>L</sub>) and reported whether those were perceived as being synchronous or asynchronous. Importantly, they also reported changes of perceived timing <italic>during</italic> the presentation of each stimulus. This design enabled us to dissociate those neural processes that were related to perceptual switches and those related to stable perceptual states during the presentation of audiovisual speech sequences. To anticipate, we found differential BOLD-effects for the different temporal percepts (A<sub>L</sub>, V<sub>L</sub>, and synchrony (AV<sub>S</sub>)] within adjacent subregions in human STS-c, plus differential interregional connectivity with prefrontal cortex.</p>
</sec>
<sec sec-type="methods" id="s2">
<title>Methods</title>
<p>A temporal-threshold experiment was conducted prior to scanning, to account for any individual differences in temporal perception. By choosing bistable stimuli for each subject we maximized the number of trials per condition during the fMRI-experiment (see below). Subjects (<italic>n</italic> &#x0003D; 14, 7 female) were placed in a dark, sound-attenuated chamber after providing written informed consent in accord with local ethics. They had to report the perceived synchrony or direction of asynchrony of auditory and visual information of video sequences by pressing one of three buttons (thereby indicating A<sub>L</sub>, AV<sub>S</sub>, V<sub>L</sub>). Importantly, subjects could change their judgements <italic>during</italic> each video presentation. The stimuli consisted of 20 video clips (length 23.7 s), depicting the face of a trained female speaker reading sentences (see Figure <xref ref-type="fig" rid="F1">1</xref>). Stimuli were randomized with MATLAB 6.1 and presented using Presentation 9.11 (Neurobehavioral Systems, Inc., CA). Initially, 20 synchronous sequences plus 80 temporally shifted sequences were presented (&#x02212;130 ms, &#x02212;60 ms (A<sub>L</sub>) and 200 ms/400 ms (V<sub>L</sub>), 20 video clips each, see Figure <xref ref-type="fig" rid="F2">2A</xref>). These asynchronies for threshold-determination were chosen in accord with previous reports (Dixon and Spitz, <xref ref-type="bibr" rid="B17">1980</xref>). For the fMRI-experiment, those stimuli were chosen for each subject that had a similar number of synchrony and asynchrony judgments (called near-threshold below).</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>Overlap of visual and auditory BOLD-modulations for unisensory stimulus presentations (<italic>p</italic> &#x0003C; 0.005; <italic>k</italic> &#x0003E; 10).</bold> This activation map was used as the search volume for the fMRI-analysis in the main experiment.</p></caption>
<graphic xlink:href="fnint-06-00064-g0001.tif"/>
</fig>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p><bold>Experimental design and behavioral results. (A)</bold> Depicts an example of a video-clip presented in three conditions [i.e., auditory leading (top left, temporal lag from 60&#x02013;120 ms), auditory and video synchronous (top middle), or visual leading (top right, temporal lag from 200&#x02013;400 ms)]. Auditory and visual lags were determined in a preliminary threshold-determination-experiment. Stimuli were presented at 1&#x000B0; visual angle above fixation (lower boundary) up to 7&#x000B0; (upper boundary). The duration of all 20 video-clips was 23.7 s, the interstimulus interval was 20 s. Partcipants indicated whether they perceived the auditory stream leading, the visual stream leading, or the 2 streams as being synchronous. They were encouraged to report any changes in their perception during the presentation of each video. Note that the physical lag was fixed within each video clip near the individual&#x00027;s synchrony/asynchrony-threshold. <bold>(B)</bold> Radar graph depictsmean durations (time from one keypress to the next) of subjects&#x00027; (a)synchrony-percepts for each experimental condition during fMRI-scanning: perceptual states were longest when perception of (a)synchrony was congruent with physical stimulation. Therefore, in the fMRI-analysis, hemodynamic response functions (HRF) could be specifically modeled and extracted for each stable percept and perceptual switches using a mixed model (see below). <bold>(C)</bold> An example trial modeled with hemodynamic response functions for an auditory leading-stimulus (A<sub>L</sub>). Gamma-curves depict perceptual switches/decisions, whereas box-car functions illustrate the sensory procesing prior to the first decision and perceptual states. Purple curves stand for AV<sub>S</sub>, red for A<sub>L</sub>. Note that each box-car function was individually specified based on the trial-by-trial inter-button-press duration.</p></caption>
<graphic xlink:href="fnint-06-00064-g0002.tif"/>
</fig>
<sec>
<title>fMRI-data acquisition</title>
<p>fMRI-data was acquired on a whole body Siemens 3 T Trio-scanner (Siemens, Erlangen, Germany) using a circular-polarized whole-head coil (BrukerBioSpin, Ettlingen, Germany). Subjects performed the same task as they had outside the scanner, reporting their responses with their right index, middle, and ring finger. Within the scanner subjects were presented three conditions: near-threshold V<sub>L</sub>, near-threshold A<sub>L</sub> plus the AV<sub>S</sub> condition. All other stimulus parameters were kept as in the behavioral experiment outside the scanner except for the following: first, a baseline period of 20 s was introduced after each video clip. Second, eye movements were monitored using an fMRI-compatible infrared recording system (Kanowski et al., <xref ref-type="bibr" rid="B26">2007</xref>) plus evaluation software (PupilTracker, HumanScan, Erlangen, Germany). The eye movement data was analysed with MATLAB 6.5. Third, before the main fMRI-experiment, a functional localizer was run in which only unimodal auditory or unimodal visual stimuli from the videos were presented (331 volumes covering the whole head, TR 2 s, TE 30 ms, flip 80&#x000B0;, resolution 64 &#x000D7; 64 &#x000D7; 32 at 3.5 &#x000D7; 3.5 &#x000D7;4 mm). The derived overlapping audio-visual activation map was then used to identify candidate multisensory areas (see below). Fourth, subjects wore earplugs; perceived loudness and balance were adapted individually to ensure easy comprehension of the auditory speech sequences despite the scanner noise. The stimuli were presented using MR-compatible, electrodynamic headphones (MRconfon, Magdeburg, Germany).</p>
<p>During the main experiment functional volumes were collected in four sessions (331 volumes each, covering the whole head, TR 2 s, TE 30 ms, flip 80&#x000B0;, resolution 64 &#x000D7; 64 &#x000D7; 32 at 3.5 &#x000D7; 3.5 &#x000D7;4 mm). Additionally, for anatomical localization an inversion-recovery EPI was acquired (TR 2 s, TE 30 ms, TI:1450 ms, resolution 64 &#x000D7; 64 &#x000D7; 32 at 3.5 &#x000D7; 3.5 &#x000D7;4 mm, same slice orientation and distortions as the functional volumes). The first five volumes from each session were excluded from further analysis. The remaining volumes were acquisition-corrected to the first acquired slice of each volume, motion-corrected, normalized at 2 mm<sup>3</sup> voxel size and smoothed (6 mm), using SPM2 (Wellcome Department of Cognitive Neurology, London, UK).</p>
</sec>
<sec>
<title>Group-level statistics</title>
<p>After pre-processing the data from a localizer run were modeled with two box-car functions convolved with the hemodynamic response function (HRF) for the auditory and visual trials. For the localizer runs, blocks were compared to the baseline during which no stimulus was present (<italic>p</italic> &#x0003C; 0.005; <italic>k</italic> &#x0003E; 10). An audiovisual mask (i.e., overlap of unisensory visual and auditory activations) was computed to identify candidate multisensory structures (see Figure <xref ref-type="fig" rid="F1">1</xref>; cf. Beauchamp et al., <xref ref-type="bibr" rid="B5">2004b</xref>; Beauchamp, <xref ref-type="bibr" rid="B3">2005b</xref>; Noesselt et al., <xref ref-type="bibr" rid="B35">2007</xref>; Szycik et al., <xref ref-type="bibr" rid="B44">2008</xref>).</p>
<p>Next, all experimental conditions were modeled with the HRF with variable durations when appropriate (mixed model; see Figure <xref ref-type="fig" rid="F2">2C</xref>). In particular, 21 conditions were defined in a mixed model: three perceptual switches (subjects&#x00027; button press, event-related), three perceptual states (time after button press, variable block) and the initial stimulation (time before the first button press, variable block) for every stimulus condition (A<sub>L</sub>, V<sub>L</sub>, and AV<sub>S</sub>). To test condition effects, linear contrasts were used for each subject and condition and masked inclusively with the audiovisual overlap from the functional localizer. The resulting contrast images were applied to perform random effects second-level analyses. The statistical parametric maps of the <italic>t</italic>-statistics at each voxel were thresholded at <italic>p</italic> &#x0003C; 0.05 (small-volume-corrected) and the spatial extent threshold was set at <italic>k</italic> &#x0003E; 5 voxels.</p>
<p>The following contrasts were computed: First, we identified regions that responded to physical synchrony and asynchronous conditions. Second, we identified regions that showed differential fMRI-signals for perceived synchrony vs. asynchrony conditions. Finally, we computed interaction effects for differential perceptual states with identical physical stimulation (i.e., asynchronous vs. synchronous percepts separately for A<sub>L</sub>, V<sub>L</sub>, and AV<sub>S</sub> stimulation).</p>
</sec>
<sec>
<title>Single-subject statistics</title>
<p>We also analysed the data from individual subjects in order to confirm our group-level results and to test the interaction between stimulation and percepts formally. We identified for each subject regions within STS-c using the identical contrasts as in the group analysis above: for A<sub>L</sub> stimulation: veridical A<sub>L</sub> percept &#x0003E; non-veridical synchronous percept; for AV<sub>S</sub> stimulation: veridical synchrony percept &#x0003E; both non-veridical percepts; for V<sub>L</sub> stimulation: veridical V<sub>L</sub> percept &#x0003E; non-veridical synchronous percept. Subject-specific regions of interest (ROI) were identified by searching for significant clusters of the three contrasts of interest within subregions of the STS-c (anatomical criterion) which expressed unisensory responses to both modalities (additional functional criterion). We extracted the beta-weights of all experimental conditions from these three local maxima for each subject and tested whether these local maxima would express significantly different results across stimulations. Note that this analysis is non-trivial and provides additional information, since any BOLD-modulation of different perceptual states to the AV<sub>S</sub>-stimulation was left unspecified in the A<sub>L</sub> and V<sub>L</sub> stimulation contrasts and vice versa.</p>
</sec>
<sec>
<title>Analysis of interregional connectivity</title>
<p>Complementary to the analysis of local modulations of the BOLD-response we investigated the effects of interregional connectivity in the context of perception of AV<sub>S</sub>, V<sub>L</sub>, and A<sub>L</sub> as described above (Friston et al., <xref ref-type="bibr" rid="B21">1997</xref>). We seeded our analyses in the subject-specific local maxima in STS-c and analyzed which other regions showed enhanced functional coupling in the context of A<sub>L</sub> percepts in the A<sub>L</sub> condition (relative to non-veridical synchronous percept in the A<sub>L</sub> condition), in the context of V<sub>L</sub> percepts in the V<sub>L</sub> condition (relative to the non-veridical synchronous percept in the A<sub>L</sub> condition) and in the context of synchronous percepts in the synchronous condition (relative to the non-veridical asynchronous percept in the synchronous condition) using a model with 21 regressors (see above) plus the physiological response and the psychophysiological interaction as two additional regressors (see e.g., Noesselt et al., <xref ref-type="bibr" rid="B35">2007</xref> for a similar approach) to reveal differential functional interregional connections in the psychological context of synchronous or asynchronous percepts. Three models were calculated separately for each STS-local maximum (corresponding to veridical A<sub>L</sub>-percepts, veridical V<sub>L</sub>-percepts, and veridical AV<sub>S</sub> percepts).</p>
<p>Differential group-level effects were calculated with an analysis of variance (ANOVA) pertaining the three PPIs from the three connectivity analysis.</p>
</sec>
<sec>
<title>Analysis of consistent patterning of subregions</title>
<p>Finally, distances between single subject maxima in STS-c were computed and analysed to reveal any systematic anatomical distribution of subjects&#x00027; local maxima for the A<sub>L</sub>, V<sub>L</sub>, and AV<sub>S</sub> representation. For this we used a three step approach: normalization of MNI-coordinates, calculation of distances by subtracting the normalized MNI-coordinates and calculation of Euclidian distances in three-dimensional space. In particular, for the calculation of distances, the MNI coordinates (in millimeters) of the three contrasts and their respective local maxima were scaled by adding the maximum negative value to all coordinates of one dimension so that all values were positive. This procedure was applied for the <italic>y</italic> and <italic>z</italic> extension/dimension; <italic>x</italic> coordinates were converted into their absolute value. Second, coordinate values of the same dimension but different local maxima were subtracted from each other (A<sub>L</sub>/V<sub>L</sub> minus synchrony and A<sub>L</sub> minus V<sub>L</sub>). Finally, we computed Euclidean distances for the difference measures: following Pythagoras&#x00027; Theorem, difference values of the <italic>x</italic> and <italic>y</italic> dimension (cathetuses) were squared and added together and the resulting value (hypotenuse) added to the squared <italic>z</italic> dimension difference. The square roots of the resulting values (again hypotenuse) represent the reported distances between voxels.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec>
<title>Behavioral results</title>
<p>The results of the behavioral experiment outside the scanner revealed that subjects&#x00027; judgments became more consistent with stimulation as the audiovisual delay increased. For the auditory stream leading condition, the mean delay for near-threshold stimuli was 105 ms (&#x000B1;35 ms) while for the visual stream leading condition it was 227 ms (&#x000B1;47 ms). Inside the scanner, subjects again judged temporal relations of the video clips while fMRI-data were acquired. The eye-movement data were analysed using both deviations from fixation and eye blinks (Kanowski et al., <xref ref-type="bibr" rid="B26">2007</xref>). Three subjects who showed extensive eye movements or blinking were excluded from further analysis. In the remaining 11 subjects, neither &#x0201C;real&#x0201D; eye movements nor eye blinks showed any differential effect across the experimental conditions (i.e., eye movements &#x0003C; 1&#x000B0;).</p>
<p>During each video subjects (<italic>n</italic> &#x0003D; 11) switched 5.72 (2.34 SD) times toward a &#x0201C;congruent&#x0201D; perceptual state [i.e., one during which perception and the physical stimulus were identical] vs. 3.97 (2.0) times toward a non-veridical one. Moreover, subjects maintained veridical percepts for 9.13 (3.38) s on average, whereas non-veridical percepts lasted 6.04 (2.02) s (see Figure <xref ref-type="fig" rid="F2">2B</xref> for length of stable durations as a function of the stimulus type). This allowed for an unbiased mixed model design (see Figure <xref ref-type="fig" rid="F2">2C</xref> and Kleinschmidt et al., <xref ref-type="bibr" rid="B28">1998</xref>; Dosenbach et al., <xref ref-type="bibr" rid="B18">2006</xref> for similar approaches).</p>
</sec>
<sec>
<title>Neuroimaging results</title>
<sec>
<title>Voxel-based group results</title>
<p>First, we computed candidate multisensory structures (i.e., the overlap of activation patterns found with unisensory visual and auditory stimuli before the main experiment; see Beauchamp et al., <xref ref-type="bibr" rid="B5">2004b</xref>; Noesselt et al., <xref ref-type="bibr" rid="B35">2007</xref>; Szycik et al., <xref ref-type="bibr" rid="B44">2008</xref>, for similar approaches). These candidate multisensory structures comprised of bilateral superior temporal sulcus, bilateral anterior insula extending into prefrontal cortex plus bilateral premotor cortex.</p>
<p>When comparing stable perceptual states with switches we found stronger fMRI-responses in bilateral STS-c and lateral prefrontal cortex for the maintenance of perceptual states relative to switches whereas perceptual switches engaged posterior parietal regions plus anterior cingulate in accord with earlier studies (e.g., Heekeren et al., <xref ref-type="bibr" rid="B24">2008</xref>). Since perceptual switches did not significantly modulate voxels within temporal regions, we then focused on the experimental effects of the different stimulus types and of stable perceptual states (i.e., inter-response intervals) within multisensory regions.</p>
<p>First, comparisons of AV<sub>S</sub> vs. (V<sub>L</sub>&#x0002B;A<sub>L</sub>) perceptual states (collapsed over stimulus types) revealed modulations in adjacent subregions of bilateral multisensory STS-c, in right insular cortex, and in bilateral prefrontal areas (see Figure <xref ref-type="fig" rid="F3">3A</xref> and Tables <xref ref-type="table" rid="T1">1A,B</xref>); note that both asynchronous and synchronous perceptual states modulated regions within STS-c, whereas only asynchronous perceptual states additionally modulated the anterior insula and prefrontal cortex (see Table <xref ref-type="table" rid="T1">1</xref>). Second, comparisons of the physically AV<sub>S</sub> minus (V<sub>L</sub>&#x0002B;A<sub>L</sub>) stimuli (regardless of perceptual states) revealed right-lateralised modulations in middle and posterior STS-c plus prefrontal cortex (see Figure <xref ref-type="fig" rid="F3">3B</xref>, purple spots). A<sub>L</sub> and V<sub>L</sub> stimuli (relative to synchronous stimuli; see Figure <xref ref-type="fig" rid="F3">3B</xref>, red and blue spots, respectively) showed enhanced BOLD-responses in bilateral STS-c, prefrontal cortex, and anterior insula (see Tables <xref ref-type="table" rid="T2">2A&#x02013;C</xref> for local maxima). Please note, that the time-related modulations are more widespread in the left hemisphere, which might be a reason for the left-sided dominance of synchronous representation reported in other studies (e.g., Miller and D&#x00027;Esposito, <xref ref-type="bibr" rid="B34">2005</xref>; Marchant et al., <xref ref-type="bibr" rid="B32">2012</xref>).</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p><bold>Voxel-based group BOLD-effect of subjects&#x00027; stable perceptual states (time from one keypress to the next, upper row) and the effects of the different stimulus types (lower row) within audiovisual activation maps (as defined by the overlap of unisensory stimuli) thresholded at <bold><italic>p</italic></bold> &#x0003C; 0.05 (small-volume-corrected).</bold> Note that the distribution of time-sensitive regions differed in the left and right hemisphere, with the left hemisphere showing a more widespread pattern than the right hemisphere (as evidenced by the formatting). <bold>(A)</bold> Comparison of synchrony &#x0003E; asynchrony percepts collapsed over stimulus type (purple spots) highlights modulations reaching from posterior to middle STS-c. Adjacent regions within STS-c were also found to be relevant for stable asynchrony percepts &#x0003E; synchronous ones (red spots; additionally, the asynchrony &#x0003E; synchrony percepts-contrast produced significant modulations in prefrontal areas; not shown, see Table <xref ref-type="table" rid="T1">1</xref>). <bold>(B)</bold> Differential BOLD-responses for the three stimulus types collapsed over peceptual state show significant effects at the right posterior STS-c (purple spots; plus premotor regions; not shown, but see Table <xref ref-type="table" rid="T2">2</xref>) for synchronous relative to asynchronous stimulation; at both STS-c (blue spots) and prefrontal areas (not shown) for visual leading relative to synchronous stimuli, and at left posterior STS-c and right anterior/posterior STS-c (red spots plus modulations at precentral gyrus and prefrontal areas; not shown, see Table <xref ref-type="table" rid="T2">2</xref>) for auditory leading relative to synchronous stimulation.</p></caption>
<graphic xlink:href="fnint-06-00064-g0003.tif"/>
</fig>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p><bold>Local maxima (<italic>p</italic> &#x0003C; 0.05, <italic>k</italic> &#x0003E; 5 small-volume-corrected) for (A) synchrony minus asynchrony perception within multisensory regions (see Figure <xref ref-type="fig" rid="F3">3A</xref>, purple spots) and (B) asynchrony minus synchrony percepts (see Figure <xref ref-type="fig" rid="F3">3A</xref>, red spots) collapsed across physical stimulation</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left"><bold>Anatomical structure</bold></th>
<th align="left"><bold>Hemisphere</bold></th>
<th align="left"><bold>Cluster size (voxels)</bold></th>
<th align="left"><bold><italic>t</italic>-value</bold></th>
<th align="center" colspan="3"><bold>MNI coordinates</bold></th>
</tr>
<tr>
<th/>
<th/>
<th/>
<th/>
<th align="left"><bold><italic>x</italic></bold></th>
<th align="left"><bold><italic>y</italic></bold></th>
<th align="left"><bold><italic>z</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="7"><bold>A.SYNCHRONY PERCEPT &#x0003E; ASYNCHRONY PERCEPT</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal Lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">R</td>
<td align="left">96</td>
<td align="left">4.95 (0.001)</td>
<td align="left">60</td>
<td align="left">&#x02212;22</td>
<td align="left">&#x02212;2</td>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">&#x02212;</td>
<td align="left">22</td>
<td align="left">3.68 (0.005)</td>
<td align="left">62</td>
<td align="left">&#x02212;10</td>
<td align="left">&#x02212;10</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">R</td>
<td align="left">32</td>
<td align="left">2.84 (0.01)</td>
<td align="left">48</td>
<td align="left">&#x02212;38</td>
<td align="left">8</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">13</td>
<td align="left">3.47 (0.007)</td>
<td align="left">&#x02212;58</td>
<td align="left">&#x02212;34</td>
<td align="left">&#x02212;4</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>B. ASYNCHRONY PERCEPT &#x0003E; SYNCHRONY PERCEPT</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal Lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">R</td>
<td align="left">206</td>
<td align="left">5.63 (0.000)</td>
<td align="left">66</td>
<td align="left">&#x02212;34</td>
<td align="left">&#x02212;6</td>
</tr>
<tr>
<td align="left">Posterior/middle STS/MTG</td>
<td align="left">R</td>
<td align="left">10</td>
<td align="left">2.98 (0.007)</td>
<td align="left">58</td>
<td align="left">&#x02212;40</td>
<td align="left">&#x02212;8</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">R</td>
<td align="left">14</td>
<td align="left">2.92 (0.008)</td>
<td align="left">54</td>
<td align="left">&#x02212;44</td>
<td align="left">18</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">17</td>
<td align="left">3.56 (0.006)</td>
<td align="left">&#x02212;66</td>
<td align="left">&#x02212;50</td>
<td align="left">2</td>
</tr>
<tr>
<td align="left"><bold>Frontal Lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior insula</td>
<td align="left">R</td>
<td align="left">14</td>
<td align="left">2.66 (0.01)</td>
<td align="left">42</td>
<td align="left">36</td>
<td align="left">&#x02212;10</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">R</td>
<td align="left">644</td>
<td align="left">10.09 (0.000)</td>
<td align="left">56</td>
<td align="left">24</td>
<td align="left">22</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">L</td>
<td align="left">9</td>
<td align="left">2.44 (0.02)</td>
<td align="left">&#x02212;54</td>
<td align="left">30</td>
<td align="left">14</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>MNI, Montreal Neurological institute; L, left; R, right.</italic></p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p><bold>Local maxima (<italic>p</italic> &#x0003C; 0.05, <italic>k</italic> &#x0003E; 5 small-volume-corrected) for (A) AV<sub>S</sub> minus (A<sub>L</sub>&#x0002B;V<sub>L</sub>) stimulation within multisensory regions (see Figure <xref ref-type="fig" rid="F3">3B</xref>, purple spots); (B) V<sub>L</sub> minus synchrony stimulation (see Figure <xref ref-type="fig" rid="F3">3B</xref>, blue spots); and (C) A<sub>L</sub> minus synchrony stimulation (see Figure <xref ref-type="fig" rid="F3">3B</xref>, red spots) collapsed across perceptual states</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left"><bold>Anatomical structure</bold></th>
<th align="left"><bold>Hemisphere</bold></th>
<th align="left"><bold>Cluster size (voxels)</bold></th>
<th align="left"><bold><italic>t</italic>-value</bold></th>
<th align="center" colspan="3"><bold>MNI coordinates</bold></th>
</tr>
<tr>
<th/>
<th/>
<th/>
<th/>
<th align="left"><bold><italic>x</italic></bold></th>
<th align="left"><bold><italic>y</italic></bold></th>
<th align="left"><bold><italic>z</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="7"><bold>A. PHYSICAL SYNCHRONY &#x0003E; PHYSICAL ASYNCHRONY</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal cortex</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">STS</td>
<td align="left">R</td>
<td align="left">9</td>
<td align="left">2.39 (0.03)</td>
<td align="left">54</td>
<td align="left">&#x02212;46</td>
<td align="left">14</td>
</tr>
<tr>
<td align="left">STS</td>
<td align="left">R</td>
<td align="left">8</td>
<td align="left">2.31 (0.04)</td>
<td align="left">62</td>
<td align="left">&#x02212;50</td>
<td align="left">10</td>
</tr>
<tr>
<td align="left"><bold>Frontal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">R</td>
<td align="left">14</td>
<td align="left">2.92 (0.008)</td>
<td align="left">50</td>
<td align="left">36</td>
<td align="left">12</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>B. PHYSICAL VISUAL LEADING ASYNCHRONY &#x0003E; PHYSICAL SYNCHRONY</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">R</td>
<td align="left">370</td>
<td align="left">3.81 (0.001)</td>
<td align="left">64</td>
<td align="left">&#x02212;20</td>
<td align="left">&#x02212;12</td>
</tr>
<tr>
<td align="left">Middle STS</td>
<td align="left">L</td>
<td align="left">9</td>
<td align="left">3.05 (0.005)</td>
<td align="left">&#x02212;68</td>
<td align="left">&#x02212;38</td>
<td align="left">14</td>
</tr>
<tr>
<td align="left">Middle STS</td>
<td align="left">L</td>
<td align="left">14</td>
<td align="left">2.95 (0.007)</td>
<td align="left">&#x02212;60</td>
<td align="left">&#x02212;30</td>
<td align="left">8</td>
</tr>
<tr>
<td align="left"><bold>Frontal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">R</td>
<td align="left">41</td>
<td align="left">2.78 (0.01)</td>
<td align="left">38</td>
<td align="left">18</td>
<td align="left">26</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">L</td>
<td align="left">11</td>
<td align="left">2.73 (0.01)</td>
<td align="left">&#x02212;46</td>
<td align="left">20</td>
<td align="left">28</td>
</tr>
<tr>
<td align="left">Anterior insula</td>
<td align="left">R</td>
<td align="left">8</td>
<td align="left">2.99 (0.006)</td>
<td align="left">50</td>
<td align="left">42</td>
<td align="left">2</td>
</tr>
<tr>
<td align="left">Anterior insula/IFG</td>
<td align="left">L</td>
<td align="left">41</td>
<td align="left">3.88 (0.001)</td>
<td align="left">&#x02212;36</td>
<td align="left">38</td>
<td align="left">&#x02212;16</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>C. PHYSICAL AUDITORY LEADING ASYNCHRONY &#x0003E; PHYSICAL SYNCHRONY</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal Lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">R</td>
<td align="left">177</td>
<td align="left">3.55 (0.002)</td>
<td align="left">62</td>
<td align="left">&#x02212;14</td>
<td align="left">&#x02212;8</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">R</td>
<td align="left">122</td>
<td align="left">3.32 (0.002)</td>
<td align="left">54</td>
<td align="left">&#x02212;46</td>
<td align="left">&#x02212;2</td>
</tr>
<tr>
<td align="left">Middle STS</td>
<td align="left">L</td>
<td align="left">12</td>
<td align="left">4.07 (0.001)</td>
<td align="left">&#x02212;68</td>
<td align="left">&#x02212;38</td>
<td align="left">14</td>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">L</td>
<td align="left">57</td>
<td align="left">2.86 (0.009)</td>
<td align="left">&#x02212;54</td>
<td align="left">&#x02212;54</td>
<td align="left">8</td>
</tr>
<tr>
<td align="left"><bold>Frontal Lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Precentral gyrus</td>
<td align="left">R</td>
<td align="left">17</td>
<td align="left">3.27 (0.003)</td>
<td align="left">44</td>
<td align="left">0</td>
<td align="left">40</td>
</tr>
<tr>
<td align="left">Precentral gyrus</td>
<td align="left">R</td>
<td align="left">7</td>
<td align="left">2.47 (0.02)</td>
<td align="left">48</td>
<td align="left">6</td>
<td align="left">44</td>
</tr>
<tr>
<td align="left">Anterior insula/IFG</td>
<td align="left">L</td>
<td align="left">6</td>
<td align="left">3.31 (0.002)</td>
<td align="left">&#x02212;36</td>
<td align="left">40</td>
<td align="left">&#x02212;18</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">L</td>
<td align="left">19</td>
<td align="left">2.46 (0.02)</td>
<td align="left">&#x02212;46</td>
<td align="left">22</td>
<td align="left">24</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>MNI, Montreal neurological institute; L, left; R, right.</italic></p>
</table-wrap-foot>
</table-wrap>
<p>Finally, we compared different perceptual states separately for each stimulus type (and not collapsed across stimulus type as above). Note that these stimulus-type-specific comparisons were designed to reveal perceptual effects for identical physical stimuli. Differential non-overlapping BOLD-modulations were again found in anterior insula, prefrontal cortex, and STS-c; with only asynchronous perceptions expressing higher activations in the insula and prefrontal cortex (see Figure <xref ref-type="fig" rid="F4">4</xref>, plus Tables <xref ref-type="table" rid="T3">3A&#x02013;C</xref>). Within STS-c, distinct regions for synchronous and asynchronous perceptions were observed as a function of stimulus type. BOLD-modulations for AL and VL conditions (veridically perceived as asynchronous) enclosed a region with an enhanced BOLD-response for veridically perceived AVS stimuli within the left hemisphere (see Figure <xref ref-type="fig" rid="F4">4</xref>, middle row and lower left panel). In the right hemisphere, regions within the STS-c responded to veridically perceived AVS and VL stimuli (see Figure <xref ref-type="fig" rid="F4">4</xref>, middle and bottom row). We also investigated whether we would find modulations in the fMRI-signal for the main effects of stimulus type, perception and perceptual states for each stimulus type outside the multisensory ROI. However no significant modulations were observed (p<sub>FWE&#x02212;corrected</sub> &#x0003C; 0.05, since we did not have any <italic>a priori</italic> hypothesis).</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>Activation maps comparing participants&#x00027; stable veridical percepts (i.e., identical with the physical stimulus) against non-veridical percepts within audiovisual regions thresholded at <italic>p</italic> &#x0003C; 0.05; <italic>k</italic> &#x0003E; 5 (small-volume-corrected).</bold> Whereas the right-sided STS-c is only modulated by synchronous &#x0003E; asynchronous percepts (orange spots, upper left panel) when processing synchronous stimuli, the results also show higher activations for asynchronous judgments during asynchronous stimulation [both auditory (red spots) and visual leading (blue spots); represents coherence] compared to synchronous judgments during asynchronous stimulation (incoherence) within prefrontal regions (top right) and the left STS-c region (middle and lower row). Note that within this area analysis revealed distinct spots for each veridical percept. Prefrontal regions were only modulated by veridical percepts during asynchrony stimulation (see Table <xref ref-type="table" rid="T3">3</xref>).</p></caption>
<graphic xlink:href="fnint-06-00064-g0004.tif"/>
</fig>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p><bold>Local maxima (<italic>p</italic> &#x0003C; 0.05, <italic>k</italic> &#x0003E; 5 small-volume-corrected) for (A) Auditory leading minus synchrony percepts during A<sub>L</sub> stimulation within multisensory regions (see Figure <xref ref-type="fig" rid="F4">4</xref>, red spots); (B) synchrony minus (A<sub>L</sub> &#x0002B; V<sub>L</sub>) percepts during AV<sub>S</sub> stimulation (see Figure <xref ref-type="fig" rid="F4">4</xref>, purple spots); and (C) V<sub>L</sub> minus synchrony percepts during V<sub>L</sub> stimulation (see Figure <xref ref-type="fig" rid="F4">4</xref>, blue spots)</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left"><bold>Anatomical structure</bold></th>
<th align="left"><bold>Hemisphere</bold></th>
<th align="left"><bold>Cluster size (voxels)</bold></th>
<th align="left"><bold><italic>t</italic>-value</bold></th>
<th align="center" colspan="3"><bold>MNI coordinates</bold></th>
</tr>
<tr>
<th/>
<th/>
<th/>
<th/>
<th align="left"><bold><italic>x</italic></bold></th>
<th align="left"><bold><italic>y</italic></bold></th>
<th align="left"><bold><italic>z</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="7"><bold>A. COHERENT AUDITORY LEADING PERCEPT &#x0003E; COHERENT SYNCHRONY PERCEPT</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">57</td>
<td align="left">2.62 (0.01)</td>
<td align="left">&#x02212;52</td>
<td align="left">&#x02212;42</td>
<td align="left">4</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">18</td>
<td align="left">2.96 (0.01)</td>
<td align="left">&#x02212;64</td>
<td align="left">&#x02212;38</td>
<td align="left">10</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">6</td>
<td align="left">2.85 (0.01)</td>
<td align="left">&#x02212;64</td>
<td align="left">&#x02212;36</td>
<td align="left">&#x02212;8</td>
</tr>
<tr>
<td align="left">Middle STS</td>
<td align="left">R</td>
<td align="left">8</td>
<td align="left">3.43 (0.005)</td>
<td align="left">&#x02212;54</td>
<td align="left">&#x02212;30</td>
<td align="left">&#x02212;14</td>
</tr>
<tr>
<td align="left">Anterior/middle STS</td>
<td align="left">R</td>
<td align="left">15</td>
<td align="left">2.95 (0.007)</td>
<td align="left">&#x02212;64</td>
<td align="left">&#x02212;42</td>
<td align="left">12</td>
</tr>
<tr>
<td align="left"><bold>Frontal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior insula</td>
<td align="left">L</td>
<td align="left">20</td>
<td align="left">3.64 (0.001)</td>
<td align="left">&#x02212;32</td>
<td align="left">28</td>
<td align="left">&#x02212;6</td>
</tr>
<tr>
<td align="left">anterior insula</td>
<td align="left">R</td>
<td align="left">66</td>
<td align="left">6.52 (0.000)</td>
<td align="left">42</td>
<td align="left">32</td>
<td align="left">&#x02212;6</td>
</tr>
<tr>
<td align="left">Precentral gyrus</td>
<td align="left">R</td>
<td align="left">86</td>
<td align="left">3.98 (0.002)</td>
<td align="left">48</td>
<td align="left">10</td>
<td align="left">40</td>
</tr>
<tr>
<td align="left">precentral gyrus</td>
<td align="left">L</td>
<td align="left">28</td>
<td align="left">4.67 (0.000)</td>
<td align="left">&#x02212;36</td>
<td align="left">8</td>
<td align="left">60</td>
</tr>
<tr>
<td align="left">Precentral gyrus</td>
<td align="left">L</td>
<td align="left">11</td>
<td align="left">3.41 (0.006)</td>
<td align="left">&#x02212;40</td>
<td align="left">8</td>
<td align="left">38</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">R</td>
<td align="left">191</td>
<td align="left">3.25 (0.003)</td>
<td align="left">54</td>
<td align="left">28</td>
<td align="left">12</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">L</td>
<td align="left">12</td>
<td align="left">2.97 (0.01)</td>
<td align="left">&#x02212;50</td>
<td align="left">6</td>
<td align="left">44</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>B. COHERENT SYNCHRONY PERCEPT &#x0003E; COHERENT ASYNCHRONY PERCEPT</bold></td>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">R</td>
<td align="left">447</td>
<td align="left">4.24 (0.001)</td>
<td align="left">62</td>
<td align="left">&#x02212;16</td>
<td align="left">&#x02212;4</td>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">L</td>
<td align="left">62</td>
<td align="left">4.05 (0.002)</td>
<td align="left">&#x02212;62</td>
<td align="left">&#x02212;24</td>
<td align="left">4</td>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">R</td>
<td align="left">5</td>
<td align="left">2.66 (0.02)</td>
<td align="left">48</td>
<td align="left">&#x02212;52</td>
<td align="left">10</td>
</tr>
<tr>
<td align="left">Posterior/middle STS</td>
<td align="left">L</td>
<td align="left">29</td>
<td align="left">3.28 (0.007)</td>
<td align="left">&#x02212;50</td>
<td align="left">&#x02212;46</td>
<td align="left">8</td>
</tr>
<tr>
<td align="left"><bold>Frontal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior Insula/Prefrontal</td>
<td align="left">L</td>
<td align="left">14</td>
<td align="left">2.63 (0.01)</td>
<td align="left">&#x02212;34</td>
<td align="left">40</td>
<td align="left">&#x02212;14</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>C. COHERENT VISUAL LEADING PERCEPT &#x0003E; COHERENT SYNCHRONY PERCEPT</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">R</td>
<td align="left">18</td>
<td align="left">3.5 (0.002)</td>
<td align="left">64</td>
<td align="left">&#x02212;50</td>
<td align="left">2</td>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">R</td>
<td align="left">5</td>
<td align="left">2.41 (0.02)</td>
<td align="left">62</td>
<td align="left">&#x02212;50</td>
<td align="left">14</td>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">L</td>
<td align="left">29</td>
<td align="left">3.47 (0.002)</td>
<td align="left">&#x02212;58</td>
<td align="left">&#x02212;50</td>
<td align="left">8</td>
</tr>
<tr>
<td align="left"><bold>Frontal lobe</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior insula</td>
<td align="left">R</td>
<td align="left">72</td>
<td align="left">4.77 (0.000)</td>
<td align="left">44</td>
<td align="left">40</td>
<td align="left">&#x02212;10</td>
</tr>
<tr>
<td align="left">Anterior insula</td>
<td align="left">L</td>
<td align="left">20</td>
<td align="left">2.71 (0.008)</td>
<td align="left">&#x02212;34</td>
<td align="left">30</td>
<td align="left">&#x02212;2</td>
</tr>
<tr>
<td align="left">Precentral gyrus</td>
<td align="left">R</td>
<td align="left">21</td>
<td align="left">3.42 (0.002)</td>
<td align="left">42</td>
<td align="left">8</td>
<td align="left">46</td>
</tr>
<tr>
<td align="left">Prefrontal cortex</td>
<td align="left">R</td>
<td align="left">451</td>
<td align="left">3.76 (0.001)</td>
<td align="left">50</td>
<td align="left">24</td>
<td align="left">24</td>
</tr>
<tr>
<td align="left">Precentral cortex</td>
<td align="left">L</td>
<td align="left">134</td>
<td align="left">3.56 (0.002)</td>
<td align="left">&#x02212;42</td>
<td align="left">16</td>
<td align="left">26</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>MNI, Montreal Neurological institute; L, left; R, right.</italic></p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec>
<title>Single-subject region-of-interest approach</title>
<p>Because of the possible anatomical differences between subjects within the STS-c (Ochiai et al., <xref ref-type="bibr" rid="B36">2004</xref>), a ROI analysis was performed within single subjects to confirm and extend voxel-based group-level responses to physical and/or perceptual (a)synchrony.</p>
<p>For this ROI analysis, three differential temporal percepts were evaluated for each subject with the following contrasts: veridical (asynchronous) minus non-veridical synchronous perception for A<sub>L</sub> and V<sub>L</sub> speech stimuli; plus synchronous minus asynchronous perception for AV<sub>S</sub> stimulus trains. This analysis was again applied within subjects&#x00027; audiovisual masks. Mean beta weights responses (proportional to percent signal change) for the subjects&#x00027; perceptual states in every experimental condition were assessed for the three (a)synchrony areas and their respective local maxima. (Note, that these local maxima were identified by conducting comparisons of a limited number of perceptual states, regardless of any other differential effects between conditions. Thus, the analysis of BOLD-effects reported below, will provide additional information concerning the overall response patterns within the STS-c-subregions):</p>
<p>We extracted the beta weights for all perceptual states (3 states &#x000D7; 3 stimulus types) from the three local maxima within STS-c and conducted a 2 &#x000D7; 3 &#x000D7; 3 &#x000D7; 3 repeated measures ANOVA with the factors of hemisphere, type of (a)synchrony area, percept, and stimulus type (see Figure <xref ref-type="fig" rid="F5">5B</xref>). As no effect of hemisphere was found [<italic>F</italic><sub>(1, 10)</sub> &#x0003C; 1; n.s.], beta weights averaged over hemispheres are displayed in Figure <xref ref-type="fig" rid="F5">5C</xref>. Interaction effects occurred between type of area, percept, and stimulus type [<italic>F</italic><sub>(8, 80)</sub> &#x0003D; 3.1; <italic>p</italic> &#x0003C; 0.01] suggesting that, within each (a)synchrony area, beta weights change as a function of the subjects&#x00027; percept and stimulus type. Main effects were observed for type of (a)synchrony area [<italic>F</italic><sub>(2, 20)</sub> &#x0003D; 4.9; <italic>p</italic> &#x0003C; 0.05] and percept [<italic>F</italic><sub>(1.33, 13.26)</sub> &#x0003D; 10.9; <italic>p</italic> &#x0003C; 0.01]. Although <italic>post-hoc t</italic>-tests showed no significant effects, responses within the &#x0201C;V<sub>L</sub> areas&#x0201D; were lower than in the other two areas. BOLD responses to synchronous stimuli were significantly lower than to asynchronous stimuli [<italic>t</italic><sub>(20)</sub> &#x0003D; &#x02212;3.53; <italic>p</italic> &#x0003C; 0.01]. Interaction effects occurred between hemisphere and type of area [<italic>F</italic><sub>(2, 20)</sub> &#x0003D; 8.04; <italic>p</italic> &#x0003C; 0.01], type of area and percept [<italic>F</italic><sub>(4, 40)</sub> &#x0003D; 3.48; <italic>p</italic> &#x0003C; 0.05], type of area and stimulus type [<italic>F</italic><sub>(1.73, 17.3)</sub> &#x0003D; 9.17; <italic>p</italic> &#x0003C; 0.01], percept and stimulus type [<italic>F</italic><sub>(1.7, 17.06)</sub> &#x0003D; 4.7; <italic>p</italic> &#x0003C; 0.05].</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><bold>Panels showing the results of single-subject analyses.</bold> The contrasts displayed here represent subjective perceptions that were congruent with physical stimulation &#x0003E; incogruent perceptions for auditory/visual leading (AL,VL, red/blue spots) and synchronous stimuli (AVS, purple spots). <bold>(A)</bold> The colored spots indicate average local maxima (11 subjects) of areas that express higher activations for veridical percepts (see main text for contrast definitions) than for non-veridical ones within the STS-c region (region of interest). The white lines and their corresponding numbers display averaged distances in millimeters from one activation spot to the two others (see &#x0201C;Methods&#x0201D; section for details). Note that asynchrony spots are always more distant from each other than from synchrony activation. Colored lines show the individual anatomical curvatures of STS-c of the all subjects after normalization. <bold>(B)</bold> The middle row depicts the activation maps of three individual subjects for the above-described contrasts. Note that synchrony spots (purple) are enclosed by two asynchrony spots [blue and red spots; see also distances in panel <bold>(A)</bold>]. Such activation patterns were found in both left and right hemispheres. <bold>(C)</bold> Mean beta-weights (proportional to % signal change) for the local maxima in panel <bold>(A)</bold> were collapsed over hemispheres. Bars show the height of the BOLD-effect (y-axis) for each stable percept [auditory leading (red bars), synchronous (orange bars), and visual leading (blue bars)] for the three stimulus types (auditory leading, visual leading, and synchrony, x-axis) within each each of the local maxima shown in panel <bold>(A)</bold> [auditory leading percept maximum (left graph section), synchrony percept maximum (middle section of graph), and visual leading percept (right graph section)]. BOLD-responses to asynchrony percepts <italic>within</italic> asynchrony percept maxima were always higher (outer left and right bars) than to any other percept for the different stimulus types. Within the synchrony percept maximum BOLD-responses to synchrony percepts were higher than asynchrony percepts whenever synchronous video clips were presented.</p></caption>
<graphic xlink:href="fnint-06-00064-g0005.tif"/>
</fig>
<p>Further analysis of the ANOVA-data (<italic>post-hoc t</italic>-tests) revealed that for each stimulus category, subjects&#x00027; BOLD responses were highest when a veridical judgment was made. Within the &#x0201C;A<sub>L</sub> area&#x0201D; (red), the mean BOLD response was highest when subjects perceived an A<sub>L</sub> stimulus as A<sub>L</sub> (veridical percept). The according beta weight differed statistically from the two other beta weights and their respective perceptual states [<italic>t</italic><sub>(10)</sub> &#x0003D; 3.12; <italic>p</italic> &#x0003C; 0.05], whereas the beta weights of the non-veridical percepts did not differ statistically from each other. The same pattern of results was also observed for the AV<sub>S</sub> region (yellow) [<italic>t</italic><sub>(10)</sub> &#x0003D; 4.76; <italic>p</italic> &#x0003C; 0.001] and V<sub>L</sub> percepts (blue) [<italic>t</italic><sub>(10)</sub> &#x0003D; 2.72; <italic>p</italic> &#x0003C; 0.05]. Since, in the AV<sub>S</sub> area, veridical responses were not significantly different from BOLD-responses for other stimulus types, this region may serve additional sub-functions on top of the maintenance of synchrony perception. In general, these ROI-results reaffirm the functional micro-compartmentalization of the STS-c found in the voxel-based group results into areas specialized for the perception of distinct audiovisual temporal patterns.</p>
</sec>
<sec>
<title>Interregional connectivity of STS-c-regions</title>
<p>Moreover, we assessed whether the subregions within STS-c that consistently expressed differential local activity (see Figure <xref ref-type="fig" rid="F5">5</xref>) would also be functionally linked to other multisensory regions. We used the assumption-free &#x0201C;psychophysiological interaction&#x0201D; (PPI; Friston et al., <xref ref-type="bibr" rid="B21">1997</xref>) and seeded our analysis in subject-specific STS-c maxima. We analysed whether the strength of functional coupling of these adjacent STS-c-regions with other multisensory regions would differ. We found that both A<sub>L</sub> and V<sub>L</sub>-regions in bilateral STS-c showed a significantly stronger coupling with right prefrontal regions than did the AV<sub>S</sub>-region (see Figure <xref ref-type="fig" rid="F6">6</xref> and Table <xref ref-type="table" rid="T4">4</xref>). Moreover, synchronous patches with the middle STS-c expressed a stronger functional connection with posterior STS-c regions in the left hemisphere, whereas asynchronous patches showed a stronger coupling with posterior STS-c in the right hemisphere (see Table <xref ref-type="table" rid="T4">4</xref>).</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p><bold>Interregional connectivity of subjects&#x00027; stable veridical percepts (i.e., identical with the physical stimulus) during asynchronous stimulation thresholded at <italic>p</italic> &#x0003C; 0.05; <italic>k</italic> &#x0003E; 5 (small-volume-corrected).</bold> Left column: right prefrontal regions only expressed stronger coupling with temporal regions in the context of veridical asynchronous &#x0003E; non-veridical synchronous perceptions during A<sub>L</sub> and V<sub>L</sub> stimulation (see Table <xref ref-type="table" rid="T4">4</xref> for all maxima). Right column shows the origin of the brain sections depicted on the left on a lateral group mean view.</p></caption>
<graphic xlink:href="fnint-06-00064-g0006.tif"/>
</fig>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p><bold>Local Maxima (<italic>p</italic> &#x0003C; 0.05, <italic>k</italic> &#x0003E; 5 small-volume-corected) of interregional connectivity in the context of (A) veridical synchrony percepts (relative to non-veridical asynchrony percepts) during AV<sub>S</sub> stimulation; (B) veridical auditory leading percepts (relative to non-veridical synchrony percepts) during A<sub>L</sub> stimulation; (C) veridical visual leading percepts (relative to non-veridical synchrony percepts) during V<sub>L</sub> stimulation</bold>.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left"><bold>Anatomical structure</bold></th>
<th align="left"><bold>Hemisphere</bold></th>
<th align="left"><bold>Cluster size (voxels)</bold></th>
<th align="left"><bold><italic>p</italic>-value</bold></th>
<th align="center" colspan="3"><bold>MNI coordinates</bold></th>
</tr>
<tr>
<th/>
<th/>
<th/>
<th/>
<th align="left"><bold><italic>x</italic></bold></th>
<th align="left"><bold><italic>y</italic></bold></th>
<th align="left"><bold><italic>z</italic></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="7"><bold>A. PSYCHOPHYSICAL INTERACTION OF SYNCHRONY PERCEPTS</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal regions</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">L</td>
<td align="left">31</td>
<td align="left">4.32 (0.001)</td>
<td align="left">&#x02212;54</td>
<td align="left">&#x02212;54</td>
<td align="left">12</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>B. PSYCHOPHYSICAL INTERACTION OF AUDITORY LEADING PERCEPTS</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal regions</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Anterior STS</td>
<td align="left">R</td>
<td align="left">73</td>
<td align="left">3.57 (0.001)</td>
<td align="left">64</td>
<td align="left">&#x02212;12</td>
<td align="left">&#x02212;8</td>
</tr>
<tr>
<td align="left">Posterior STS</td>
<td align="left">R</td>
<td align="left">22</td>
<td align="left">3.32 (0.001)</td>
<td align="left">&#x02212;50</td>
<td align="left">&#x02212;46</td>
<td align="left">16</td>
</tr>
<tr>
<td align="left"><bold>Frontal regions</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Middle/inferior frontal gyrus</td>
<td align="left">R</td>
<td align="left">168</td>
<td align="left">2.94 (0.002)</td>
<td align="left">40</td>
<td align="left">22</td>
<td align="left">16</td>
</tr>
<tr>
<td align="left">precentral gyrus</td>
<td align="left">L</td>
<td align="left">54</td>
<td align="left">2.68 (0.005)</td>
<td align="left">&#x02212;48</td>
<td align="left">&#x02212;2</td>
<td align="left">46</td>
</tr>
<tr>
<td align="left" colspan="7"><bold>C. PSYCHOPHYSICAL INTERACTION OF VISUAL LEADING PERCEPTS</bold></td>
</tr>
<tr>
<td align="left"><bold>Temporal regions</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Posterior STS/STG</td>
<td align="left">R</td>
<td align="left">44</td>
<td align="left">2.42 (0.009)</td>
<td align="left">52</td>
<td align="left">&#x02212;46</td>
<td align="left">0</td>
</tr>
<tr>
<td align="left"><bold>Frontal regions</bold></td>
<td/>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">Inferior frontal gyrus</td>
<td align="left">R</td>
<td align="left">21</td>
<td align="left">2.51 (0.008)</td>
<td align="left">58</td>
<td align="left">22</td>
<td align="left">14</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>Spatial configuration of STS-c-subregions</title>
<p>In addition, we evaluated whether the spatial configuration of the identified sub-regions within bilateral STS-c showed a systematic spatial distribution across subjects: the analysis revealed that perceived asynchrony (A<sub>L</sub> or V<sub>L</sub>) and synchrony modulated distinct regions along the STS-c which were situated adjacent to one another (with asynchrony enclosing synchrony modulations). For every subject, this specific pattern differed in its position along STS-c but occurred regularly (see Figure <xref ref-type="fig" rid="F5">5A</xref> for average, Figure <xref ref-type="fig" rid="F5">5B</xref> for illustrative subjects). Distances between the areas modulated by an interaction of stimulus type and perception were calculated. We found that, on average, the local maxima of the &#x0201C;A<sub>L</sub>&#x0201D; and &#x0201C;V<sub>L</sub> areas&#x0201D; were situated closer to &#x0201C;synchrony areas&#x0201D; (12.1 and 11.1 mm) than to each other (17.6 mm).</p>
<p>A 2 &#x000D7; 3 repeated measures ANOVA with the factors hemisphere and distance showed an effect of distance [<italic>F</italic><sub>(2, 20)</sub> &#x0003D; 10.2; <italic>p</italic> &#x0003C; 0.001]. The distance between the &#x0201C;asynchrony areas&#x0201D; was statistically different from their respective distance to the &#x0201C;synchrony area&#x0201D; [A<sub>L</sub>: <italic>t</italic><sub>(10)</sub> &#x0003D; 3.77; <italic>p</italic> &#x0003C; 0.05; V<sub>L</sub>: <italic>t</italic><sub>(10)</sub> &#x0003D; 3.40; <italic>p</italic> &#x0003C; 0.05]; the distances between the asynchrony areas and the &#x0201C;synchrony area&#x0201D; were similar [<italic>t</italic><sub>(10)</sub> &#x0003D; 0.63; <italic>p</italic> &#x0003D; 0.55]. There was no effect of hemisphere [<italic>F</italic><sub>(1, 10)</sub> &#x0003C; 1; n.s.], nor any interaction between hemisphere and distance [<italic>F</italic><sub>(2, 20)</sub> &#x0003C; 1; n.s.].</p>
</sec>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>The present study investigated the neural basis of both the processing of physical properties and subjective perception of the temporal relationship between auditory and visual speech stimuli, thereby pinpointing the functional neuroanatomy of audiovisual temporal processing and perception in multisensory cortex in humans. We found that sub-regions within the superior temporal sulcus have a distinct response pattern during the maintenance of perceptual states and for the processing of physical stimulus differences regardless of subjects&#x00027; perceptual state. Within lateral prefrontal regions and anterior insula only the perception of asynchrony was consistently linked to an increase in BOLD-response. A ROI-based single-subject analysis corroborated and extended this pattern: three subregions within the STS-c showed a differential response for the different physical stimuli (AL, VL, and AVS). Responses were further enhanced if subjects&#x00027; perceptual states were congruent to the physical stimulus being presented. Further, analyses of interregional connectivity suggest that during the perception of asynchronous stimuli AL and VL regions within the STS-c are coupled more strongly to lateral prefrontal regions, whereas connectivity within posterior STS-c was lateralized with stronger connections of the middle with posterior STS-c in the left hemisphere for synchrony patches and with posterior STS-c in the right hemisphere for asynchronous patches. Finally, analysis of the anatomical patterning of these regions suggests that they are distributed regularly within the STS-c with a synchrony region being enclosed by asynchrony regions.</p>
<p>Previous neuroimaging studies have reported that the STS-c (among other structures) is involved in audiovisual temporal processing and synchrony perception (Calvert, <xref ref-type="bibr" rid="B9">2001</xref>; Macaluso et al., <xref ref-type="bibr" rid="B31">2004</xref>; Miller and D&#x00027;Esposito, <xref ref-type="bibr" rid="B34">2005</xref>; Dhamala et al., <xref ref-type="bibr" rid="B16">2007</xref>; Noesselt et al., <xref ref-type="bibr" rid="B35">2007</xref>; Stevenson et al., <xref ref-type="bibr" rid="B43">2010</xref>; Marchant et al., <xref ref-type="bibr" rid="B32">2012</xref>; see Driver and Noesselt, <xref ref-type="bibr" rid="B19">2008</xref>, for a review). However, most of these studies investigating the crossmodal binding of semantically meaningful stimuli (Calvert et al., <xref ref-type="bibr" rid="B11">2000</xref>; Calvert and Campbell, <xref ref-type="bibr" rid="B10">2003</xref>; Macaluso et al., <xref ref-type="bibr" rid="B31">2004</xref>) did not separate task- and perception-related effects; their reported modulations may therefore reflect a mixture of stimulus-, decision-, and perception-related processing.</p>
<p>Previous research (Miller and D&#x00027;Esposito, <xref ref-type="bibr" rid="B34">2005</xref>; Stevenson et al., <xref ref-type="bibr" rid="B43">2010</xref>) reported effects of the temporal fusion of short AV-syllables using event-related fMRI. Stevenson and his colleagues (<xref ref-type="bibr" rid="B43">2010</xref>) reported functional subregions within STS-c, that preferentially processed asynchronous or synchronous speech. Miller and D&#x00027;Esposito (<xref ref-type="bibr" rid="B34">2005</xref>) reported left-hemispheric modulations within STS-c for perceptual fusion and right hemispheric effects for perceptual segregation. However, the differences in stimulus materials used in the various conditions may explain the different activation maps reported there. Nonetheless, while we did not find lateralized effects of the local fMRI-signal, our interregional connectivity analysis revealed a lateralized pattern, that accord with Miller and D&#x00027;Esposito.</p>
<p>Other studies have investigated the effects of audiovisual timing with streams of simple stimuli: Calvert et al. (<xref ref-type="bibr" rid="B12">2001</xref>) investigated multisensory interactions using simple synchronized and desynchronized audiovisual stimulus sequences. Synchronous or asynchronous bimodal inputs showed non-linear enhancements or suppressions (respectively) of BOLD-responses in multisensory areas, including STS-c, plus frontal regions. Noesselt et al. (<xref ref-type="bibr" rid="B35">2007</xref>) reported effects in contralateral STS-c for the processing of lateralized non-semantic synchronous audiovisual stimuli, but did not report effects for asynchronous audiovisual stimuli. In a related study, Marchant et al. (<xref ref-type="bibr" rid="B32">2012</xref>) observed left-sided synchrony representations in left STS-c. Meanwhile, van Atteveldt et al. (<xref ref-type="bibr" rid="B46">2004</xref>, <xref ref-type="bibr" rid="B47">2007</xref>) identified lateral temporal areas (PT, STP, and STS-c) as major integration sites whenever audiovisual grapheme-morpheme pairs were being processed. While the intensity of modulations increased in auditory areas for semantically congruent conditions, the location of modulations within the STS-c changed as a function of the temporal distance/delay between vision and sound: asynchrony was predominantly processed at the eccentricity of the STS-c activation pattern, whereas smaller temporal delays were related to the activation&#x00027;s core region. However, no effect of synchrony was reported for synchronous audiovisual letters in the STS-c and the reported activations for different audiovisual lags overlapped substantially.</p>
<p>In the present study, asynchronous percepts engaged the posterior STS-c, the anterior insula, and the prefrontal cortex bilaterally. Our results accord with previous imaging studies on temporal asynchrony which reported right-sided effects within the STS-c, supplementary motor areas (Miller and D&#x00027;Esposito, <xref ref-type="bibr" rid="B34">2005</xref>) and prefrontal (MFG, IFG) cortices (Bushara et al., <xref ref-type="bibr" rid="B8">2001</xref>; Dhamala et al., <xref ref-type="bibr" rid="B16">2007</xref>) in the perception of asynchrony. Our findings corroborate previous results and suggest that audiovisual prefrontal areas and the STS-c are functionally linked during the maintenance of the perception of audiovisual asynchrony. There is also corroborating anatomical evidence that the STS-c is reciprocally linked to prefrontal regions (see e.g., Yeterian et al., <xref ref-type="bibr" rid="B52">2012</xref>). We speculate that the perception of asynchronous percepts may be more demanding than synchrony perception and requires the on-line updating of two separate working memory representations in prefrontal cortex with input from the STS-c. Alternatively, the separation of auditory and visual input may be processed by prefrontal cortical regions (in line with the notion of a hierarchical multisensory processing model, see e.g., Noppeney et al., <xref ref-type="bibr" rid="B34a">2010</xref>) and fed back into the STS-c. Future research in non-human primates or in humans using transcranial magnetic stimulation/transcranial direct current stimulation is needed to disentangle these two possibilities.</p>
<p>Most remarkably of all, our results indicate that the multisensory superior temporal sulcus complex (mSTS-c) can be further differentiated into subregions that process particular audiovisual temporal patterns. Anatomical studies in non-human primates that have investigated the anatomical texture of TPO (the likely homologue to the human STS-c; Beauchamp, <xref ref-type="bibr" rid="B2">2005a</xref>) have provided evidence for three caudal-to-rostral subdivisions within this region (Cusick et al., <xref ref-type="bibr" rid="B13">1995</xref>). Those subdivisions are distinct in terms of their chemoarchitecture. Seltzer and Pandya (<xref ref-type="bibr" rid="B39">1991</xref>) provided evidence that TPO consists of cytoarchitectonic subdivisons of which particularly the rostral part is directly connected to the insula. Further chemoarchitectonic results support the view that the upper bank of TPO in the rhesus monkey contains several different anatomical and functional zones (Padberg et al., <xref ref-type="bibr" rid="B37">2003</xref>). They demonstrated that within those distinct neurochemical/connectional modules the STS-c shows a patchy organization of connections toward other cerebral regions. Those patches within the STS-c may have functional relevance. In a functional imaging study, Beauchamp et al. (<xref ref-type="bibr" rid="B4">2004a</xref>) reported that STS-c can be parcellated into unisensory auditory, visual, and multisensory patches. Our imaging analysis extends these findings and reveals distinct multisensory patches along the STS-c that encode separate audiovisual temporal patterns when the synchrony/asynchrony of continuous speech is being judged. Given that the identified synchrony patches lie in-between auditory- and visual-leading audiovisual patches, these modulations build up a chronological array that suggests the existence of a &#x0201C;time line.&#x0201D; Moreover, another publication (Fairhall and Macaluso, <xref ref-type="bibr" rid="B20">2009</xref>) also reported a modulation of the fMRI-signal due to attention within middle but not posterior STS-c, when subjects processed congruent audiovisual speech, thereby suggesting a large-scale segregation of the STS-c along the anterior-posterior axis (though asynchronous representations seem to be more variable; see Tables <xref ref-type="table" rid="T1">1</xref>&#x02013;<xref ref-type="table" rid="T3">3</xref>). Moreover, Marchant et al. (<xref ref-type="bibr" rid="B32">2012</xref>) investigated the correspondence of an audiovisual behavioral benefit on BOLD-modulations in the cerebrum and found significant effects in middle but not posterior STS-c for synchronous stimulus trains. The results from our study&#x02014;revealing an interaction effects in middle STS-c specific for temporal patterns and their perception plus an enhanced connectivity with more posterior regions&#x02014;are in accord with this proposition (though note that our results did not reveal a clear anterior-posterior distinction for the main effects of physical vs. perceptual states). Finally, our results could be applied to nonhuman primates to enable more invasive measures [combined with fMRI (see Tsao et al., <xref ref-type="bibr" rid="B45">2006</xref>)] to identify the pathways and neural mechanisms involved. A study in non-human primates on audiovisual face-voice integration (Ghazanfar et al., <xref ref-type="bibr" rid="B22">2008</xref>) reported enhanced coupling of STS-c-neurons with auditory areas when processing audiovisual stimuli (Schroeder et al., <xref ref-type="bibr" rid="B38">2008</xref>). Our results would predict the existence of distinct patches within mSTS-c that may differentially engage unisensory cortices via feedback connections (Driver and Noesselt, <xref ref-type="bibr" rid="B19">2008</xref>).</p>
<p>In conclusion, we found a distinct pattern of modulations within mSTS-c reflecting an interaction between perceptual state and the physical properties of audiovisual speech stimuli. Our data therefore suggest that there is an aligned spatial representation of audiovisual temporal patterns parcellating the multisensory STS-c in humans, with differential functional connections to multisensory prefrontal regions.</p>
<sec>
<title>Conflict of interest statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
</sec>
</body>
<back>
<ack>
<p>T&#x000F6;mme Noesselt, Daniel Bergmann, and Hans-Jochen Heinze were supported by DFG-SFB-TR31/TPA8; Thomas M&#x000FC;nte by DFG-SFB-TR31/TPA7 and Charles Spence by the Alexander von Humboldt Foundation, Germany. T&#x000F6;mme Noesselt, Daniel Bergmann, and Charles Spence planned the experiment. Daniel Bergmann collected and analysed the data. All authors were involved in writing the MS.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barraclough</surname> <given-names>N. E.</given-names></name> <name><surname>Xiao</surname> <given-names>D.</given-names></name> <name><surname>Baker</surname> <given-names>C. I.</given-names></name> <name><surname>Oram</surname> <given-names>M. W.</given-names></name> <name><surname>Perrett</surname> <given-names>D. I.</given-names></name></person-group> (<year>2005</year>). <article-title>Integration of visual and auditory information by superior temporal sulcus neurons responsive to the sight of actions</article-title>. <source>J. Cogn. Neurosci</source>. <volume>17</volume>, <fpage>377</fpage>&#x02013;<lpage>391</lpage>. <pub-id pub-id-type="doi">10.1162/0898929053279586</pub-id><pub-id pub-id-type="pmid">15813999</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beauchamp</surname> <given-names>M. S.</given-names></name></person-group> (<year>2005a</year>). <article-title>See me, hear me, touch me: multisensory integration in lateral occipital-temporal cortex</article-title>. <source>Curr. Opin. Neurobiol</source>. <volume>15</volume>, <fpage>145</fpage>&#x02013;<lpage>153</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2005.03.011</pub-id><pub-id pub-id-type="pmid">15831395</pub-id></citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beauchamp</surname> <given-names>M. S.</given-names></name></person-group> (<year>2005b</year>). <article-title>Statistical criteria in fMRI studies of multisensory integration</article-title>. <source>Neuroinformatics</source> <volume>3</volume>, <fpage>93</fpage>&#x02013;<lpage>113</lpage>. <pub-id pub-id-type="doi">10.1385/NI:3:2:093</pub-id><pub-id pub-id-type="pmid">15988040</pub-id></citation>
</ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beauchamp</surname> <given-names>M. S.</given-names></name> <name><surname>Argall</surname> <given-names>B. D.</given-names></name> <name><surname>Bodurka</surname> <given-names>J.</given-names></name> <name><surname>Duyn</surname> <given-names>J. H.</given-names></name> <name><surname>Martin</surname> <given-names>A.</given-names></name></person-group> (<year>2004a</year>). <article-title>Unraveling multisensory integration: patchy organization within human STS-c multisensory cortex</article-title>. <source>Nat. Neurosci</source>. <volume>7</volume>, <fpage>1190</fpage>&#x02013;<lpage>1192</lpage>. <pub-id pub-id-type="doi">10.1038/nn1333</pub-id><pub-id pub-id-type="pmid">15475952</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Beauchamp</surname> <given-names>M. S.</given-names></name> <name><surname>Lee</surname> <given-names>K. E.</given-names></name> <name><surname>Argall</surname> <given-names>B. D.</given-names></name> <name><surname>Martin</surname> <given-names>A.</given-names></name></person-group> (<year>2004b</year>). <article-title>Integration of auditory and visual information about objects in superior temporal sulcus</article-title>. <source>Neuron</source> <volume>41</volume>, <fpage>809</fpage>&#x02013;<lpage>823</lpage>. <pub-id pub-id-type="doi">10.1016/S0896-6273(04)00070-4</pub-id><pub-id pub-id-type="pmid">15003179</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Benevento</surname> <given-names>L. A.</given-names></name> <name><surname>Fallon</surname> <given-names>J.</given-names></name> <name><surname>Davis</surname> <given-names>B. J.</given-names></name> <name><surname>Rezak</surname> <given-names>M.</given-names></name></person-group> (<year>1977</year>). <article-title>Auditory-visual interaction in single cells in the cortex of the superior temporal sulcus and the orbital frontal cortex of the macaque monkey</article-title>. <source>Exp. Neurol</source>. <volume>57</volume>, <fpage>849</fpage>&#x02013;<lpage>872</lpage>. <pub-id pub-id-type="pmid">411682</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bruce</surname> <given-names>C.</given-names></name> <name><surname>Desimone</surname> <given-names>R.</given-names></name> <name><surname>Gross</surname> <given-names>C. G.</given-names></name></person-group> (<year>1981</year>). <article-title>Visual properties of neurons in a polysensory area in superior temporal sulcus of the macaque</article-title>. <source>J. Neurophysiol</source>. <volume>46</volume>, <fpage>369</fpage>&#x02013;<lpage>384</lpage>. <pub-id pub-id-type="pmid">6267219</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bushara</surname> <given-names>K. O.</given-names></name> <name><surname>Grafman</surname> <given-names>J.</given-names></name> <name><surname>Hallett</surname> <given-names>M.</given-names></name></person-group> (<year>2001</year>). <article-title>Neural correlates of auditory-visual stimulus onset asynchrony detection</article-title>. <source>J. Neurosci</source>. <volume>21</volume>, <fpage>300</fpage>&#x02013;<lpage>304</lpage>. <pub-id pub-id-type="pmid">11150347</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Calvert</surname> <given-names>G. A.</given-names></name></person-group> (<year>2001</year>). <article-title>Crossmodal processing in the human brain: insights from functional neuroimaging studies</article-title>. <source>Cereb. Cortex</source> <volume>11</volume>, <fpage>1110</fpage>&#x02013;<lpage>1123</lpage>. <pub-id pub-id-type="pmid">11709482</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Calvert</surname> <given-names>G. A.</given-names></name> <name><surname>Campbell</surname> <given-names>R.</given-names></name></person-group> (<year>2003</year>). <article-title>Reading speech from still and moving faces: the neural substrates of visible speech</article-title>. <source>J. Cogn. Neurosci</source>. <volume>15</volume>, <fpage>57</fpage>&#x02013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1162/089892903321107828</pub-id><pub-id pub-id-type="pmid">12590843</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Calvert</surname> <given-names>G. A.</given-names></name> <name><surname>Campbell</surname> <given-names>R.</given-names></name> <name><surname>Brammer</surname> <given-names>M. J.</given-names></name></person-group> (<year>2000</year>). <article-title>Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex</article-title>. <source>Curr. Biol</source>. <volume>10</volume>, <fpage>649</fpage>&#x02013;<lpage>657</lpage>. <pub-id pub-id-type="doi">10.1016/S0960-9822(00)00513-3</pub-id><pub-id pub-id-type="pmid">10837246</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Calvert</surname> <given-names>G. A.</given-names></name> <name><surname>Hansen</surname> <given-names>P. C.</given-names></name> <name><surname>Iversen</surname> <given-names>S. D.</given-names></name> <name><surname>Brammer</surname> <given-names>M. J.</given-names></name></person-group> (<year>2001</year>). <article-title>Detection of audio-visual integration sites in humans by application of electrophysiological criteria to the BOLD effect</article-title>. <source>Neuroimage</source> <volume>14</volume>, <fpage>427</fpage>&#x02013;<lpage>438</lpage>. <pub-id pub-id-type="doi">10.1006/nimg.2001.0812</pub-id><pub-id pub-id-type="pmid">11467916</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cusick</surname> <given-names>C. G.</given-names></name> <name><surname>Seltzer</surname> <given-names>B.</given-names></name> <name><surname>Cola</surname> <given-names>M.</given-names></name> <name><surname>Griggs</surname> <given-names>E.</given-names></name></person-group> (<year>1995</year>). <article-title>Chemoarchitectonics and corticocortical terminations within the superior temporal sulcus of the rhesus monkey: evidence for subdivisions of superior temporal polysensory cortex</article-title>. <source>J. Comp. Neurol</source>. <volume>360</volume>, <fpage>513</fpage>&#x02013;<lpage>535</lpage>. <pub-id pub-id-type="doi">10.1002/cne.903600312</pub-id><pub-id pub-id-type="pmid">8543656</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dennett</surname> <given-names>D.</given-names></name></person-group> (<year>1991</year>). <source>Consciousness Explained</source>. <publisher-loc>London</publisher-loc>: <publisher-name>Penguin Press</publisher-name>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Desimone</surname> <given-names>R.</given-names></name> <name><surname>Gross</surname> <given-names>C. G.</given-names></name></person-group> (<year>1979</year>). <article-title>Visual areas in the temporal cortex of the macaque</article-title>. <source>Brain Res</source>. <volume>178</volume>, <fpage>363</fpage>&#x02013;<lpage>380</lpage>. <pub-id pub-id-type="doi">10.1016/0006-8993(79)90699-1</pub-id><pub-id pub-id-type="pmid">116712</pub-id></citation>
</ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dhamala</surname> <given-names>M.</given-names></name> <name><surname>Assisi</surname> <given-names>C. G.</given-names></name> <name><surname>Jirsa</surname> <given-names>V. K.</given-names></name> <name><surname>Steinberg</surname> <given-names>F. L.</given-names></name> <name><surname>Kelso</surname> <given-names>J. A.</given-names></name></person-group> (<year>2007</year>). <article-title>Multisensory integration for timing engages different brain networks</article-title>. <source>Neuroimage</source> <volume>34</volume>, <fpage>764</fpage>&#x02013;<lpage>773</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2006.07.044</pub-id><pub-id pub-id-type="pmid">17098445</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dixon</surname> <given-names>N. F.</given-names></name> <name><surname>Spitz</surname> <given-names>L.</given-names></name></person-group> (<year>1980</year>). <article-title>The detection of auditory visual desynchrony</article-title>. <source>Perception</source> <volume>9</volume>, <fpage>719</fpage>&#x02013;<lpage>721</lpage>. <pub-id pub-id-type="pmid">7220244</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dosenbach</surname> <given-names>N. U.</given-names></name> <name><surname>Visscher</surname> <given-names>K. M.</given-names></name> <name><surname>Palmer</surname> <given-names>E. D.</given-names></name> <name><surname>Miezin</surname> <given-names>F. M.</given-names></name> <name><surname>Wenger</surname> <given-names>K. K.</given-names></name> <name><surname>Kang</surname> <given-names>H. C.</given-names></name> <name><surname>Burgund</surname> <given-names>E. D.</given-names></name> <name><surname>Grimes</surname> <given-names>A. L.</given-names></name> <name><surname>Schlaggar</surname> <given-names>B. L.</given-names></name> <name><surname>Petersen</surname> <given-names>S. E.</given-names></name></person-group> (<year>2006</year>). <article-title>A core system for the implementation of task sets</article-title>. <source>Neuron</source> <volume>50</volume>, <fpage>799</fpage>&#x02013;<lpage>812</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2006.04.031</pub-id><pub-id pub-id-type="pmid">16731517</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Driver</surname> <given-names>J.</given-names></name> <name><surname>Noesselt</surname> <given-names>T.</given-names></name></person-group> (<year>2008</year>). <article-title>Multisensory interplay reveals crossmodal influences on &#x02018;sensory-specific&#x02019; brain regions, neural responses, and judgments</article-title>. <source>Neuron</source> <volume>57</volume>, <fpage>11</fpage>&#x02013;<lpage>23</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2007.12.013</pub-id><pub-id pub-id-type="pmid">18184561</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fairhall</surname> <given-names>S.</given-names></name> <name><surname>Macaluso</surname> <given-names>E.</given-names></name></person-group> (<year>2009</year>). <article-title>Spatial attention can modulate audiovisual integration at multiple cortical and subcortical sites</article-title>. <source>Eur. J. Neurosci</source>. <volume>29</volume>, <fpage>1247</fpage>&#x02013;<lpage>1257</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2009.06688.x</pub-id><pub-id pub-id-type="pmid">19302160</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friston</surname> <given-names>K. J.</given-names></name> <name><surname>Buechel</surname> <given-names>C.</given-names></name> <name><surname>Fink</surname> <given-names>G. R.</given-names></name> <name><surname>Morris</surname> <given-names>J.</given-names></name> <name><surname>Rolls</surname> <given-names>E.</given-names></name> <name><surname>Dolan</surname> <given-names>R. J.</given-names></name></person-group> (<year>1997</year>). <article-title>Psychophysiological and modulatory interactions in neuroimaging</article-title>. <source>Neuroimage</source> <volume>6</volume>, <fpage>218</fpage>&#x02013;<lpage>229</lpage>. <pub-id pub-id-type="doi">10.1006/nimg.1997.0291</pub-id><pub-id pub-id-type="pmid">9344826</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ghazanfar</surname> <given-names>A. A.</given-names></name> <name><surname>Chandrasekaran</surname> <given-names>C.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name></person-group> (<year>2008</year>). <article-title>Interactions between the superior temporal sulcus and auditory cortex mediate dynamic face/voice integration in rhesus monkeys</article-title>. <source>J. Neurosci</source>. <volume>28</volume>, <fpage>4457</fpage>&#x02013;<lpage>4469</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0541-08.2008</pub-id><pub-id pub-id-type="pmid">18434524</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gil-da-Costa</surname> <given-names>R.</given-names></name> <name><surname>Braun</surname> <given-names>A.</given-names></name> <name><surname>Lopes</surname> <given-names>M.</given-names></name> <name><surname>Hauser</surname> <given-names>M. D.</given-names></name> <name><surname>Carson</surname> <given-names>R. E.</given-names></name> <name><surname>Herscovitch</surname> <given-names>P.</given-names></name> <name><surname>Martin</surname> <given-names>A.</given-names></name></person-group> (<year>2004</year>). <article-title>Toward an evolutionary perspective on conceptual representation: species-specific calls activate visual and affective processing systems in the macaque</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A</source>. <volume>101</volume>, <fpage>17516</fpage>&#x02013;<lpage>17521</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.0408077101</pub-id><pub-id pub-id-type="pmid">15583132</pub-id></citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heekeren</surname> <given-names>H. R.</given-names></name> <name><surname>Marrett</surname> <given-names>S.</given-names></name> <name><surname>Ungerleider</surname> <given-names>L. G.</given-names></name></person-group> (<year>2008</year>). <article-title>The neural systems that mediate human perceptual decision making</article-title>. <source>Nat. Rev. Neurosci</source>. <volume>9</volume>, <fpage>467</fpage>&#x02013;<lpage>479</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2374</pub-id><pub-id pub-id-type="pmid">18464792</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hikosaka</surname> <given-names>K.</given-names></name> <name><surname>Iwai</surname> <given-names>E.</given-names></name> <name><surname>Saito</surname> <given-names>H.</given-names></name> <name><surname>Tanaka</surname> <given-names>K.</given-names></name></person-group> (<year>1988</year>). <article-title>Polysensory properties of neurons in the anterior bank of the caudal superior temporal sulcus of the macaque monkey</article-title>. <source>J. Neurophysiol</source>. <volume>60</volume>, <fpage>1615</fpage>&#x02013;<lpage>1637</lpage>. <pub-id pub-id-type="pmid">2462027</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kanowski</surname> <given-names>M.</given-names></name> <name><surname>Rieger</surname> <given-names>J. W.</given-names></name> <name><surname>Noesselt</surname> <given-names>T.</given-names></name> <name><surname>Tempelmann</surname> <given-names>C.</given-names></name> <name><surname>Hinrichs</surname> <given-names>H.</given-names></name></person-group> (<year>2007</year>). <article-title>Endoscopic eye tracking system for fMRI</article-title>. <source>J. Neurosci. Methods</source> <volume>160</volume>, <fpage>10</fpage>&#x02013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneumeth.2006.08.001</pub-id><pub-id pub-id-type="pmid">16978705</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kelly</surname> <given-names>S. D.</given-names></name></person-group> (<year>2005</year>). <article-title>The puzzle of temporal experience</article-title>, in <source>Cognition and the Brain: The Philosophy and Neuroscience movement</source>, ed <person-group person-group-type="editor"><name><surname>Akins</surname> <given-names>A. B. K.</given-names></name></person-group> (<publisher-loc>Cambridge</publisher-loc>: <publisher-name>Cambridge University Press</publisher-name>), <fpage>208</fpage>&#x02013;<lpage>240</lpage>.</citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kleinschmidt</surname> <given-names>A.</given-names></name> <name><surname>Buchel</surname> <given-names>C.</given-names></name> <name><surname>Zeki</surname> <given-names>S.</given-names></name> <name><surname>Frackowiak</surname> <given-names>R. S.</given-names></name></person-group> (<year>1998</year>). <article-title>Human brain activity during spontaneously reversing perception of ambiguous figures</article-title>. <source>Proc. Biol. Sci</source>. <volume>265</volume>, <fpage>2427</fpage>&#x02013;<lpage>2433</lpage>. <pub-id pub-id-type="doi">10.1098/rspb.1998.0594</pub-id><pub-id pub-id-type="pmid">9921682</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>K&#x000F6;hler</surname> <given-names>W.</given-names></name></person-group> (<year>1947</year>). <source>Gestalt Psychology: An Introduction to New Concepts in Modern Psychology</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Liveright Publication Corporation</publisher-name>.</citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Macaluso</surname> <given-names>E.</given-names></name> <name><surname>George</surname> <given-names>N.</given-names></name> <name><surname>Dolan</surname> <given-names>R.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Driver</surname> <given-names>J.</given-names></name></person-group> (<year>2004</year>). <article-title>Spatial and temporal factors during processing of audiovisual speech: a PET study</article-title>. <source>Neuroimage</source> <volume>21</volume>, <fpage>725</fpage>&#x02013;<lpage>732</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2003.09.049</pub-id><pub-id pub-id-type="pmid">14980575</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Marchant</surname> <given-names>J. L.</given-names></name> <name><surname>Ruff</surname> <given-names>C. C.</given-names></name> <name><surname>Driver</surname> <given-names>J.</given-names></name></person-group> (<year>2012</year>). <article-title>Audiovisual synchrony enhances BOLD responses in a brain network including multisensory STS while also enhancing target-detection performance for both modalities</article-title>. <source>Hum. Brain Mapp</source>. <volume>33</volume>, <fpage>1212</fpage>&#x02013;<lpage>1224</lpage>. <pub-id pub-id-type="doi">10.1002/hbm.21278</pub-id><pub-id pub-id-type="pmid">21953980</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>McGrath</surname> <given-names>M.</given-names></name> <name><surname>Summerfield</surname> <given-names>Q.</given-names></name></person-group> (<year>1985</year>). <article-title>Intermodal timing relations and audio-visual speech recognition by normal-hearing adults</article-title>. <source>J. Acoust. Soc. Am</source>. <volume>77</volume>, <fpage>678</fpage>&#x02013;<lpage>685</lpage>. <pub-id pub-id-type="doi">10.1121/1.392336</pub-id><pub-id pub-id-type="pmid">3973239</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>L. M.</given-names></name> <name><surname>D&#x00027;Esposito</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <article-title>Perceptual fusion and stimulus coincidence in the cross-modal integration of speech</article-title>. <source>J. Neurosci</source>. <volume>25</volume>, <fpage>5884</fpage>&#x02013;<lpage>5893</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0896-05.2005</pub-id><pub-id pub-id-type="pmid">15976077</pub-id></citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Noesselt</surname> <given-names>T.</given-names></name> <name><surname>Rieger</surname> <given-names>J. W.</given-names></name> <name><surname>Schoenfeld</surname> <given-names>M. A.</given-names></name> <name><surname>Kanowski</surname> <given-names>M.</given-names></name> <name><surname>Hinrichs</surname> <given-names>H.</given-names></name> <name><surname>Heinze</surname> <given-names>H. J.</given-names></name> <name><surname>Driver</surname> <given-names>J.</given-names></name></person-group> (<year>2007</year>). <article-title>Audiovisual temporal correspondence modulates human multisensory superior temporal sulcus plus primary sensory cortices</article-title>. <source>J. Neurosci</source>. <volume>27</volume>, <fpage>11431</fpage>&#x02013;<lpage>11441</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2252-07.2007</pub-id><pub-id pub-id-type="pmid">17942738</pub-id></citation>
</ref>
<ref id="B34a">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Noppeney</surname> <given-names>U.</given-names></name> <name><surname>Ostwald</surname> <given-names>D.</given-names></name> <name><surname>Werner</surname> <given-names>S.</given-names></name></person-group> (<year>2010</year>). <article-title>Perceptual decisions formed by accumulation of audiovisual evidence in prefrontal cortex</article-title>. <source>J. Neurosci</source>. <volume>30</volume>, <fpage>7434</fpage>&#x02013;<lpage>7446</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0455-10.2010</pub-id><pub-id pub-id-type="pmid">20505110</pub-id></citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ochiai</surname> <given-names>T.</given-names></name> <name><surname>Grimault</surname> <given-names>S.</given-names></name> <name><surname>Scavarda</surname> <given-names>D.</given-names></name> <name><surname>Roch</surname> <given-names>G.</given-names></name> <name><surname>Hori</surname> <given-names>T.</given-names></name> <name><surname>Riviere</surname> <given-names>D.</given-names></name> <name><surname>Mangin</surname> <given-names>J. F.</given-names></name> <name><surname>Regis</surname> <given-names>J.</given-names></name></person-group> (<year>2004</year>). <article-title>Sulcal pattern and morphology of the superior temporal sulcus</article-title>. <source>Neuroimage</source> <volume>22</volume>, <fpage>706</fpage>&#x02013;<lpage>719</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2004.01.023</pub-id><pub-id pub-id-type="pmid">15193599</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Padberg</surname> <given-names>J.</given-names></name> <name><surname>Seltzer</surname> <given-names>B.</given-names></name> <name><surname>Cusick</surname> <given-names>C. G.</given-names></name></person-group> (<year>2003</year>). <article-title>Architectonics and cortical connections of the upper bank of the superior temporal sulcus in the rhesus monkey: an analysis in the tangential plane</article-title>. <source>J. Comp. Neurol</source>. <volume>467</volume>, <fpage>418</fpage>&#x02013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1002/cne.10932</pub-id><pub-id pub-id-type="pmid">14608603</pub-id></citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schroeder</surname> <given-names>C. E.</given-names></name> <name><surname>Lakatos</surname> <given-names>P.</given-names></name> <name><surname>Kajikawa</surname> <given-names>Y.</given-names></name> <name><surname>Partan</surname> <given-names>S.</given-names></name> <name><surname>Puce</surname> <given-names>A.</given-names></name></person-group> (<year>2008</year>). <article-title>Neuronal oscillations and visual amplification of speech</article-title>. <source>Trends Cogn. Sci</source>. <volume>12</volume>, <fpage>106</fpage>&#x02013;<lpage>113</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2008.01.002</pub-id><pub-id pub-id-type="pmid">18280772</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Seltzer</surname> <given-names>B.</given-names></name> <name><surname>Pandya</surname> <given-names>D. N.</given-names></name></person-group> (<year>1991</year>). <article-title>Post-rolandic cortical projections of the superior temporal sulcus in the rhesus monkey</article-title>. <source>J. Comp. Neurol</source>. <volume>312</volume>, <fpage>625</fpage>&#x02013;<lpage>640</lpage>. <pub-id pub-id-type="doi">10.1002/cne.903120412</pub-id><pub-id pub-id-type="pmid">1761745</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Slutsky</surname> <given-names>D. A.</given-names></name> <name><surname>Recanzone</surname> <given-names>G. H.</given-names></name></person-group> (<year>2001</year>). <article-title>Temporal and spatial dependency of the ventriloquism effect</article-title>. <source>Neuroreport</source> <volume>12</volume>, <fpage>7</fpage>&#x02013;<lpage>10</lpage>. <pub-id pub-id-type="pmid">11201094</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Squire</surname> <given-names>S.</given-names></name></person-group> (<year>2003</year>). <article-title>Multisensory integration: maintaining the perception of synchrony</article-title>. <source>Curr. Biol</source>. <volume>13</volume>, <fpage>R519</fpage>&#x02013;<lpage>R521</lpage>. <pub-id pub-id-type="doi">10.1016/S0960-9822(03)00445-7</pub-id><pub-id pub-id-type="pmid">12842029</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Stein</surname> <given-names>B. E.</given-names></name> <name><surname>Meredith</surname> <given-names>M. A.</given-names></name></person-group> (<year>1993</year>). <source>The Merging of the Senses</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stevenson</surname> <given-names>R. A.</given-names></name> <name><surname>Altieri</surname> <given-names>N. A.</given-names></name> <name><surname>Kim</surname> <given-names>S.</given-names></name> <name><surname>Pisoni</surname> <given-names>D. B.</given-names></name> <name><surname>James</surname> <given-names>T. W.</given-names></name></person-group> (<year>2010</year>). <article-title>Neural processing of asynchronous audiovisual speech perception</article-title>. <source>Neuroimage</source> <volume>49</volume>, <fpage>33</fpage>&#x02013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2009.12.001</pub-id><pub-id pub-id-type="pmid">20004723</pub-id></citation>
</ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szycik</surname> <given-names>G. R.</given-names></name> <name><surname>Tausche</surname> <given-names>P.</given-names></name> <name><surname>Munte</surname> <given-names>T. F.</given-names></name></person-group> (<year>2008</year>). <article-title>A novel approach to study audiovisual integration in speech perception: localizer fMRI and sparse sampling</article-title>. <source>Brain Res</source>. <volume>1220</volume>, <fpage>142</fpage>&#x02013;<lpage>149</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2007.08.027</pub-id><pub-id pub-id-type="pmid">17880929</pub-id></citation>
</ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tsao</surname> <given-names>D. Y.</given-names></name> <name><surname>Freiwald</surname> <given-names>W. A.</given-names></name> <name><surname>Tootell</surname> <given-names>R. B.</given-names></name> <name><surname>Livingstone</surname> <given-names>M. S.</given-names></name></person-group> (<year>2006</year>). <article-title>A cortical region consisting entirely of face-selective cells</article-title>. <source>Science</source> <volume>311</volume>, <fpage>670</fpage>&#x02013;<lpage>674</lpage>. <pub-id pub-id-type="doi">10.1126/science.1119983</pub-id><pub-id pub-id-type="pmid">16456083</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Atteveldt</surname> <given-names>N.</given-names></name> <name><surname>Formisano</surname> <given-names>E.</given-names></name> <name><surname>Goebel</surname> <given-names>R.</given-names></name> <name><surname>Blomert</surname> <given-names>L.</given-names></name></person-group> (<year>2004</year>). <article-title>Integration of letters and speech sounds in the human brain</article-title>. <source>Neuron</source> <volume>43</volume>, <fpage>271</fpage>&#x02013;<lpage>282</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2004.06.025</pub-id><pub-id pub-id-type="pmid">15260962</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Atteveldt</surname> <given-names>N. M.</given-names></name> <name><surname>Formisano</surname> <given-names>E.</given-names></name> <name><surname>Blomert</surname> <given-names>L.</given-names></name> <name><surname>Goebel</surname> <given-names>R.</given-names></name></person-group> (<year>2007</year>). <article-title>The effect of temporal asynchrony on the multisensory integration of letters and speech sounds</article-title>. <source>Cereb. Cortex</source> <volume>17</volume>, <fpage>962</fpage>&#x02013;<lpage>974</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhl007</pub-id><pub-id pub-id-type="pmid">16751298</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vatakis</surname> <given-names>A.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2006a</year>). <article-title>Audiovisual synchrony perception for music, speech, and object actions</article-title>. <source>Brain Res</source>. <volume>1111</volume>, <fpage>134</fpage>&#x02013;<lpage>142</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2006.05.078</pub-id><pub-id pub-id-type="pmid">16876772</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vatakis</surname> <given-names>A.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2006b</year>). <article-title>Audiovisual synchrony perception for speech and music assessed using a temporal order judgment task</article-title>. <source>Neurosci. Lett</source>. <volume>393</volume>, <fpage>40</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1016/j.neulet.2005.09.032</pub-id><pub-id pub-id-type="pmid">16213656</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vroomen</surname> <given-names>J.</given-names></name> <name><surname>Keetels</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Perception of intersensory synchrony: a tutorial review</article-title>. <source>Atten. Percept. Psychophys</source>. <volume>72</volume>, <fpage>871</fpage>&#x02013;<lpage>884</lpage>. <pub-id pub-id-type="doi">10.3758/APP.72.4.871</pub-id><pub-id pub-id-type="pmid">20436185</pub-id></citation>
</ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wright</surname> <given-names>T. M.</given-names></name> <name><surname>Pelphrey</surname> <given-names>K. A.</given-names></name> <name><surname>Allison</surname> <given-names>T.</given-names></name> <name><surname>McKeown</surname> <given-names>M. J.</given-names></name> <name><surname>McCarthy</surname> <given-names>G.</given-names></name></person-group> (<year>2003</year>). <article-title>Polysensory interactions along lateral temporal regions evoked by audiovisual speech</article-title>. <source>Cereb. Cortex</source> <volume>13</volume>, <fpage>1034</fpage>&#x02013;<lpage>1043</lpage>. <pub-id pub-id-type="pmid">12967920</pub-id></citation>
</ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yeterian</surname> <given-names>E. H</given-names></name> <name><surname>Pandya</surname> <given-names>D. N</given-names></name> <name><surname>Tomaiuolo</surname> <given-names>F</given-names></name> <name><surname>Petrides</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>The cortical connectivity of the prefrontal cortex in themonkey brain</article-title>. <source>Cortex</source> <volume>48</volume>, <fpage>58</fpage>&#x02013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.1016/j.cortex.2011.03.004</pub-id><pub-id pub-id-type="pmid">21481342</pub-id></citation>
</ref>
</ref-list>
</back>
</article>