Steady-state evoked potentials as an index of multisensory temporal binding Sylvie Nozaradan a, b , Isabelle Peretz b , André Mouraux a, a Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), Belgium b International Laboratory for Brain, Music and Sound Research (BRAMS), Université de Montréal (UdeM), Canada abstract article info Article history: Received 6 July 2011 Revised 10 November 2011 Accepted 22 November 2011 Available online 1 December 2011 Keywords: EEG Multisensory integration Rhythm perception Neuronal entrainment Steady-state evoked potentials Temporal congruency promotes perceptual binding of multisensory inputs. Here, we used EEG frequency- tagging to track cortical activities elicited by auditory and visual inputs separately, in the form of steady- state evoked potentials (SS-EPs). We tested whether SS-EPs could reveal a dynamic coupling of cortical activ- ities related to the binding of auditory and visual inputs conveying synchronous vs. non-synchronous tempo- ral periodicities, or beats. The temporally congruent audiovisual condition elicited markedly enhanced auditory and visual SS-EPs, as compared to the incongruent condition. Furthermore, an increased inter-trial phase coherence of both SS-EPs was observed in that condition. Taken together, these observations indicate that temporal congruency enhances the processing of multisensory inputs at sensory-specic stages of corti- cal processing, possibly through a dynamic binding by synchrony of the elicited activities and/or improved dynamic attending. Moreover, we show that EEG frequency-tagging with SS-EPs constitutes an effective tool to explore the neural dynamics of multisensory integration in the human brain. © 2011 Elsevier Inc. All rights reserved. Introduction Building coherent representations of the external world requires integrating and merging information concurrently sampled through our different senses (Gibson, 1966; Spence and Driver, 2004). Most events occurring in the environment concomitantly activate afferents from different sensory modalities. For example, the perception of an explosion simultaneously emitting light, noise, vibrations and heat requires the integration of combined visual, auditory and somatosen- sory inputs. Because the information conveyed by these different sen- sory modalities is often complementary, cross-modal integration of these inputs may provide information about the environment that is absent in any one modality presented in isolation and, hence, cross- modal integration may improve behavior (Adrian, 1949; Elliott et al., 2010; Stein and Meredith, 1993). Temporal congruency facilitates cross-modal integration (Bertelson, 1999; Fujisaki and Nishida, 2005; Petrini et al., 2009; Sekuler et al., 1997; Vatakis and Spence, 2006; Vroomen and Keetels, 2010; Welch and Warren, 1980; Zampini et al., 2003). Multisensory perception may result from a process of binding by synchrony of the cortical responses to sensory inputs sharing similar temporal dynamics (Kayser, 2009; Luo et al., 2010; Schroeder et al., 2008; Senkowski et al., 2008). Support for this hypothesis can be found in the electrophysiological recordings per- formed in the sensory cortices of monkeys where congruent multisenso- ry inputs elicit an increased phase coherence of neuronal oscillatory activity within the activated sensory cortices, as compared to incongru- ent multisensory inputs (Kayser and Logothetis, 2007; Kayser et al., 2008; Senkowski et al., 2007). Similarly, in humans, electroencephalo- graphic (EEG) recordings reveal that the congruency of combined audi- tory and visual stimulation enhances the magnitude of stimulus- induced EEG oscillations across both auditory and visual cortices (Luo et al., 2010; Schall et al., 2009; Schroeder et al., 2008). However, because of the unavoidable temporal overlap between the neural responses to concurrent streams of sensory input, disentangling the neural activities related to each sensory stream, although critical to study multisensory integration, is difcult (Besle et al., 2009). Hence, current knowledge of how the human brain extracts, integrates and exploits the temporal dy- namics of sensory input remains, at present, poorly understood. Frequency-tagging using EEG steady-state evoked potentials (SS- EPs) could overcome this limitation, and thus may constitute a mean to study, non-invasively, multisensory integration in humans (Regan, 1989; Regan and Heron, 1969). SS-EPs are elicited by the con- tinuous presentation of a sensory stimulus in which a given feature is modulated periodically at a given frequency. SS-EPs appear as an in- crease in the EEG frequency spectrum peaking specically at the fre- quency of stimulation (Regan, 1989). Therefore, different SS-EP frequencies can be used to tag the different sensory inputs constitut- ing a multimodal stimulus and, thereby, isolate the neural activity re- lated specically to each stream of sensory input (Morgan et al., 1996; Regan, 1989; Tononi et al., 1998). This frequency-tagging approach has already been used to characterize the neural activity triggered by intermodal interactions of selective attention, using simultaneous auditory and visual inputs and comparing the magnitude of SS-EPs obtained in unisensory vs. multisensory conditions, according to the NeuroImage 60 (2012) 2128 Corresponding author at: Institute of Neurosciences, Université catholique de Lou- vain, 53, Avenue Mounier UCL 53.75, B-1200 Bruxelles, Belgium. E-mail address: andre.mouraux@uclouvain.be (A. Mouraux). 1053-8119/$ see front matter © 2011 Elsevier Inc. All rights reserved. doi:10.1016/j.neuroimage.2011.11.065 Contents lists available at SciVerse ScienceDirect NeuroImage journal homepage: www.elsevier.com/locate/ynimg