UNI-MB - logo
UMNIK - logo
 
E-resources
Full text
Peer reviewed
  • Spatial and temporal
    Keller, Peter E; Varlet, Manuel; Nozaradan, Sylvie; Lapenta, Olivia Morgan

    Experimental brain research, 03/2023, Volume: 241, Issue: 3
    Journal Article

    Human movement synchronisation with moving objects strongly relies on visual input. However, auditory information also plays an important role, since real environments are intrinsically multimodal. We used electroencephalography (EEG) frequency tagging to investigate the selective neural processing and integration of visual and auditory information during motor tracking and tested the effects of spatial and temporal congruency between audiovisual modalities. EEG was recorded while participants tracked with their index finger a red flickering (rate f.sub.V = 15 Hz) dot oscillating horizontally on a screen. The simultaneous auditory stimulus was modulated in pitch (rate f.sub.A = 32 Hz) and lateralised between left and right audio channels to induce perception of a periodic displacement of the sound source. Audiovisual congruency was manipulated in terms of space in Experiment 1 (no motion, same direction or opposite direction), and timing in Experiment 2 (no delay, medium delay or large delay). For both experiments, significant EEG responses were elicited at f.sub.V and f.sub.A tagging frequencies. It was also hypothesised that intermodulation products corresponding to the nonlinear integration of visual and auditory stimuli at frequencies f.sub.V ± f.sub.A would be elicited, due to audiovisual integration, especially in Congruent conditions.sub.. However, these components were not observed. Moreover, synchronisation and EEG results were not influenced by congruency manipulations, which invites further exploration of the conditions which may modulate audiovisual processing and the motor tracking of moving objects.