AB054. Audio—visual multiple object tracking
Brain and Perception

AB054. Audio—visual multiple object tracking

Vanessa Harrar1,2, Eugenie Roudaia1,2, Jocelyn Faubert1,2

1Visual Psychophysics and Perception Laboratory, School of Optometry, University of Montreal, Montreal, QC, Canada;2NSERC-Essilor Industrial Research Chair, Montreal, QC, Canada


Background: The ability to track objects as they move is critical for successful interaction with objects in the world. The multiple object tracking (MOT) paradigm has demonstrated that, within limits, our visual attention capacity allows us to track multiple moving objects among distracters. Very little is known about dynamic auditory attention and the role of multisensory binding in attentional tracking. Here, we examined whether dynamic sounds congruent with visual targets could facilitate tracking in a 3D-MOT task.

Methods: Participants tracked one or multiple target-spheres among identical distractor-spheres during 8 seconds of movement in a virtual cube. In the visual condition, targets were identified with a brief colour change, but were then indistinguishable from the distractors during the movement. In the audio-visual condition, the target-spheres were accompanied by a sound, which moved congruently with the change in the target’s position. Sound amplitude varied with distance from the observer and inter-aural amplitude difference varied with azimuth.

Results: Results with one target showed that performance was better in the audiovisual condition, which suggests that congruent sounds can facilitate attentional visual tracking. However, with multiple targets, the sounds did not facilitate tracking.

Conclusions: This suggests that audiovisual binding may not be possible when attention is divided between multiple targets.

Keywords: Tracking; attention; multisensory; audiovisual; integration


doi: 10.21037/aes.2018.AB054
Cite this abstract as: Harrar V, Roudaia E, Faubert J. Audio—visual multiple object tracking. Ann Eye Sci 2018;3:AB054.

Download Citation