Discriminating Audiovisual Speed: Optimal Integration of Speed Defaults to Probability Summation when Component Reliabilities Diverge
Overview
Authors
Affiliations
We investigated audiovisual speed perception to test the maximum-likelihood-estimation (MLE) model of multisensory integration. According to MLE, audiovisual speed perception will be based on a weighted average of visual and auditory speed estimates, with each component weighted by its inverse variance, a statistically optimal combination that produces a fused estimate with minimised variance and thereby affords maximal discrimination. We use virtual auditory space to create ecologically valid auditory motion, together with visual apparent motion around an array of 63 LEDs. To degrade the usual dominance of vision over audition, we added positional jitter to the motion sequences, and also measured peripheral trajectories. Both factors degraded visual speed discrimination, while auditory speed perception was unaffected by trajectory location. In the bimodal conditions, a speed conflict was introduced (48 degrees versus 60 degrees s(-1)) and two measures were taken: perceived audiovisual speed, and the precision (variability) of audiovisual speed discrimination. These measures showed only a weak tendency to follow MLE predictions. However, splitting the data into two groups based on whether the unimodal component weights were similar or disparate revealed interesting findings: similarly weighted components were integrated in a manner closely matching MLE predictions, while dissimilarity weighted components (greater than 3 : 1 difference) were integrated according to probability-summation predictions. These results suggest that different multisensory integration strategies may be implemented depending on relative component reliabilities, with MLE integration vetoed when component weights are highly disparate.
A model of audio-visual motion integration during active self-movement.
Gallagher M, Haynes J, Culling J, Freeman T J Vis. 2025; 25(2):8.
PMID: 39969485 PMC: 11841688. DOI: 10.1167/jov.25.2.8.
Statistically Optimal Cue Integration During Human Spatial Navigation.
Newman P, Qi Y, Mou W, McNamara T Psychon Bull Rev. 2023; 30(5):1621-1642.
PMID: 37038031 DOI: 10.3758/s13423-023-02254-w.
A review of interactions between peripheral and foveal vision.
Stewart E, Valsecchi M, Schutz A J Vis. 2020; 20(12):2.
PMID: 33141171 PMC: 7645222. DOI: 10.1167/jov.20.12.2.
But Still It Moves: Static Image Statistics Underlie How We See Motion.
Rideaux R, Welchman A J Neurosci. 2020; 40(12):2538-2552.
PMID: 32054676 PMC: 7083528. DOI: 10.1523/JNEUROSCI.2760-19.2020.
Integration of audiovisual spatial signals is not consistent with maximum likelihood estimation.
Meijer D, Veselic S, Calafiore C, Noppeney U Cortex. 2019; 119:74-88.
PMID: 31082680 PMC: 6864592. DOI: 10.1016/j.cortex.2019.03.026.