» Articles » PMID: 31013361

Look at Me when I'm Talking to You: Selective Attention at a Multisensory Cocktail Party Can Be Decoded Using Stimulus Reconstruction and Alpha Power Modulations

Overview
Journal Eur J Neurosci
Specialty Neurology
Date 2019 Apr 24
PMID 31013361
Citations 11
Authors
Affiliations
Soon will be listed here.
Abstract

Recent work using electroencephalography has applied stimulus reconstruction techniques to identify the attended speaker in a cocktail party environment. The success of these approaches has been primarily based on the ability to detect cortical tracking of the acoustic envelope at the scalp level. However, most studies have ignored the effects of visual input, which is almost always present in naturalistic scenarios. In this study, we investigated the effects of visual input on envelope-based cocktail party decoding in two multisensory cocktail party situations: (a) Congruent AV-facing the attended speaker while ignoring another speaker represented by the audio-only stream and (b) Incongruent AV (eavesdropping)-attending the audio-only speaker while looking at the unattended speaker. We trained and tested decoders for each condition separately and found that we can successfully decode attention to congruent audiovisual speech and can also decode attention when listeners were eavesdropping, i.e., looking at the face of the unattended talker. In addition to this, we found alpha power to be a reliable measure of attention to the visual speech. Using parieto-occipital alpha power, we found that we can distinguish whether subjects are attending or ignoring the speaker's face. Considering the practical applications of these methods, we demonstrate that with only six near-ear electrodes we can successfully determine the attended speech. This work extends the current framework for decoding attention to speech to more naturalistic scenarios, and in doing so provides additional neural measures which may be incorporated to improve decoding accuracy.

Citing Articles

Convolutional neural networks can identify brain interactions involved in decoding spatial auditory attention.

Mahjoory K, Bahmer A, Henry M PLoS Comput Biol. 2024; 20(8):e1012376.

PMID: 39116183 PMC: 11335149. DOI: 10.1371/journal.pcbi.1012376.


The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.

Ahmed F, Nidiffer A, Lalor E Front Hum Neurosci. 2024; 17:1283206.

PMID: 38162285 PMC: 10754997. DOI: 10.3389/fnhum.2023.1283206.


Distinct roles of delta- and theta-band neural tracking for sharpening and predictive coding of multi-level speech features during spoken language processing.

Mai G, Wang W Hum Brain Mapp. 2023; 44(17):6149-6172.

PMID: 37818940 PMC: 10619373. DOI: 10.1002/hbm.26503.


The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.

Ahmed F, Nidiffer A, Lalor E bioRxiv. 2023; .

PMID: 37662393 PMC: 10473711. DOI: 10.1101/2023.08.23.554451.


Age-Related Changes to Multisensory Integration and Audiovisual Speech Perception.

Pepper J, Nuttall H Brain Sci. 2023; 13(8).

PMID: 37626483 PMC: 10452685. DOI: 10.3390/brainsci13081126.