Cortical Tracking of Voice Pitch in the Presence of Multiple Speakers Depends on Selective Attention
Overview
Authors
Affiliations
Voice pitch carries linguistic and non-linguistic information. Previous studies have described cortical tracking of voice pitch in clean speech, with responses reflecting both pitch strength and pitch value. However, pitch is also a powerful cue for auditory stream segregation, especially when competing streams have pitch differing in fundamental frequency, as is the case when multiple speakers talk simultaneously. We therefore investigated how cortical speech pitch tracking is affected in the presence of a second, task-irrelevant speaker. We analyzed human magnetoencephalography (MEG) responses to continuous narrative speech, presented either as a single talker in a quiet background or as a two-talker mixture of a male and a female speaker. In clean speech, voice pitch was associated with a right-dominant response, peaking at a latency of around 100 ms, consistent with previous electroencephalography and electrocorticography results. The response tracked both the presence of pitch and the relative value of the speaker's fundamental frequency. In the two-talker mixture, the pitch of the attended speaker was tracked bilaterally, regardless of whether or not there was simultaneously present pitch in the speech of the irrelevant speaker. Pitch tracking for the irrelevant speaker was reduced: only the right hemisphere still significantly tracked pitch of the unattended speaker, and only during intervals in which no pitch was present in the attended talker's speech. Taken together, these results suggest that pitch-based segregation of multiple speakers, at least as measured by macroscopic cortical tracking, is not entirely automatic but strongly dependent on selective attention.
EEG-based cross-subject passive music pitch perception using deep learning models.
Meng Q, Tian L, Liu G, Zhang X Cogn Neurodyn. 2025; 19(1):6.
PMID: 39758357 PMC: 11699146. DOI: 10.1007/s11571-024-10196-9.
Neural encoding of melodic expectations in music across EEG frequency bands.
Galeano-Otalvaro J, Martorell J, Meyer L, Titone L Eur J Neurosci. 2024; 60(11):6734-6749.
PMID: 39469882 PMC: 11612851. DOI: 10.1111/ejn.16581.
Episodic long-term memory formation during slow-wave sleep.
Schmidig F, Ruch S, Henke K Elife. 2024; 12.
PMID: 38661727 PMC: 11045222. DOI: 10.7554/eLife.89601.
Schuller A, Schilling A, Krauss P, Rampp S, Reichenbach T J Neurosci. 2023; 43(44):7429-7440.
PMID: 37793908 PMC: 10621774. DOI: 10.1523/JNEUROSCI.1247-23.2023.
Hemispheric asymmetries for music and speech: Spectrotemporal modulations and top-down influences.
Zatorre R Front Neurosci. 2023; 16:1075511.
PMID: 36605556 PMC: 9809288. DOI: 10.3389/fnins.2022.1075511.