» Articles » PMID: 30378469

Multimodal and Spectral Degradation Effects on Speech and Emotion Recognition in Adult Listeners

Overview
Journal Trends Hear
Date 2018 Nov 1
PMID 30378469
Citations 1
Authors
Affiliations
Soon will be listed here.
Abstract

For cochlear implant (CI) users, degraded spectral input hampers the understanding of prosodic vocal emotion, especially in difficult listening conditions. Using a vocoder simulation of CI hearing, we examined the extent to which informative multimodal cues in a talker's spoken expressions improve normal hearing (NH) adults' speech and emotion perception under different levels of spectral degradation (two, three, four, and eight spectral bands). Participants repeated the words verbatim and identified emotions (among four alternative options: happy, sad, angry, and neutral) in meaningful sentences that are semantically congruent with the expression of the intended emotion. Sentences were presented in their natural speech form and in speech sampled through a noise-band vocoder in sound (auditory-only) and video (auditory-visual) recordings of a female talker. Visual information had a more pronounced benefit in enhancing speech recognition in the lower spectral band conditions. Spectral degradation, however, did not interfere with emotion recognition performance when dynamic visual cues in a talker's expression are provided as participants scored at ceiling levels across all spectral band conditions. Our use of familiar sentences that contained congruent semantic and prosodic information have high ecological validity, which likely optimized listener performance under simulated CI hearing and may better predict CI users' outcomes in everyday listening contexts.

Citing Articles

Age-Related Changes in the Perception of Emotions in Speech: Assessing Thresholds of Prosody and Semantics Recognition in Noise for Young and Older Adults.

Dor Y, Algom D, Shakuf V, Ben-David B Front Neurosci. 2022; 16:846117.

PMID: 35546888 PMC: 9082150. DOI: 10.3389/fnins.2022.846117.

References
1.
Pals C, Sarampalis A, van Rijn H, Baskent D . Validation of a simple response-time measure of listening effort. J Acoust Soc Am. 2015; 138(3):EL187-92. DOI: 10.1121/1.4929614. View

2.
Cleary M, Pisoni D, Kirk K . Influence of voice similarity on talker discrimination in children with normal hearing and children with cochlear implants. J Speech Lang Hear Res. 2005; 48(1):204-23. PMC: 3422886. DOI: 10.1044/1092-4388(2005/015). View

3.
Green T, Faulkner A, Rosen S . Enhancing temporal cues to voice pitch in continuous interleaved sampling cochlear implants. J Acoust Soc Am. 2004; 116(4 Pt 1):2298-310. DOI: 10.1121/1.1785611. View

4.
Tinnemore A, Zion D, Kulkarni A, Chatterjee M . Children's Recognition of Emotional Prosody in Spectrally Degraded Speech Is Predicted by Their Age and Cognitive Status. Ear Hear. 2018; 39(5):874-880. PMC: 6046271. DOI: 10.1097/AUD.0000000000000546. View

5.
Luo X, Fu Q, Galvin 3rd J . Vocal emotion recognition by normal-hearing listeners and cochlear implant users. Trends Amplif. 2007; 11(4):301-15. PMC: 4111530. DOI: 10.1177/1084713807305301. View