» Articles » PMID: 37237693

Automatic Multiple Articulator Segmentation in Dynamic Speech MRI Using a Protocol Adaptive Stacked Transfer Learning U-NET Model

Overview
Date 2023 May 27
PMID 37237693
Authors
Affiliations
Soon will be listed here.
Abstract

Dynamic magnetic resonance imaging has emerged as a powerful modality for investigating upper-airway function during speech production. Analyzing the changes in the vocal tract airspace, including the position of soft-tissue articulators (e.g., the tongue and velum), enhances our understanding of speech production. The advent of various fast speech MRI protocols based on sparse sampling and constrained reconstruction has led to the creation of dynamic speech MRI datasets on the order of 80-100 image frames/second. In this paper, we propose a stacked transfer learning U-NET model to segment the deforming vocal tract in 2D mid-sagittal slices of dynamic speech MRI. Our approach leverages (a) low- and mid-level features and (b) high-level features. The low- and mid-level features are derived from models pre-trained on labeled open-source brain tumor MR and lung CT datasets, and an in-house airway labeled dataset. The high-level features are derived from labeled protocol-specific MR images. The applicability of our approach to segmenting dynamic datasets is demonstrated in data acquired from three fast speech MRI protocols: Protocol 1: 3 T-based radial acquisition scheme coupled with a non-linear temporal regularizer, where speakers were producing French speech tokens; Protocol 2: 1.5 T-based uniform density spiral acquisition scheme coupled with a temporal finite difference (FD) sparsity regularization, where speakers were producing fluent speech tokens in English, and Protocol 3: 3 T-based variable density spiral acquisition scheme coupled with manifold regularization, where speakers were producing various speech tokens from the International Phonetic Alphabetic (IPA). Segments from our approach were compared to those from an expert human user (a vocologist), and the conventional U-NET model without transfer learning. Segmentations from a second expert human user (a radiologist) were used as ground truth. Evaluations were performed using the quantitative DICE similarity metric, the Hausdorff distance metric, and segmentation count metric. This approach was successfully adapted to different speech MRI protocols with only a handful of protocol-specific images (e.g., of the order of 20 images), and provided accurate segmentations similar to those of an expert human.

Citing Articles

Estimating Palatal and Pharyngeal Muscle Contraction in Hindi Syllable Pronunciation using Computational Modeling.

Vathulya M, Sarkar S, Singh I, Prajapati T, Sharma P Indian J Plast Surg. 2025; 57(Suppl 1):S24-S29.

PMID: 39741722 PMC: 11684914. DOI: 10.1055/s-0044-1788591.


Multi-label deep learning for comprehensive optic nerve head segmentation through data of fundus images.

Kako N, Abdulazeez A, Abdulqader D Heliyon. 2024; 10(18):e36996.

PMID: 39309959 PMC: 11416576. DOI: 10.1016/j.heliyon.2024.e36996.


A machine learning approach for vocal fold segmentation and disorder classification based on ensemble method.

Nobel S, Swapno S, Islam M, Safran M, Alfarhood S, Mridha M Sci Rep. 2024; 14(1):14435.

PMID: 38910146 PMC: 11758383. DOI: 10.1038/s41598-024-64987-5.

References
1.
Bresch E, Narayanan S . Region segmentation in the frequency domain applied to upper airway real-time magnetic resonance images. IEEE Trans Med Imaging. 2009; 28(3):323-38. PMC: 2718576. DOI: 10.1109/TMI.2008.928920. View

2.
Ruthven M, Miquel M, King A . Deep-learning-based segmentation of the vocal tract and articulators in real-time magnetic resonance images of speech. Comput Methods Programs Biomed. 2020; 198:105814. PMC: 7732702. DOI: 10.1016/j.cmpb.2020.105814. View

3.
Burdumy M, Traser L, Richter B, Echternach M, Korvink J, Hennig J . Acceleration of MRI of the vocal tract provides additional insight into articulator modifications. J Magn Reson Imaging. 2015; 42(4):925-35. DOI: 10.1002/jmri.24857. View

4.
Ha J, Sung I, Son J, Stone M, Ord R, Cho Y . Analysis of speech and tongue motion in normal and post-glossectomy speaker using cine MRI. J Appl Oral Sci. 2016; 24(5):472-480. PMC: 5083024. DOI: 10.1590/1678-775720150421. View

5.
Echternach M, Sundberg J, Arndt S, Breyer T, Markl M, Schumacher M . Vocal tract and register changes analysed by real-time MRI in male professional singers-a pilot study. Logoped Phoniatr Vocol. 2008; 33(2):67-73. DOI: 10.1080/14015430701875653. View