Authors
Jihwan Lee1, Sean Foley1,2, Thanathai Lertpetchpun1, Kevin Huang1, Yoonjeong Lee1, Tiantian Feng1, Louis Goldstein2, Dani Byrd2Shrikanth Narayanan1,2
1Signal Analysis and Interpretation Laboratory, University of Southern California, USA
2Department of Linguistics, University of Southern California, USA
Abstract
We propose ARTI-6, a compact six-dimensional articulatory speech encoding framework derived from real-time MRI data that captures crucial vocal tract regions including the velum, tongue root, and larynx. ARTI-6 consists of three components: (1) a six-dimensional articulatory feature set representing key regions of the vocal tract; (2) an articulatory inversion model, which predicts articulatory features from speech acoustics leveraging speech foundation models, achieving a prediction correlation of 0.87; and (3) an articulatory synthesis model, which reconstructs intelligible speech directly from articulatory features, showing that even a low-dimensional representation can generate natural-sounding speech. Together, ARTI-6 provides an interpretable, computationally efficient, and physiologically grounded framework for advancing articulatory inversion, synthesis, and broader speech technology applications. The source code and speech samples are publicly available.
