We release the USC Long Single-Speaker (LSS) dataset containing real-time MRI video of the vocal tract dynamics and simultaneous audio obtained during speech production. This unique dataset contains roughly one hour of video and audio data from a single native speaker of American English, making it one of the longer publicly available single-speaker datasets of real-time MRI speech data. Along with the articulatory and acoustic raw data, we release derived representations of the data that are suitable for a range of downstream tasks. This includes video cropped to the vocal tract region, sentence-level splits of the data, restored and denoised audio, and regions-of-interest timeseries. We also benchmark this dataset on articulatory synthesis and phoneme recognition tasks, providing baseline performance for these tasks on this dataset which future research can aim to improve upon.
Prochains événements
Voir la liste d'événementsSRPP Beyond reaction time: Articulatory evidence of perception-production link in speech using the Stimulus-Response Compatibility paradigm.
Takayuki Nagamine (Department of Speech Hearing and Phonetic Sciences, University College London)
SRPP 13/03/2026 Christophe Corbier
Christophe Corbier (CNRS, IReMUS)
SRPP 20/03/2026 Claire Njoo
Claire Njoo (Université Paris-Sud)
SRPP 27/03/2026 Rasmus Puggaard-Rode
Rasmus Puggaard-Rode(University of Oxford)


