Can we build models of language acquisition from raw acoustic data in an unsupervised manner? Can deep convolutional neural networks learn to generate speech using linguistically meaningful representations? In this talk, I propose that language acquisition can be modeled with Generative Adversarial Networks (GANs) and that such modeling has implications both for the understanding of language acquisition and for the understanding of how deep neural networks learn internal representations. I propose a technique that allows us to wug-test neural networks trained on raw speech. I further propose an extension of the GAN architecture in which learning of meaningful linguistic units emerges from a requirement that the networks output informative data. With this model, we can test what the networks can and cannot learn, how their biases match human learning biases (by comparing both behavioral and neural data with networks’ outputs), how they represent linguistic structure internally, and what GAN’s innovative outputs can teach us about productivity in human language. This talk also makes a more general case for probing deep neural networks with raw speech data, as dependencies in speech are often better understood than those in the visual domain and because behavioral data on speech acquisition are relatively easily accessible
Prochains événements
Voir la liste d'événementsSRPP Beyond reaction time: Articulatory evidence of perception-production link in speech using the Stimulus-Response Compatibility paradigm.
Takayuki Nagamine (Department of Speech Hearing and Phonetic Sciences, University College London)
SRPP 13/03/2026 Christophe Corbier
Christophe Corbier (CNRS, IReMUS)
SRPP 20/03/2026 Claire Njoo
Claire Njoo (Université Paris-Sud)
SRPP 27/03/2026 Rasmus Puggaard-Rode
Rasmus Puggaard-Rode(University of Oxford)


