SRPP: Interpretable comparison between auditory brainstem response and intermediate convolutional layers in deep neural networks

Gasper Begus (UC Berkeley)
19 November 2021, 18h3020h00

Can we build models of language acquisition from raw acoustic data in an unsupervised manner? Can deep convolutional neural networks learn to generate speech using linguistically meaningful representations? In this talk, I propose that language acquisition can be modeled with Generative Adversarial Networks (GANs) and that such modeling has implications both for the understanding of language acquisition and for the understanding of how deep neural networks learn internal representations.  I propose a technique that allows us to wug-test neural networks trained on raw speech. I further propose an extension of the GAN architecture in which learning of meaningful linguistic units emerges from a requirement that the networks output informative data. With this model, we can test what the networks can and cannot learn, how their biases match human learning biases (by comparing both behavioral and neural data with networks’ outputs), how they represent linguistic structure internally, and what GAN’s innovative outputs can teach us about productivity in human language. This talk also makes a more general case for probing deep neural networks with raw speech data, as dependencies in speech are often better understood than those in the visual domain and because behavioral data on speech acquisition are relatively easily accessible

Prochains événements

Voir la liste d'événements
14 April 2023

SRPP de Mortaza Taheri-Aldari

Mortaza Taheri-Aldari (Department of English, Shahrekord University, Iran & IfL Phonetics, University of Cologne, Germany)