A rainbow in deep network black boxes

Abstract

What do neural networks learn? A major difficulty is that every training run results in a different set of weights but nevertheless leads to the same performance. We introduce a model of the probability distribution of these weights. Layers are not independent, but their dependencies can be captured by an alignment procedure. We use this model to show that networks learn the same features no matter their initialization. We also compress trained weights to a reduced set of summary statistics, from which a family of networks with equivalent performance can be reconstructed.

Publication
ArXiv

Check out the video of my talk at DeepMath 2023, or our keynote and tutorial at the CCN 2023 conference with Mick Bonner and his lab!

Florentin Guth
Florentin Guth
Postdoctoral Researcher in Deep Learning Theory

I’m a Faculty Fellow in the Center for Data Science at NYU and a Research Fellow in the Center for Computational Neuroscience at the Flatiron Institute. I’m interested in improving our scientific understanding of deep learning, e.g., understand how it escapes the curse of dimensionality, particularly in the context of image generative models.