Asymptotics of representation learning in finite Bayesian neural networks*

Type: Article

Publication Date: 2022-11-01

Citations: 19

DOI: https://doi.org/10.1088/1742-5468/ac98a6

Abstract

Abstract Recent works have suggested that finite Bayesian neural networks may sometimes outperform their infinite cousins because finite networks can flexibly adapt their internal representations. However, our theoretical understanding of how the learned hidden layer representations of finite networks differ from the fixed representations of infinite networks remains incomplete. Perturbative finite-width corrections to the network prior and posterior have been studied, but the asymptotics of learned features have not been fully characterized. Here, we argue that the leading finite-width corrections to the average feature kernels for any Bayesian network with linear readout and Gaussian likelihood have a largely universal form. We illustrate this explicitly for three tractable network architectures: deep linear fully-connected and convolutional networks, and networks with a single nonlinear hidden layer. Our results begin to elucidate how task-relevant learning signals shape the hidden layer representations of wide Bayesian neural networks.

Locations

  • Journal of Statistical Mechanics Theory and Experiment - View
  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Asymptotics of representation learning in finite Bayesian neural networks 2021 Jacob A. Zavatone-Veth
AbdĂźlkadir Canatar
Benjamin S. Ruben
Cengiz Pehlevan
+ Exact marginal prior distributions of finite Bayesian neural networks 2021 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ Exact priors of finite neural networks. 2021 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ PDF Chat Feature learning in finite-width Bayesian deep linear networks with multiple outputs and convolutional layers 2024 Federico Bassetti
Marco Gherardi
Alessandro Ingrosso
Mauro Pastore
Pietro Rotondo
+ PDF Chat Kernel shape renormalization explains output-output correlations in finite Bayesian one-hidden-layer networks 2024 Piero Baglioni
Lorenzo Giambagli
A. Vezzani
R. Burioni
Pietro Rotondo
Rosalba Pacelli
+ PDF Chat Contrasting random and learned features in deep Bayesian linear regression 2022 Jacob A. Zavatone-Veth
William L. Tong
Cengiz Pehlevan
+ Predicting the Outputs of Finite Networks Trained with Noisy Gradients 2021 Gadi Naveh
Oded Ben-David
Haim Sompolinsky
Zohar Ringel
+ An Infinite-Feature Extension for Bayesian ReLU Nets That Fixes Their Asymptotic Overconfidence 2020 Agustinus Kristiadi
Matthias Hein
Philipp Hennig
+ PDF Chat Depth induces scale-averaging in overparameterized linear Bayesian neural networks 2021 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ Predicting the outputs of finite deep neural networks trained with noisy gradients 2020 Gadi Naveh
Oded Ben-David
Haim Sompolinsky
Zohar Ringel
+ An Infinite-Feature Extension for Bayesian ReLU Nets That Fixes Their Asymptotic Overconfidence 2020 Agustinus Kristiadi
Matthias Hein
Philipp Hennig
+ PDF Chat Predicting the outputs of finite deep neural networks trained with noisy gradients 2021 Gadi Naveh
Oded Ben David
Haim Sompolinsky
Zohar Ringel
+ A theory of representation learning gives a deep generalisation of kernel methods 2021 Adam X. Yang
Maxime Robeyns
Edward Milsom
Nandi Schoots
Laurence Aitchison
+ PDF Chat Proportional infinite-width infinite-depth limit for deep linear neural networks 2024 Federico Bassetti
Lucia Ladelli
Pietro Rotondo
+ Bayesian Interpolation with Deep Linear Networks 2022 Boris Hanin
Alexander Zlokapa
+ Bayesian interpolation with deep linear networks 2023 Boris Hanin
Alexander Zlokapa
+ Deep Neural Networks as Gaussian Processes 2017 Jaehoon Lee
Yasaman Bahri
Roman Novak
Samuel S. Schoenholz
Jeffrey Pennington
Jascha Sohl‐Dickstein
+ Deep Neural Networks as Gaussian Processes 2018 Jaehoon Lee
Yasaman Bahri
Roman Novak
Samuel S. Schoenholz
Jeffrey Pennington
Jascha Sohl‐Dickstein
+ Wide Neural Networks with Bottlenecks are Deep Gaussian Processes 2020 Devanshu Agrawal
Theodore Papamarkou
Jacob Hinkle
+ PDF Chat Critical feature learning in deep neural networks 2024 Kirsten Fischer
Javed Lindner
David Dahmen
Zohar Ringel
M. Krämer
Moritz Helias

Works That Cite This (11)

Action Title Year Authors
+ A statistical mechanics framework for Bayesian deep neural networks beyond the infinite-width limit 2022 R. Pacelli
S. Ariosto
Mauro Pastore
Francesco Ginelli
Marco Gherardi
Pietro Rotondo
+ PDF Chat Replica method for eigenvalues of real Wishart product matrices 2023 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ A statistical mechanics framework for Bayesian deep neural networks beyond the infinite-width limit 2023 R. Pacelli
S. Ariosto
Mauro Pastore
Francesco Ginelli
Marco Gherardi
Pietro Rotondo
+ PDF Chat Asymptotics of representation learning in finite Bayesian neural networks* 2022 Jacob A. Zavatone-Veth
AbdĂźlkadir Canatar
Benjamin S. Ruben
Cengiz Pehlevan
+ Learning curves for deep structured Gaussian feature models<sup>*</sup> 2024 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ PDF Chat Neural network field theories: non-Gaussianity, actions, and locality 2023 Mehmet Demirtaş
James Halverson
Anindita Maiti
Matthew D. Schwartz
Keegan Stoner
+ Resolution of similar patterns in a solvable model of unsupervised deep learning with structured data 2024 Andrea Baroffio
Pietro Rotondo
Marco Gherardi
+ PDF Chat On Neural Network Kernels and the Storage Capacity Problem 2022 Jacob A. Zavatone-Veth
Cengiz Pehlevan
+ A fast point solver for deep nonlinear function approximators. 2021 Laurence Aitchison
+ A Primer on Bayesian Neural Networks: Review and Debates 2023 Julyan Arbel
Konstantinos Pitas
Mariia Vladimirova
Vincent Fortuin