More is less: inducing sparsity via overparameterization

Type: Article

Publication Date: 2023-03-22

Citations: 3

DOI: https://doi.org/10.1093/imaiai/iaad012

Abstract

Abstract In deep learning, it is common to overparameterize neural networks, that is, to use more parameters than training samples. Quite surprisingly training the neural network via (stochastic) gradient descent leads to models that generalize very well, while classical statistics would suggest overfitting. In order to gain understanding of this implicit bias phenomenon, we study the special case of sparse recovery (compressed sensing) which is of interest on its own. More precisely, in order to reconstruct a vector from underdetermined linear measurements, we introduce a corresponding overparameterized square loss functional, where the vector to be reconstructed is deeply factorized into several vectors. We show that, if there exists an exact solution, vanilla gradient flow for the overparameterized loss functional converges to a good approximation of the solution of minimal $\ell _1$-norm. The latter is well-known to promote sparse solutions. As a by-product, our results significantly improve the sample complexity for compressed sensing via gradient flow/descent on overparameterized models derived in previous works. The theory accurately predicts the recovery rate in numerical experiments. Our proof relies on analyzing a certain Bregman divergence of the flow. This bypasses the obstacles caused by non-convexity and should be of independent interest.

Locations

  • Information and Inference A Journal of the IMA - View
  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ More is Less: Inducing Sparsity via Overparameterization 2021 Hung-Hsu Chou
Johannes Maly
Holger Rauhut
+ Small random initialization is akin to spectral learning: Optimization and generalization guarantees for overparameterized low-rank matrix reconstruction 2021 Dominik Stöger
Mahdi Soltanolkotabi
+ Small random initialization is akin to spectral learning: Optimization and generalization guarantees for overparameterized low-rank matrix reconstruction 2021 Dominik Stöger
Mahdi Soltanolkotabi
+ Small random initialization is akin to spectral learning: Optimization and generalization guarantees for overparameterized low-rank matrix reconstruction 2021 Dominik Stöger
Mahdi Soltanolkotabi
+ PDF Chat Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks 2021 Xiangyu Chang
Yingcong Li
Samet Oymak
Christos Thrampoulidis
+ A Relaxation Argument for Optimization in Neural Networks and Non-Convex Compressed Sensing 2020 Gerrit Welper
+ Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks 2020 Xiangyu Chang
Yingcong Li
Samet Oymak
Christos Thrampoulidis
+ A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of Overparameterized Machine Learning 2021 Yehuda Dar
Vidya Muthukumar
Richard G. Baraniuk
+ A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of Overparameterized Machine Learning 2021 Yehuda Dar
V. Sai Muthukumar
Richard G. Baraniuk
+ Overparameterized Nonlinear Learning: Gradient Descent Takes the Shortest Path? 2018 Samet Oymak
Mahdi Soltanolkotabi
+ Overparameterized Nonlinear Learning: Gradient Descent Takes the Shortest Path? 2018 Samet Oymak
Mahdi Soltanolkotabi
+ Noise Regularizes Over-parameterized Rank One Matrix Recovery, Provably 2022 Tianyi Liu
Yan Li
Enlu Zhou
Tuo Zhao
+ PDF Chat Harmless interpolation of noisy data in regression 2019 V. Sai Muthukumar
Kailas Vodrahalli
Anant Sahai
+ Overfitting Can Be Harmless for Basis Pursuit, But Only to a Degree 2020 Peizhong Ju
Xiaojun Lin
Jia Liu
+ Blessing of Nonconvexity in Deep Linear Models: Depth Flattens the Optimization Landscape Around the True Solution 2022 Jianhao Ma
Salar Fattahi
+ Overparameterized ReLU Neural Networks Learn the Simplest Models: Neural Isometry and Exact Recovery 2022 Yifei Wang
Yixuan Hua
Emmanuel Candés
Mert Pilancı
+ Algorithmic Regularization in Over-parameterized Matrix Recovery 2017 Yuanzhi Li
Tengyu Ma
Hongyang Zhang
+ Implicit Balancing and Regularization: Generalization and Convergence Guarantees for Overparameterized Asymmetric Matrix Sensing 2023 Mahdi Soltanolkotabi
Dominik Stöger
Changzhi Xie
+ The Power of Preconditioning in Overparameterized Low-Rank Matrix Sensing 2023 Xingyu Xu
Yandi Shen
Yuejie Chi
Cong Ma
+ Efficient Compression of Overparameterized Deep Models through Low-Dimensional Learning Dynamics 2023 Soo Min Kwon
Zekai Zhang
Dogyoon Song
Laura Balzano
Qing Qu