A model of double descent for high-dimensional binary linear classification

Type: Article

Publication Date: 2021-01-27

Citations: 68

DOI: https://doi.org/10.1093/imaiai/iaab002

Abstract

Abstract We consider a model for logistic regression where only a subset of features of size $p$ is used for training a linear classifier over $n$ training samples. The classifier is obtained by running gradient descent on logistic loss. For this model, we investigate the dependence of the classification error on the ratio $\kappa =p/n$. First, building on known deterministic results on the implicit bias of gradient descent, we uncover a phase-transition phenomenon for the case of Gaussian features: the classification error of the gradient descent solution is the same as that of the maximum-likelihood solution when $\kappa <\kappa _\star $, and that of the support vector machine when $\kappa>\kappa _\star $, where $\kappa _\star $ is a phase-transition threshold. Next, using the convex Gaussian min–max theorem, we sharply characterize the performance of both the maximum-likelihood and the support vector machine solutions. Combining these results, we obtain curves that explicitly characterize the classification error for varying values of $\kappa $. The numerical results validate the theoretical predictions and unveil double-descent phenomena that complement similar recent findings in linear regression settings as well as empirical observations in more complex learning scenarios.

Locations

  • Information and Inference A Journal of the IMA - View
  • arXiv (Cornell University) - View - PDF
  • King Abdullah University of Science and Technology Repository (King Abdullah University of Science and Technology) - View - PDF

Similar Works

Action Title Year Authors
+ A Model of Double Descent for High-dimensional Binary Linear Classification 2019 Zeyu Deng
Abla Kammoun
Christos Thrampoulidis
+ PDF Chat Analytic Study of Double Descent in Binary Classification: The Impact of Loss 2020 Ganesh Ramachandra Kini
Christos Thrampoulidis
+ Analytic Study of Double Descent in Binary Classification: The Impact of Loss 2020 Ganesh Ramachandra Kini
Christos Thrampoulidis
+ Analytic Study of Double Descent in Binary Classification: The Impact of Loss 2020 Ganesh Ramachandra Kini
Christos Thrampoulidis
+ Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability 2023 Jingfeng Wu
Vladimir Braverman
Jason D. Lee
+ Implicit bias of deep linear networks in the large learning rate phase 2020 Wei Huang
Weitao Du
Richard Yi Da Xu
Chunrui Liu
+ PDF Chat From Logistic Regression to the Perceptron Algorithm: Exploring Gradient Descent with Large Step Sizes 2024 Alexander Tyurin
+ Reconciling modern machine learning practice and the bias-variance trade-off 2018 Mikhail Belkin
Daniel Hsu
Siyuan Ma
Soumik Mandal
+ PDF Chat Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency 2024 Jingfeng Wu
Peter L. Bartlett
Matus Telgarsky
Bin Yu
+ Reconciling modern machine learning and the bias-variance trade-off 2018 Mikhail Belkin
Daniel Hsu
Siyuan Ma
Soumik Mandal
+ Gradient Descent Converges Linearly for Logistic Regression on Separable Data 2023 Kyriakos Axiotis
Maxim Sviridenko
+ Reconciling modern machine-learning practice and the classical bias–variance trade-off 2019 Mikhail Belkin
Daniel Hsu
Siyuan Ma
Soumik Mandal
+ PDF Chat Two Models of Double Descent for Weak Features 2020 Mikhail A. Belkin
Daniel Hsu
Ji Xu
+ Generalization error in high-dimensional perceptrons: Approaching Bayes error with convex optimization 2020 Benjamin Aubin
Florent Krząkała
Yue M. Lu
Lenka Zdeborová
+ Multi-scale Feature Learning Dynamics: Insights for Double Descent 2021 Mohammad Pezeshki
Amartya Mitra
Yoshua Bengio
Guillaume Lajoie
+ PDF Chat Multi-scale Feature Learning Dynamics: Insights for Double Descent 2021 Mohammad Zakaria Pezeshki
Amartya Mitra
Yoshua Bengio
Guillaume Lajoie
+ PDF Chat Double Descent Meets Out-of-Distribution Detection: Theoretical Insights and Empirical Analysis on the role of model complexity 2024 Mouïn Ben Ammar
David Brellmann
Arturo Mendoza Ramos
Antoine Manzanera
Gianni Franchi
+ PDF Chat A brief prehistory of double descent 2020 Marco Loog
Tom J. Viering
Alexander Mey
Jesse H. Krijthe
David M. J. Tax
+ A Brief Prehistory of Double Descent 2020 Marco Loog
Tom J. Viering
Alexander Mey
Jesse H. Krijthe
David M. J. Tax
+ Double Descent Demystified: Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle 2023 Rylan Schaeffer
Mikail Khona
Zachary Robertson
Akhilan Boopathy
Kateryna Pistunova
Jason W. Rocks
Ila Fiete
Oluwasanmi Koyejo