Discussion: The Dantzig selector: Statistical estimation when p is much larger than n

Type: Article

Publication Date: 2007-12-01

Citations: 32

DOI: https://doi.org/10.1214/009053607000000442

Abstract

Professors Candès and Tao are to be congratulated for their innovative and valuable contribution to high-dimensional sparse recovery and model selection.The analysis of vast data sets now commonly arising in scientific investigations poses many statistical challenges not present in smaller scale studies.Many of these data sets exhibit sparsity where most of the data corresponds to noise and only a small fraction is of interest.The needs of this research have excited much interest in the statistical community.In particular, high-dimensional model selection has attracted much recent attention and has become a central topic in statistics.The main difficulty of such a problem comes from collinearity between the predictor variables.It is clear from the geometric point of view that the collinearity increases as the dimensionality grows.A common approach taken in the statistics literature is the penalized likelihood, for example, Lasso (Tibshirani [11]) and adaptive Lasso (Zou [12]), SCAD (Fan and Li [7] and Fan and Peng [9]) and nonnegative garrote (Breiman [1]).Commonly used algorithms include LARS (Efron, Hastie, Johnstone and Tibshirani [6]), LQA (Fan and Li [7]) and MM (Hunter and Li [10]).In the present paper, Candès and Tao take a new approach, called the Dantzig selector, which uses 1 -minimization with regularization on the residuals.One promising fact is that the Dantzig selector solves a linear program, usually faster than the existing methods.In addition, the authors establish that, under the Uniform Uncertainty Principle (UUP), with large probability the Dantzig selector mimics the risk of the oracle estimator up to a logarithmic factor log p, where p denotes the number of variables.We appreciate the opportunity to comment on several aspects of this article.Our discussion here will focus on four issues: (1) connection to sparse signal recovery in the noiseless case; (2) the UUP condition and identifiability of the model;(3) computation and model selection; (4) minimax rate.

Locations

  • The Annals of Statistics - View - PDF
  • arXiv (Cornell University) - View - PDF
  • ScholarlyCommons (University of Pennsylvania) - View - PDF
  • DataCite API - View

Similar Works

Action Title Year Authors
+ PDF Chat The Dantzig selector: Statistical estimation when p is much larger than n 2007 Emmanuel J. Candès
Terence Tao
+ PDF Chat Discussion: A tale of three cousins: Lasso, L2Boosting and Dantzig 2007 Nicolai Meinshausen
Guilherme V. Rocha
Bin Yu
+ PDF Chat Model Selection With Lasso-Zero: Adding Straw to the Haystack to Better Find Needles 2021 Pascaline Descloux
Sylvain Sardy
+ Model Selection With Lasso-Zero: Adding Straw to the Haystack to Better Find Needles 2021 Pascaline Descloux
Sylvain Sardy
+ The constrained Dantzig selector with enhanced consistency 2016 Yinfei Kong
Zemin Zheng
Jinchi Lv
+ Adapting to Unknown Sparsity by controlling the False Discovery Rate 2005 Felix Abramovich
Yoav Benjamini
David L. Donoho
Iain M. Johnstone
+ Simultaneous Lasso and Dantzig Selector in High Dimensional Nonparametric Regression 2013 Shiqing Wang
Limin Su
+ Thresholding Procedures for High Dimensional Variable Selection and Statistical Estimation 2009 Shuheng Zhou
+ Lasso Degrees of Freedom when p > n 2011 Ryan J. Tibshirani
Jonathan Taylor
+ Model selection with lasso-zero: adding straw to the haystack to better find needles 2018 Pascaline Descloux
Sylvain Sardy
+ PDF Chat Randomized pick-freeze for sparse Sobol indices estimation in high dimension 2015 Yohann de Castro
Alexandre Janon
+ PDF Chat Adapting to unknown sparsity by controlling the false discovery rate 2006 Felix Abramovich
Yoav Benjamini
David L. Donoho
Iain M. Johnstone
+ The Squared-Error of Generalized LASSO: A Precise Analysis 2013 Samet Oymak
Christos Thrampoulidis
Babak Hassibi
+ The Squared-Error of Generalized LASSO: A Precise Analysis 2013 Samet Oymak
Christos Thrampoulidis
Babak Hassibi
+ Thresholded Lasso for high dimensional variable selection and statistical estimation 2010 Shuheng Zhou
+ Thresholded Lasso for high dimensional variable selection 2023 Shuheng Zhou
+ Variable Selection with Big Data based on Zero Norm and via Sequential Monte Carlo 2019 Jin‐Chuan Duan
+ Adaptive post-Dantzig estimation and prediction for non-sparse "large $p$ and small $n$" models 2010 Lu Lin
Lixing Zhu
Yujie Gai
+ optimal procedures in high-dimensional variable selection 2013 Qi Zhang
+ Subset Selection with Shrinkage: Sparse Linear Modeling when the SNR is low 2017 Rahul Mazumder
Peter Radchenko
Antoine Dedieu