Estimating high-dimensional covariance and precision matrices under general missing dependence

Type: Article

Publication Date: 2021-01-01

Citations: 11

DOI: https://doi.org/10.1214/21-ejs1892

Abstract

A sample covariance matrix S of completely observed data is the key statistic in a large variety of multivariate statistical procedures, such as structured covariance/precision matrix estimation, principal component analysis, and testing of equality of mean vectors. However, when the data are partially observed, the sample covariance matrix from the available data is biased and does not provide valid multivariate procedures. To correct the bias, a simple adjustment method called inverse probability weighting (IPW) has been used in previous research, yielding the IPW estimator. The estimator can play the role of S in the missing data context, thus replacing S in off-the-shelf multivariate procedures such as the graphical lasso algorithm. However, theoretical properties (e.g. concentration) of the IPW estimator have been only established in earlier work under very simple missing structures; every variable of each sample is independently subject to missingness with equal probability. We investigate the deviation of the IPW estimator when observations are partially observed under general missing dependency. We prove the optimal convergence rate Op( logp∕n) of the IPW estimator based on the element-wise maximum norm, even when two unrealistic assumptions (known mean and/or missing probabilities) frequently assumed to be known in the past work are relaxed. The optimal rate is especially crucial in estimating a precision matrix, because of the "meta-theorem" [26] that claims the rate of the IPW estimator governs that of the resulting precision matrix estimator. In the simulation study, we discuss one of practically important issues, non-positive semi-definiteness of the IPW estimator, and compare the estimator with imputation methods.

Locations

  • arXiv (Cornell University) - View - PDF
  • Electronic Journal of Statistics - View - PDF

Similar Works

Action Title Year Authors
+ Estimating High-dimensional Covariance and Precision Matrices under General Missing Dependence 2020 Seongoh Park
Xinlei Wang
Johan Lim
+ Estimating High-dimensional Covariance and Precision Matrices under General Missing Dependence 2020 Seongoh Park
Xinlei Wang
Johan Lim
+ High-dimensional Covariance/Precision Matrix Estimation under General Missing Dependency 2020 박성오
+ Robust high-dimensional precision matrix estimation 2015 Viktoria Öllerer
Christophe Croux
+ Robust high-dimensional precision matrix estimation 2015 Viktoria Öllerer
Christophe Croux
+ New estimation methods for high dimensional inverse covariance matrices 2016 Vahe Avagyan
+ Estimating sparse precision matrices from data with missing values 2012 Mladen Kolar
Eric P. Xing
+ PDF Chat Robust High-Dimensional Precision Matrix Estimation 2014 Viktoria Oellerer
Christophe Croux
+ On the Precision Matrix in Semi-High-Dimensional Settings 2020 Kentaro Hayashi
Ke‐Hai Yuan
Ge Jiang
+ Estimating high-dimensional covariance matrices with misses for Kronecker product expansion models 2016 Mahdi Zamanighomi
Zhengdao Wang
Georgios B. Giannakis
+ A unified theory of confidence intervals for high-dimensional precision matrix 2021 Yue Wang
Yang Li
Zemin Zheng
+ Tests of Missing Completely At Random based on sample covariance matrices 2024 Alberto Bordino
Thomas B. Berrett
+ Precision Matrix Estimation with Noisy and Missing Data 2019 Roger Fan
Byoungwook Jang
Yuekai Sun
Shuheng Zhou
+ Bayesian Estimation of the Precision Matrix with Monotone Missing Data 2020 Emna Ghorbel
Kaouthar Kammoun
Mahdi Louati
+ PDF Chat Concentration of a sparse Bayesian model with Horseshoe prior in estimating high-dimensional precision matrix 2024 The Tien Mai
+ Advanced Computation of Sparse Precision Matrices for Big Data 2017 Abdelkader Baggag
Halima Bensmail
Jaideep Srivastava
+ High-dimensional covariance matrix estimation 2020 Clifford Lam
+ PDF Chat Concentration of a Sparse Bayesian Model With Horseshoe Prior in Estimating High‐Dimensional Precision Matrix 2024 The Tien Mai
+ Regularized estimation of precision matrix for high-dimensional multivariate longitudinal data 2019 Fang Qian
Yu Chen
Weiping Zhang
+ Estimation and imputation in Probabilistic Principal Component Analysis with Missing Not At Random data 2019 Aude Sportisse
Claire Boyer
Julie Josse