Processing math: 45%

WeiYa's Work Yard

A traveler with endless curiosity, who fell into the ocean of statistics, tries to write down his ideas and notes to save himself.

High Dimensional Covariance Matrix Estimation

Posted on
Tags: Covariance Matrix Estimation

This note is based on Cai TT, Ren Z, Zhou HH. Estimating structured high-dimensional covariance and precision matrices: Optimal rates and adaptive estimation. Electronic Journal of Statistics. 2016;10(1):1-59..

Introduction

The standard and most natural estimator, the sample covariance matrix, performs poorly and can lead to invalid conclusions in the high-dimensional settings.

  • When p/nc(0,], the largest eigenvalue of the sample covariance matrix is not a consistent estimate of the largest eigenvalue of the population covariance matrix, and the eigenvectors of the sample covariance matrix can be nearly orthogonal to the truth.
  • When p>n, the sample covariance matrix is not invertible, and thus cannot be applied in many applications that require estimation of the precision matrix.

Overcome the difficulty by structural assumptions:

  • bandable covariance matrices
  • sparse covariance matrices
  • spiked covariance matrices
  • covariances with a tensor product structure
  • sparse precision matrices
  • bandable precision matrix via Cholesky decomposition
  • latent graphical model

Regularization methods:

  • banding method
  • tapering
  • thresholding
  • penalized likelihood estimation
  • regularized principal components
  • penalized regression for precision matrix estimation

Theoretical studies of the fundamental difficulty of the various estimation problems in terms of the minimax risks

  • the optimal rates of convergence for estimating a class of high-dimensional bandable covariance matrices under the spectral norm and Frobenious norm losses.
  • optimal estimation of sparse covariance and sparse precision matrices under a range of losses
  • optimal estimation of a Toeplitz covariance matrix
  • the minimax estimation for a large class of sparse spiked covariance matrices under the spectral norm loss.

Goal: a survey of recent optimality results on estimation of structured high-dimensional covariance and precision matrices, and discuss some key technical tools that are used in the theoretical analyses.

  • The optimal procedures for estimating the bandable, Toeplitz, and sparse covariance matrices are obtained by smoothing or thresholding the sample covariance matrices based on various sparsity assumptions.
  • In contrast, estimation of sparse spiked covariance matrices, which have sparse principal components, requires significantly different techniques to achieve optimality results.

Notation:

  • n random sample {X(1),,X(n)}
  • p-dimensional random vector X=(X1,,Xp) follows some distribution with covariance matrix Σ=(σij).
  • Goal: Estimate the covariance matrix Σ and its inverse.
  • ω operator norm:
The distribution of a random vector X is said to be sub-Gaussian with constant \rho > 0 if P\{\vert v'(X-EX)\vert > t\} \le 2e^{-t^2\rho/2}\,, for all t > 0 and all deterministic unit vector \Vert v\Vert=1.

Estimation of structured covariance matrices

Bandable covariance matrices

Sparse covariance matrices

no information on the “order” among the variables

Sparse spiked covariance matrices

Spiked covariance matrix

\Sigma = \sum_{i=1}^r\lambda_rv_iv_i'+I\,,

where \lambda_1 \ge \lambda_2 \ge \ldots \ge \lambda_r > 0 and the vector v_1,\ldots,v_r are orthonormal. Since the spectrum of \Sigma has r spikes, it was named spiked covariance model.

Estimation of structured precision matrices


Published in categories Memo