THE SPARSE LAPLACIAN SHRINKAGE ESTIMATOR FOR HIGH-DIMENSIONAL REGRESSION

被引:84
作者
Huang, Jian [1 ]
Ma, Shuangge [2 ]
Li, Hongzhe [3 ]
Zhang, Cun-Hui [4 ]
机构
[1] Univ Iowa, Dept Stat & Actuarial Sci, Iowa City, IA 52242 USA
[2] Yale Univ, Sch Publ Hlth, Div Biostat, New Haven, CT 06520 USA
[3] Univ Penn, Sch Med, Dept Biostat & Epidemiol, Philadelphia, PA 19104 USA
[4] Rutgers State Univ, Dept Stat & Biostat, Piscataway, NJ 08854 USA
基金
美国国家科学基金会;
关键词
Graphical structure; minimax concave penalty; penalized regression; high-dimensional data; variable selection; oracle property; COORDINATE DESCENT ALGORITHMS; VARIABLE SELECTION; MODEL SELECTION; PENALIZED REGRESSION; ELASTIC-NET; LASSO; REGULARIZATION; NORMALIZATION; EXPLORATION;
D O I
10.1214/11-AOS897
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We propose a new penalized method for variable selection and estimation that explicitly incorporates the correlation patterns among predictors. This method is based on a combination of the minimax concave penalty and Laplacian quadratic associated with a graph as the penalty function. We call it the sparse Laplacian shrinkage (SLS) method. The SLS uses the minimax concave penalty for encouraging sparsity and Laplacian quadratic penalty for promoting smoothness among coefficients associated with the correlated predictors. The SLS has a generalized grouping property with respect to the graph represented by the Laplacian quadratic. We show that the SLS possesses an oracle property in the sense that it is selection consistent and equal to the oracle Laplacian shrinkage estimator with high probability. This result holds in sparse, high-dimensional settings with p >> n under reasonable conditions. We derive a coordinate descent algorithm for computing the SLS estimates. Simulation studies are conducted to evaluate the performance of the SLS method and a real data example is used to illustrate its application.
引用
收藏
页码:2021 / 2046
页数:26
相关论文
共 37 条
[1]  
[Anonymous], 1997, C BOARD MATH SCI
[2]  
[Anonymous], 2006, Journal of the Royal Statistical Society, Series B
[3]  
[Anonymous], SPARSENET COORDINATE
[4]   A comparison of normalization methods for high density oligonucleotide array data based on variance and bias [J].
Bolstad, BM ;
Irizarry, RA ;
Åstrand, M ;
Speed, TP .
BIOINFORMATICS, 2003, 19 (02) :185-193
[5]   Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR [J].
Bondell, Howard D. ;
Reich, Brian J. .
BIOMETRICS, 2008, 64 (01) :115-123
[6]   COORDINATE DESCENT ALGORITHMS FOR NONCONVEX PENALIZED REGRESSION, WITH APPLICATIONS TO BIOLOGICAL FEATURE SELECTION [J].
Breheny, Patrick ;
Huang, Jian .
ANNALS OF APPLIED STATISTICS, 2011, 5 (01) :232-253
[7]   Atomic decomposition by basis pursuit [J].
Chen, SSB ;
Donoho, DL ;
Saunders, MA .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1998, 20 (01) :33-61
[8]   Homozygosity mapping with SNP arrays identifies TRIM32 an E3 ubiquitin ligase, as a Bardet-Biedl syndrome gene (BBS11) [J].
Chiang, AP ;
Beck, JS ;
Yen, HJ ;
Tayeh, MK ;
Scheetz, TE ;
Swiderski, RE ;
Nishimura, DY ;
Braun, TA ;
Kim, KYA ;
Huang, J ;
Elbedour, K ;
Carmi, R ;
Slusarski, DC ;
Casavant, TL ;
Stone, EM ;
Sheffield, VC .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2006, 103 (16) :6287-6292
[9]  
Chung F., 2006, CBMS REGIONAL C SERI, V107
[10]   Shrinkage and model selection with correlated variables via weighted fusion [J].
Daye, Z. John ;
Jeng, X. Jessie .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2009, 53 (04) :1284-1298