Sparsity and smoothness via the fused lasso

被引:1783
作者
Tibshirani, R
Saunders, M
Rosset, S
Zhu, J
Knight, K
机构
[1] Stanford Univ, Dept Hlth Res & Policy, Stanford, CA 94305 USA
[2] IBM Corp, Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
[3] Univ Michigan, Ann Arbor, MI 48109 USA
[4] Univ Toronto, Toronto, ON, Canada
关键词
fused lasso; gene expression; lasso; least squares regression; protein mass spectroscopy; sparse solutions; support vector classifier;
D O I
10.1111/j.1467-9868.2005.00490.x
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The lasso penalizes a least squares regression by the sum of the absolute values (L-1-norm) of the coefficients. The form of this penalty encourages sparse solutions (with many coefficients equal to 0). We propose the 'fused lasso', a generalization that is designed for problems with features that can be ordered in some meaningful way. The fused lasso penalizes the L-1-norm of both the coefficients and their successive differences. Thus it encourages sparsity of the coefficients and also sparsity of their differences-i.e. local constancy of the coefficient profile. The fused lasso is especially useful when the number of features p is much greater than N, the sample size. The technique is also extended to the 'hinge' loss function that underlies the support vector classifier. We illustrate the methods on examples from protein mass spectroscopy and gene expression data.
引用
收藏
页码:91 / 108
页数:18
相关论文
共 23 条
[1]  
ADAM B.L, 2003, CANCER RES, V63, P3609
[2]  
[Anonymous], 1996, ASYMPTOTICS CONVEX S
[3]  
BOSER B, 1992, P COMPUTATIONAL LEAR, V2
[4]  
Chen SSB, 2001, SIAM REV, V43, P129, DOI [10.1137/S003614450037906X, 10.1137/S1064827596304010]
[5]   IDEAL SPATIAL ADAPTATION BY WAVELET SHRINKAGE [J].
DONOHO, DL ;
JOHNSTONE, IM .
BIOMETRIKA, 1994, 81 (03) :425-455
[6]  
Efron B., 2002, LEAST ANGLE REGRESSI
[7]  
Friedman J., 2001, The elements of statistical learning, V1, DOI DOI 10.1007/978-0-387-21606-5
[8]  
GILL PE, 1997, 974 NA U CAL
[9]   Molecular classification of cancer: Class discovery and class prediction by gene expression monitoring [J].
Golub, TR ;
Slonim, DK ;
Tamayo, P ;
Huard, C ;
Gaasenbeek, M ;
Mesirov, JP ;
Coller, H ;
Loh, ML ;
Downing, JR ;
Caligiuri, MA ;
Bloomfield, CD ;
Lander, ES .
SCIENCE, 1999, 286 (5439) :531-537
[10]   RIDGE REGRESSION - BIASED ESTIMATION FOR NONORTHOGONAL PROBLEMS [J].
HOERL, AE ;
KENNARD, RW .
TECHNOMETRICS, 1970, 12 (01) :55-&