期刊
ANNALS OF STATISTICS
卷 37, 期 6A, 页码 3498-3528出版社
INST MATHEMATICAL STATISTICS
DOI: 10.1214/09-AOS683
关键词
Model selection; sparse recovery; high dimensionality; concave penalty; regularized least squares; weak oracle property
资金
- NSF [DMS-08-06030, DMS-09-06784]
- USC's James H. Zumberge Faculty Research and Innovation Fund
- Direct For Mathematical & Physical Scien
- Division Of Mathematical Sciences [0806030] Funding Source: National Science Foundation
Model selection and sparse recovery are two important problems for which many regularization methods have been proposed. We study the properties of regularization methods in both problems under the unified framework of regularized least squares with concave penalties. For model selection, we establish conditions under which a regularized least squares estimator enjoys a nonasymptotic property, called the weak oracle property, where the dimensionality can grow exponentially with sample size. For sparse recovery, we present a sufficient condition that ensures the recoverability of the sparsest solution. In particular, we approach both problems by considering a family of penalties that give a smooth homotopy between L-0 and L-1 penalties. We also propose the sequentially and iteratively reweighted squares (SIRS) algorithm for sparse recovery. Numerical studies support our theoretical results and demonstrate the advantage of our new methods for model selection and sparse recovery.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据