Open Access Green as soon as Postprint is submitted to ZB.
Feature selection guided by structural information.
Ann. Appl. Stat. 4, 1056-1080 (2010)
In generalized linear regression problems with an abundant number of features, lasso-type regularization which imposes an `1-constraint on the regression coefficients has become a widely established technique. Crucial deficiencies of the lasso were unmasked when Zhou and Hastie (2005) introduced the elastic net. In this paper, we propose to extend the elastic net by admitting general nonnegative quadratic constraints as second form of regularization. The generalized ridge-type constraint will typically make use of the known association structure of features, e.g. by using temporal- or spatial closeness. We study properties of the resulting ’structured elastic net’ regression estimation procedure, including basic asymptotics and the issue of model selection consistency. In this vein, we provide an analog to the so-called ’irrepresentable condition’ which holds for the lasso. An oracle property is established by incorporating a scaled `1-constraint. Moreover, we outline algorithmic solutions for the structured elastic net within the generalized linear model family. The rationale and the performance of our approach is illustrated by means of simulated and real world data.
Edit extra informations Login
Publication type Article: Journal article
Document type Scientific Article
Keywords Generalized linear model; Regularization; sparsity; p => n; Llasso; Elastic net; Random fields; Consistency; Epiconvergence; Model selection; Signal regression
ISSN (print) / ISBN 1932-6157
Quellenangaben Volume: 4, Issue: 2, Pages: 1056-1080
Publisher Institute of Mathematical Statistics (IMS)
Publishing Place Cleveland, OH
Reviewing status Peer reviewed
Institute(s) Institute of Biomathematics and Biometry (IBB)