TY - JOUR
T1 - Nonparametric independence screening in sparse ultra-high-dimensional additive models
AU - Fan, Jianqing
AU - Feng, Yang
AU - Song, Rui
N1 - Funding Information:
Jianqing Fan is Frederick L. Moore Professor of Finance, Department of Operations Research and Financial Engineering, Princeton University, Princeton, NJ 08544 (E-mail: [email protected]). Yang Feng is Assistant Professor, Department of Statistics, Columbia University, New York, NY 10027 (E-mail: [email protected]). Rui Song is Assistant Professor, Department of Statistics, Colorado State University, Fort Collins, CO 80523 (E-mail: [email protected]). Financial support was provided by National Science Foundation grants DMS-0714554, DMS-0704337, and DMS-1007698 and National Institutes of Health grant R01-GM072611. The authors are in deep debt to Dr. Lukas Meier for sharing the codes of penGAM. The authors thank the editor, the associate editor, and referees for their constructive comments.
PY - 2011/6
Y1 - 2011/6
N2 - A variable screening procedure via correlation learning was proposed by Fan and Lv (2008) to reduce dimensionality in sparse ultra-highdimensional models. Even when the true model is linear, the marginal regression can be highly nonlinear. To address this issue, we further extend the correlation learning to marginal nonparametric learning. Our nonparametric independence screening (NIS) is a specific type of sure independence screening. We propose several closely related variable screening procedures. We show that with general nonparametric models, under some mild technical conditions, the proposed independence screening methods have a sure screening property. The extentto which the dimensionality can be reduced by independence screening is also explicitly quantified. As a methodological extension, we also propose a data-driven thresholding and an iterative nonparametric independence screening (INIS) method to enhance the finite- sample performance for fitting sparse additive models. The simulation results and a real data analysis demonstrate that the proposed procedure works well with moderate sample size and large dimension and performs better than competing methods.
AB - A variable screening procedure via correlation learning was proposed by Fan and Lv (2008) to reduce dimensionality in sparse ultra-highdimensional models. Even when the true model is linear, the marginal regression can be highly nonlinear. To address this issue, we further extend the correlation learning to marginal nonparametric learning. Our nonparametric independence screening (NIS) is a specific type of sure independence screening. We propose several closely related variable screening procedures. We show that with general nonparametric models, under some mild technical conditions, the proposed independence screening methods have a sure screening property. The extentto which the dimensionality can be reduced by independence screening is also explicitly quantified. As a methodological extension, we also propose a data-driven thresholding and an iterative nonparametric independence screening (INIS) method to enhance the finite- sample performance for fitting sparse additive models. The simulation results and a real data analysis demonstrate that the proposed procedure works well with moderate sample size and large dimension and performs better than competing methods.
KW - Additive model
KW - Independent learning
KW - Nonparametric independence screening
KW - Nonparametric regression
KW - Sparsity
KW - Sure independence screening
KW - Variable selection
UR - http://www.scopus.com/inward/record.url?scp=79960138168&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79960138168&partnerID=8YFLogxK
U2 - 10.1198/jasa.2011.tm09779
DO - 10.1198/jasa.2011.tm09779
M3 - Article
AN - SCOPUS:79960138168
SN - 0162-1459
VL - 106
SP - 544
EP - 557
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 494
ER -