石跃勇等 Generalized Newton-Raphson algorithm for high dimensional LASSO regression
我校英国威廉希尔公司石跃勇老师在T2级别期刊——《Statistics and Its Interface》上发表题为“Generalized Newton-Raphson algorithm for high dimensional LASSO regression”。论文第一作者石跃勇为英国威廉希尔公司副教授。
Abstract /摘要:
The least absolute shrinkage and selection operator (LASSO) penalized regression is a state-of-the-art statistical method in high dimensional data analysis, when the number of predictors exceeds the number of observations. The commonly used Newton–Raphson algorithm is not very successful in solving the non-smooth optimization in LASSO. In this paper, we propose a fast generalized Newton–Raphson (GNR) algorithm for LASSO-type problems. The proposed algorithm, derived from a suitable Karush–Kuhn–Tucker (KKT) conditions based on generalized Newton derivatives, is a non-smooth Newton-type method. We first establish the local one-step convergence of GNR and then show that it is very efficient and accurate when coupled with a constinuation strategy. We also develop a novel parameter selection method. Numerical studies of simulated and real data analysis suggest that the GNR algorithm, with better (or comparable) accuracy, is faster than the algorithm implemented in the popular glmnet package.
论文信息;
Title/题目:
Generalized Newton-Raphson algorithm for high dimensional LASSO regression
Authors/作者:
Shi Yueyong; Huang Jian; Jiao Yuling; Kang Yicheng; Zhang Hu
Key Words /关键词:
LASSO; generalized Newton–Raphson; continuation; local one-step convergence; voting
Indexed by /核心评价:
WAJCI; SCI; Scopus;
DOI:DOI:10.4310/20-SII643
全文链接:https://dx.doi.org/10.4310/20-SII643