基于预测校正算法的Spike and Slab Lasso逻辑回归模型
The Spike and Slab Lasso Logistic Regression Model Based on Prediction Correction Algorithm
摘要: 尽管Spike and Slab方法广泛应用于贝叶斯变量选择,但其惩罚似然估计的潜力在很大程度上被忽视了。通过在贝叶斯模态中引入惩罚化似然观点,本文提出了新的Spike and Slab Lasso逻辑回归模型,将两个拉普拉斯密度的混合先验置于单个坐标上,可以自适应地收缩系数,即弱收缩重要预测量,强收缩不相关预测量,从而可以得到准确的估计和预测。同时,我们使用了预测–校正算法来求解Spike and Slab Lasso逻辑回归模型,并将该算法扩展到不可分离惩罚的情况。该算法利用凸优化的预测校正算法,沿着整个正则化路径有效的计算解,方便了模型选择,避免了正则化参数值不同时的独立优化。最后,模拟学习和实证结果表明本文所提模型比Lasso逻辑回归模型具有更优的性能。
Abstract: Although the Spike and Slab method is widely used in Bayesian variable selection, its potential for penalized likelihood estimation is ignored. By introducing the penalized likelihood into the Bayes-ian case, this paper proposes a new Spike and Slab Lasso logistic regression model, which places the mixed priors of two Laplace densities on a single coordinate, and can adaptively adjust the shrink-age coefficient, that is, the weak shrinkage important predictor and the strong shrinkage irrelevant predictor, so that accurate estimation and prediction can be obtained. At the same time, we use the prediction-correction algorithm for the Spike and Slab Lasso logistic regression model, and extend the algorithm to the case of non-separable penalty. The algorithm uses the prediction correction algorithm of convex optimization to effectively calculate the solution along the entire regularization path, which facilitates model selection and avoids independent optimization when the regulariza-tion parameter values are different. Finally, the simulation learning and empirical results show that the proposed model has better performance than lasso logistic regression model.
文章引用:齐琪, 张齐. 基于预测校正算法的Spike and Slab Lasso逻辑回归模型[J]. 应用数学进展, 2023, 12(1): 292-307. https://doi.org/10.12677/AAM.2023.121032

参考文献

[1] Tibshirani, R. (1996) Regression Shrinkage and Selection via the Lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58, 267-288. [Google Scholar] [CrossRef
[2] Hastie, T., Tibshirani, R. and Friedman, J. (2009) The Elements of Statistical Learning. Springer, New York, 33. [Google Scholar] [CrossRef
[3] Wainwright, M. (2015) Statistical Learning with Sparsity: The Lasso and Generalizations. Chapman and Hall/CRC, London.
[4] Efron, B., Hastie, T., Johnstone, I. and Tibshirani, R. (2004) Least Angle Regression. Annals of Statistics, 32, 407-499. [Google Scholar] [CrossRef
[5] Friedman, J., Hastie, T. and Tibshirani, R. (2010) Regulariza-tion Paths for Generalized Linear Models via Coordinate Descent. Journal of Statistical Software, 33, 1-22. [Google Scholar] [CrossRef
[6] George, E.I. and McCulloch, R.E. (1993) Variable Selection via Gibbs Sampling. Journal of the American Statistical Association, 88, 881-889. [Google Scholar] [CrossRef
[7] Popovic, M. (2022) Strain Wars 4-Darwinian Evolution through Gibbs’ Glasses: Gibbs Energies of Binding and Growth Explain Evolution of SARS-CoV-2 from Hu-1 to BA.2. Virology, 575, 36-42. [Google Scholar] [CrossRef] [PubMed]
[8] Chipman, H., George, E.I., McCulloch, R.E., Clyde, M., Foster, D.P. and Stine, R.A. (2001) The Practical Implementation of Bayesian Model Selection. IMS Lecture Notes-Monograph Series, 38, 65-134. [Google Scholar] [CrossRef
[9] Ishwaran, H. and Rao, J.S. (2005) Spike and Slab Gene Selection for Multigroup Microarray Data. Journal of the American Statistical Association, 100, 764-780. [Google Scholar] [CrossRef
[10] Ishwaran, H. and Rao, J.S. (2011) Consistency of Spike and Slab Regression. Statistics & Probability Letters, 81(12), 1920-1928. [Google Scholar] [CrossRef
[11] Narisetty, N.N. and He, X. (2014) Bayesian Variable Selection with Shrinking and Diffusing Priors. Annals of Statistics, 42, 789-817. [Google Scholar] [CrossRef
[12] Ročková, V. and George, E.I. (2018) The Spike-and-Slab Lasso. Jour-nal of the American Statistical Association, 113, 431-444. [Google Scholar] [CrossRef
[13] Ročková, V. (2018) Bayesian Estimation of Sparse Signals with a Continuous Spike-and-Slab Prior. Annals of Statistics, 46, 401-437. [Google Scholar] [CrossRef
[14] Bai, R., Moran, G.E., Antonelli, J.L., Chen, Y. and Boland, M.R. (2020) Spike-and-Slab Group Lassos for Grouped Regression and Sparse Generalized Additive Models. Journal of the Ameri-can Statistical Association, 117, 184-197. [Google Scholar] [CrossRef
[15] Hu, H.-Y., Wu, D., You, Y.-Z., Olshausen, B. and Chen, Y. (2022) RG-Flow: A Hierarchical and Explainable Flow Model Based on Renormalization Group and Sparse prior. Ma-chine Learning: Science and Technology, 3, Article ID: 035009. [Google Scholar] [CrossRef
[16] Ji, Y. and Shi, H. (2022) Constrained Bayesian Doubly Elastic Net Lasso for Linear Quantile Mixed Models. Journal of Statistical Computation and Simulation, 92, 579-609. [Google Scholar] [CrossRef
[17] Muhammedrisaevna, T.M., Shukrullaevich, A.F. and Bakhriddinovna, A.N. (2021) The Logistics Approach in Managing a Tourism Company. ResearchJet Journal of Analy-sis and Inventions, 2, 231-236.
[18] Park, M.Y. and Hastie, T. (2007) L1-Regularization Path Algorithm for Generalized Linear Models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 69, 659-677. [Google Scholar] [CrossRef
[19] Lim, M. and Hastie, T. (2015) Learning Interactions via Hierarchical Group-Lasso Regularization. Journal of Computational and Graphical Statistics, 24, 627-654. [Google Scholar] [CrossRef] [PubMed]