异质稳健分布式支持向量回归
Heterogeneous Robust Distributed Support Vector Regression
DOI: 10.12677/aam.2026.155209, PDF,   
作者: 高国庆:青岛大学数学与统计学院,山东 青岛
关键词: 分布式个性化稳健回归异质数据Distributed Computing Personalization Robust Regression Heterogeneous Data
摘要: 在分布式算法中,由于本地数据生成机制的异构性,在开发联邦学习方法时考虑个性化非常重要。在这项工作中,本文提出了一种个性化联邦学习方法来解决鲁棒回归问题。具体来说,通过求解具有稀疏融合惩罚的平滑支持向量回归损失来学习回归权重。此外,还设计了用于鲁棒稀疏回归的个性化联邦学习(PerFL-SVR)算法,以有效地解决联邦系统中的估计问题。
Abstract: In distributed algorithms, due to the heterogeneity of local data generation mechanisms, it is crucial to consider personalization when developing federated learning methods. In this work, we propose a personalized federated learning approach to address the robust regression problem. Specifically, the regression weights are learned by optimizing a smoothed support vector regression loss function coupled with a sparse fusion penalty. Furthermore, a personalized federated learning algorithm for robust sparse regression, termed PerFL-SVR, is designed to effectively solve the estimation problem within federated systems.
文章引用:高国庆. 异质稳健分布式支持向量回归[J]. 应用数学进展, 2026, 15(5): 72-87. https://doi.org/10.12677/aam.2026.155209

参考文献

[1] Vapnik, V. (2013) The Nature of Statistical Learning Theory. Springer Science & Business Media.
[2] McMahan, B., et al. (2017) Communication-Efficient Learning of Deep Networks from Decentralized Data. Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Volume 54, 1273-1282.
[3] Liu, H., Liu, S., Ji, J., Lin, Q., Chen, J. and Tan, K.C. (2024) Personalized Federated Learning with Enhanced Implicit Generalization. 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, 30 June-5 July 2024, 1-8. [Google Scholar] [CrossRef
[4] Sang, T., Chu, Z., Xuan, J., Zhang, X. and Li, X. (2025) Personalized Federated Learning in One-Shot: A Method for Heterogeneous Data Scenarios. IEEE Internet of Things Journal, 12, 40415-40425. [Google Scholar] [CrossRef
[5] Horowitz, J.L. (1998) Bootstrap Methods for Median Regression Models. Econometrica, 66, 1327-1351. [Google Scholar] [CrossRef
[6] Pang, L., Lu, W. and Wang, H.J. (2012) Variance Estimation in Censored Quantile Regression via Induced Smoothing. Computational Statistics & Data Analysis, 56, 785-796. [Google Scholar] [CrossRef] [PubMed]
[7] Chen, X., Liu, W. and Zhang, Y. (2019) Quantile Regression under Memory Constraint. The Annals of Statistics, 47, 3244-3273. [Google Scholar] [CrossRef
[8] Ma, S., Huang, J., Zhang, Z. and Liu, M. (2020) Exploration of Heterogeneous Treatment Effects via Concave Fusion. The International Journal of Biostatistics, 16, Article ID: 20180026. [Google Scholar] [CrossRef] [PubMed]
[9] Huang, F., Chen, S. and Huang, H. (2019) Faster Stochastic Alternating Direction Method of Multipliers for Nonconvex Optimization. International Conference on Machine Learning. PMLR, Long Beach, 9-15 June 2019, 2839-2848.
[10] Ma, S. and Huang, J. (2017) A Concave Pairwise Fusion Approach to Subgroup Analysis. Journal of the American Statistical Association, 112, 410-423. [Google Scholar] [CrossRef
[11] Lu, S., Lee, J., Razaviyayn, M. and Hong, M. (2021) Linearized ADMM Converges to Second-Order Stationary Points for Non-Convex Problems. IEEE Transactions on Signal Processing, 69, 4859-4874. [Google Scholar] [CrossRef