一类深度神经网络的稳定性研究
Research on Stability for a Class of Deep Neural Networks
DOI: 10.12677/AAM.2023.127324, PDF, 下载: 131  浏览: 180 
作者: 倪炜达:浙江师范大学数学科学学院,浙江 金华
关键词: 深度神经网络动力系统稳定性Deep Neural Network Dynamical System Stability
摘要: 深度神经网络近年来逐渐成为一个研究热点,它在建模复杂的数据集上面有着突出的表现。 深度 神经网络和动力系统有着潜在的联系,如何借助动力系统理论方法深入研究深度神经网络具有重 要的理论和实际意义。 本文首先介绍了Haber等人给出的三类可逆神经网络以及稳定性定理,然 后介绍了林洁对于连续模型的稳定性做出的贡献,紧接着给出一个反例以此来说明Haber等人给 出的关于离散模型的稳定性定理不严谨,然后对定理进行优化改进,得到新的判断欧拉格式稳定 性的定理,最后将稳定性定理运用到一类哈密顿网络中。
Abstract: In recent years, deep neural networks have gradually become a research hotspot, and they have relatively good performance in modeling complex data sets. There is a potential connection between deep neural networks and dynamical systems. It is of great theoretical and practical significance to deeply study deep neural networks using dynamic system theory and methods. In this paper, firstly three types of reversible neural networks and stability theorems given by Haber et al. are introduced, and then Lin Jie’s contribution to the stability of continuous models is presented. Next, we give a counter example to illustrate that the stability theorem for discrete models given by Haber et al. is not rigorous, and then optimize and improve the theorem to obtain a new theorem for judging the stability of Euler schemes. Finally, we apply the stability theorem to a class of Hamiltonian networks.
文章引用:倪炜达. 一类深度神经网络的稳定性研究[J]. 应用数学进展, 2023, 12(7): 3250-3260. https://doi.org/10.12677/AAM.2023.127324

参考文献

[1] Weinan, E. (2017) A Proposal on Machine Learning via Dynamical Systems. Communications in Mathematics and Statistics, 5, 1-11.
https://doi.org/10.1007/s40304-017-0103-z
[2] Fan, J. and Gijbels, R. (1996) Local Polynomial Modelling and Its Applications. Chapman Hall, London.
[3] Hastie, T. (2004) The Elements of Statistical Learning: Data Mining. Journal of the Royal Statistical Society, 167, 192.
https://doi.org/10.1111/j.1467-985X.2004.298 11.x
[4] LeCun, Y., Bengio, Y. and Hinton, G. (2015) Deep Learning. Nature, 521, 436-444.
https://doi.org/10.1038/nature14539
[5] He, K., Zhang, X., Ren, S., et al. (2016) Deep Residual Learning for Image Recognition. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vegas, NV, 27-30 June 2016, 770-778.
[6] Cho, K., Merrienboer, B.V., Gulcehre, C., et al. (2014) Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, October 2014, 1724- 1734.
[7] Esteva, A., et al. (2017) Dermatologist-Level Classification of Skin Cancer with Deep Neural Networks. Nature, 542, 115-118.
https://doi.org/10.1038/nature21056
[8] Chang, B., Meng, L., Haber, E., et al. (2017) Reversible Architectures for Arbitrarily Deep Residual Neural Networks. arXiv: 1709.03698
[9] Chang, B., Chen, M. and Haber, E. (2019) Antisymmetricrnn: A Dynamical System View on Recurrent Neural Networks. arXiv: 1902.09689
[10] Haber, E. and Ruthotto, L. (2018) Stable Architectures for Deep Neural Networks. Inverse Problems, 34, Article 014004.
https://doi.org/10.1088/1361-6420/aa9a90
[11] 林沽. 深度神经网络中几类动力系统的研究[D]: [硕士学位论文]. 金华: 浙江师范大学, 2021.