基于光学可调非线性函数的目标分类
Target Recognition Based on Optically Tunable Nonlinear Functions
DOI: 10.12677/mos.2025.145424, PDF,   
作者: 汪嘉伟, 李一鸣, 张启明*:上海理工大学光子芯片研究院,上海;上海理工大学光电信息与计算机工程学院,人工智能纳米光子学中心,上海
关键词: 光学计算非线性计算衍射神经网络Optical Computing Nonlinear Computing Diffractive Neural Network
摘要: 非线性函数在神经网络中扮演着重要的角色,通过对神经元引入非线性因素,从而使得神经网络得以逼近任何非线性函数,但随着网络本身规模的增加也导致训练时间和算力成本过高。尽管目前已经提出基于电子计算的优化方法,却仍存在可拓展性差、计算自由度单一等问题。因此,文章提出了一种利用大量可编程光学神经元的光子计算作为解决方法,以高鲁棒性、高精确度方法进行Tanh和Sigmoid非线性计算。实验结果表明,所提出的基于衍射所构建的非线性计算模块在神经网络的应用方面具有相对较高的泛化能力。对于5 × 5数据输入维度,Sigmoid函数实现了9.66 × 10−4的误差损失,Tanh函数实现了1.02 × 10−3的损失。在MNIST手写数字分类任务上,通过实验验证了架构在面对传输中高斯噪声等现实因素干扰时具有相当的稳定性和鲁棒性。
Abstract: Nonlinear functions play a crucial role in neural networks by introducing nonlinearity into neurons, enabling the network to approximate arbitrary nonlinear mappings. However, as the scale of neural networks continues to grow, the associated training time and computational costs have become increasingly prohibitive. Although several optimization strategies based on electronic computing have been proposed, they still suffer from limited scalability and restricted computational flexibility. To address these challenges, this paper proposes a photonic computing approach based on a large number of programmable optical neurons to perform nonlinear computations with high robustness and precision. Specifically, nonlinear functions such as Tanh and Sigmoid are implemented using a diffractive optical structure. Experimental results demonstrate that the proposed diffraction-based nonlinear computing module exhibits strong generalization capabilities when applied to neural network tasks. For a 5 × 5 input dimension, the Sigmoid function achieved a loss of 9.66 × 104, while the Tanh function yielded a loss of 1.02 × 103. Furthermore, evaluation on the MNIST handwritten digit classification task shows that the proposed architecture maintains considerable stability and robustness in the presence of practical disturbances such as Gaussian noise during transmission.
文章引用:汪嘉伟, 李一鸣, 张启明. 基于光学可调非线性函数的目标分类[J]. 建模与仿真, 2025, 14(5): 670-681. https://doi.org/10.12677/mos.2025.145424

参考文献

[1] Lecun, Y., Chopra, S., Hadsell, R., Ranzato, M. and Huang, F.J. (2006) A Tutorial on Energy-Based Learning. In: Bakir, G., Hofman, T., Schölkopf, B., Smola, A., Taskar, B., Eds., Predicting Structured Data, The MIT Press. [Google Scholar] [CrossRef
[2] Vaswani, A., Shazeer, N., et al. (2017) Attention Is All You Need.
https://ui.adsabs.harvard.edu/abs/2017arXiv170603762V>;
[3] LeCun, Y., Bengio, Y. and Hinton, G. (2015) Deep Learning. Nature, 521, 436-444. [Google Scholar] [CrossRef] [PubMed]
[4] Kitaev, N., Kaiser, Ł. and Levskaya, A. (2020) Reformer: The Efficient Transformer.
https://ui.adsabs.harvard.edu/abs/2020arXiv200104451K>;
[5] Zaremba, W., Sutskever, I. and Vinyals, O. (2014) Recurrent Neural Network Regularization.
https://ui.adsabs.harvard.edu/abs/2014arXiv1409.2329Z
[6] Liu, F., Zhang, B., Chen, G., Gong, G., Lu, H. and Li, W. (2021) A Novel Configurable High-Precision and Low-Cost Circuit Design of Sigmoid and Tanh Activation Function. 2021 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA), Zhuhai, 24-26 November 2021, 222-223. [Google Scholar] [CrossRef
[7] Pogiri, R., Ari, S. and Mahapatra, K.K. (2022) Design and FPGA Implementation of the LUT Based Sigmoid Function for DNN Applications. 2022 IEEE International Symposium on Smart Electronic Systems (iSES), Warangal, 18-22 December 2022, 410-413. [Google Scholar] [CrossRef
[8] Moposita, T., Trojman, L., Crupi, F., Lanuzza, M. and Vladimirescu, A. (2022) Voltage-to-Voltage Sigmoid Neuron Activation Function Design for Artificial Neural Networks. 2022 IEEE 13th Latin America Symposium on Circuits and System (LASCAS), Puerto Varas, 1-4 March 2022, 1-4. [Google Scholar] [CrossRef
[9] Xing, S. and Wu, C. (2020) Implementation of a Neuron Using Sigmoid Activation Function with CMOS. 2020 IEEE 5th International Conference on Integrated Circuits and Microsystems (ICICM), Nanjing, 23-25 October 2020, 201-204. [Google Scholar] [CrossRef
[10] Kagalkar, A. and Raghuram, S. (2020) CORDIC Based Implementation of the Softmax Activation Function. 2020 24th International Symposium on VLSI Design and Test (VDAT), Bhubaneswar, 23-25 July 2020, 1-4. [Google Scholar] [CrossRef
[11] Yang, Z., He, J., Yan, Z., Hu, Y., Li, X., Dong, N., et al. (2024) Inverse-Designed Integrated All-Optical Nonlinear Activators for Optical Computing. Optics Express, 32, Article 34001. [Google Scholar] [CrossRef] [PubMed]
[12] Yan, T., Wu, J., Zhou, T., Xie, H., Xu, F., Fan, J., et al. (2019) Fourier-Space Diffractive Deep Neural Network. Physical Review Letters, 123, Article 023901. [Google Scholar] [CrossRef] [PubMed]
[13] Zhao, Y., Chen, Z., Ding, Z., Ren, H. and Nelson, J.S. (2002) Real-Time Phase-Resolved Functional Optical Coherence Tomography by Use of Optical Hilbert Transformation. Optics Letters, 27, 98-100. [Google Scholar] [CrossRef] [PubMed]
[14] Zhao, Z., Wang, Y., Guan, C., Zhang, K., Wu, Q., Li, H., et al. (2022) Deep Learning-Enabled Compact Optical Trigonometric Operator with Metasurface. PhotoniX, 3, Article No. 15. [Google Scholar] [CrossRef
[15] You, J., Han, Z., Zhang, N., Zhang, Q., Zhang, Y., Liu, Y., et al. (2024) All-Optic Logical Operations Based on the Visible-Near Infrared Bipolar Optical Response. Advanced Science, 11, Article 2404336. [Google Scholar] [CrossRef] [PubMed]
[16] Lin, X., Rivenson, Y., Yardimci, N.T., Veli, M., Luo, Y., Jarrahi, M., et al. (2018) All-Optical Machine Learning Using Diffractive Deep Neural Networks. Science, 361, 1004-1008. [Google Scholar] [CrossRef] [PubMed]
[17] Wang, T., Sohoni, M.M., Wright, L.G., Stein, M.M., Ma, S., Onodera, T., et al. (2023) Image Sensing with Multilayer Nonlinear Optical Neural Networks. Nature Photonics, 17, 408-415. [Google Scholar] [CrossRef
[18] Li, Y., Li, J. and Ozcan, A. (2024) Nonlinear Encoding in Diffractive Information Processing Using Linear Optical Materials. Light: Science & Applications, 13, Article No. 173. [Google Scholar] [CrossRef] [PubMed]
[19] Dou, H., Deng, Y., Yan, T., Wu, H., Lin, X. and Dai, Q. (2020) Residual D2NN: Training Diffractive Deep Neural Networks via Learnable Light Shortcuts. Optics Letters, 45, 2688-2691. [Google Scholar] [CrossRef] [PubMed]
[20] Li, J., Gan, T., Bai, B., Luo, Y., Jarrahi, M. and Ozcan, A. (2023) Massively Parallel Universal Linear Transformations Using a Wavelength-Multiplexed Diffractive Optical Network. Advanced Photonics, 5, Article 016003. [Google Scholar] [CrossRef
[21] Zhou, T., Lin, X., Wu, J., Chen, Y., Xie, H., Li, Y., et al. (2021) Large-Scale Neuromorphic Optoelectronic Computing with a Reconfigurable Diffractive Processing Unit. Nature Photonics, 15, 367-373. [Google Scholar] [CrossRef
[22] Matsushima, K. and Shimobaba, T. (2009) Band-Limited Angular Spectrum Method for Numerical Simulation of Free-Space Propagation in Far and Near Fields. Optics Express, 17, 19662-19673.