DSP-Transformer:基于注意力机制的最小支配集问题求解方法
DSP-Transformer: An Approach for Solving the Minimum Dominating Set Problem Based on Attention Mechanism
摘要: 最小支配集问题在计算机科学与网络优化领域具有广泛应用,但其NP-完全性使得传统启发式方法计算成本高且难以泛化。本文提出了DSP-Transformer,一种基于Transformer结构的深度强化学习求解方法,通过编码器–解码器架构与近似策略优化方法,高效求解最小支配集问题。由于注意力机制更适合捕获图上的潜在关系,相较于传统基于消息传递机制的编码器,本方法在性能上也有显著提升。实验结果表明,该方法在T1基准数据集和ER随机图合成数据集上的表现优于传统启发式算法,能够提供更优质的解,同时显著减少计算成本。DSP-Transformer预训练后可泛化到更大规模的图实例,无需每次重新搜索,展现出深度学习在求解组合优化问题中的巨大潜力。
Abstract: The Minimum Dominating Set Problem has widespread applications in computer science and network optimization. However, its NP-completeness makes traditional heuristic methods computationally expensive and difficult to generalize. In this paper, we propose DSP-Transformer, a deep reinforcement learning approach based on the Transformer architecture, which efficiently solves DSP through an encoder-decoder framework and approximate policy optimization. Since the attention mechanism is better suited for capturing latent relationships in graphs, our approach achieves significant performance improvements over traditional message-passing-based encoders. Experimental results demonstrate that DSP-Transformer outperforms conventional heuristic algorithms on both the T1 benchmark dataset and ER random graphs, providing higher-quality solutions while significantly reducing computational costs. Once pretrained, DSP-Transformer can generalize to larger graph instances without requiring repeated searches, highlighting the potential of deep learning in tackling combinatorial optimization problems.
文章引用:刘斯豪, 何伟骅. DSP-Transformer:基于注意力机制的最小支配集问题求解方法[J]. 应用数学进展, 2025, 14(4): 916-926. https://doi.org/10.12677/aam.2025.144216

参考文献

[1] Du, D.-Z. and Wan, P.-J. (2012) Connected Dominating Set: Theory and Applications. Springer Science & Business Media.
[2] Haynes, T. W. Hedetniemi, S. and Slater, P. (2013) Fundamentals of Domination in Graphs. CRC Press. [Google Scholar] [CrossRef
[3] Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C. and Yu, P.S. (2021) A Comprehensive Survey on Graph Neural Networks. IEEE Transactions on Neural Networks and Learning Systems, 32, 4-24. [Google Scholar] [CrossRef] [PubMed]
[4] Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., van den Driessche, G., et al. (2016) Mastering the Game of Go with Deep Neural Networks and Tree Search. Nature, 529, 484-489. [Google Scholar] [CrossRef] [PubMed]
[5] Sutton, R.S. and Barto, A.G. (2018) Reinforcement Learning: An Introduction, 2nd Edition, The MIT Press.
[6] Ore, O. (1962) Theory of Graphs. American Mathematical Society.
https://books.google.com.sg/books?id=g85UAAAAYAAJ
[7] Bondy, J.A. and Murty, U.S.R. (2010) Graph Theory. Springer.
[8] Haynes, T.W., Hedetniemi, S.T. and Henning, M.A. (2020) Topics in Domination in Graphs. Springer.
https://link.springer.com/book/10.1007/978-3-030-51117-3
[9] Tarr, J. (2010) Domination in Graphs. USF Tampa Graduate Theses and Dissertations.
https://digitalcommons.usf.edu/etd/1786
[10] Alanko, S., Crevals, S., Isopoussu, A., Östergård, P. and Pettersson, V. (2011) Computing the Domination Number of Grid Graphs. The Electronic Journal of Combinatorics, 18, 1-18. [Google Scholar] [CrossRef
[11] Anand, R., Aggarwal, D. and Kumar, V. (2017) A Comparative Analysis of Optimization Solvers. Journal of Statistics and Management Systems, 20, 623-635. [Google Scholar] [CrossRef
[12] Hopfield, J.J. and Tank, D.W. (1985) “Neural” Computation of Decisions in Optimization Problems. Biological Cybernetics, 52, 141-152. [Google Scholar] [CrossRef] [PubMed]
[13] Bello, I., Pham, H., Le, Q.V., Norouzi, M. and Bengio, S. (2016) Neural Combinatorial Optimization with Reinforcement Learning. arXiv: 1611.09940. [Google Scholar] [CrossRef
[14] Vinyals, O., Fortunato, M. and Jaitly, N. (2015) Pointer Networks. Advances in Neural Information Processing Systems 28 (NIPS 2015).
https://papers.nips.cc/paper/2015/hash/29921001f2f04bd3baee84a12e98098f-Abstract.html
[15] Deudon, M., Cournut, P., Lacoste, A., Adulyasak, Y. and Rousseau, L. (2018) Learning Heuristics for the TSP by Policy Gradient. Integration of Constraint Programming, Artificial Intelligence, and Operations Research, Delft, 26-29 June 2018, 170-181. [Google Scholar] [CrossRef
[16] Chen, X. and Tian, Y. (2019) Learning to Perform Local Rewriting for Combinatorial Optimization. Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, 8-14 December 2019, 6281-6292.
[17] Chen, M., Liu, S. and He, W. (2024) Learn to Solve Dominating Set Problem with GNN and Reinforcement Learning. Applied Mathematics and Computation, 474, Article 128717. [Google Scholar] [CrossRef
[18] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et al. (2017) Attention Is All You Need. Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, 4-9 December 2017, 6000-6010.
https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf
[19] Kool, W., van Hoof, H. and Welling, M. (2019) Attention, Learn to Solve Routing Problems! arXiv: 1803.08475. [Google Scholar] [CrossRef
[20] Schulman, J., Wolski, F., Dhariwal, P., Radford, A. and Klimov, O. (2017) Proximal Policy Optimization Algorithms. arXiv: 1707.06347. [Google Scholar] [CrossRef
[21] Erdös, P. and Rényi, A. (2011) On the Evolution of Random Graphs. In: Newman, M., Barabási, A.-L. and Watts, D.J., Eds., The Structure and Dynamics of Networks, Princeton University Press, 38-82. [Google Scholar] [CrossRef
[22] Dai, H., Khalil, E.B., Zhang, Y., Dilkina, B. and Song, L. (2017) Learning Combinatorial Optimization Algorithms over Graphs. arXiv: 1704.01665. [Google Scholar] [CrossRef
[23] Jovanovic, R., Tuba, M. and Simian, D. (2010) Ant Colony Optimization Applied to Minimum Weight Dominating Set Problem. Proceedings of the 12th WSEAS International Conference on Automatic Control, Modelling & Simulation, Catania, 29-31 May 2010, 322-326.