基于生成对抗网络的街景图像修复模型
A Streetview Image Inpainting Method Based on Semantic Prior Guidance
DOI: 10.12677/mos.2024.136541, PDF,    科研立项经费支持
作者: 朱 莹:南京邮电大学理学院,江苏 南京
关键词: 街景图像修复语义先验生成对抗网络Streetview Image Inpainting Semantic Prior Generative Adversarial Network
摘要: 针对现有街景图像修复方法难以有效处理复杂遮挡和保持语义一致性的问题,本文提出了一种基于生成对抗网络(GAN)的图像修复模型,旨在利用语义信息提升修复效果。该模型能够有效地提取和处理不同类别的语义信息,并将其融入图像修复过程,利用语义信息引导生成器生成更逼真、更符合语义的修复结果。在街景数据集上的实验结果表明,相较于传统方法,本文提出的模型能够生成更加自然逼真的修复结果,可以精准地还原车辆、道路、建筑物等物体的轮廓、纹理细节和颜色,有效消除图像缺失部分带来的视觉突兀感,显著提升图像的整体质量。未来可进一步拓展该模型的应用范围,将其应用于人脸图像以及其他类型场景的图像修复任务中,以期提升图像修复模型的泛化能力和修复效果。
Abstract: This paper proposes a novel generative adversarial network (GAN)-based image inpainting model for street view images, aiming to leverage semantic information to enhance inpainting quality. The model effectively extracts and processes different categories of semantic information, integrating them into the inpainting process to guide the generator in producing more realistic and semantically consistent results. Experimental results on street view datasets demonstrate that, compared to traditional methods, our proposed model generates more natural and visually appealing inpainted images. It accurately restores the contours, textures, and colors of objects such as vehicles, roads, and buildings, effectively eliminating the visual abruptness caused by missing parts and significantly improving the overall image quality. In the future, we plan to extend the application of this model to face images and other types of scenes to improve the generalization ability and inpainting performance of image inpainting models, enabling them to better address diverse image inpainting needs across various scenarios.
文章引用:朱莹. 基于生成对抗网络的街景图像修复模型[J]. 建模与仿真, 2024, 13(6): 5934-5941. https://doi.org/10.12677/mos.2024.136541

参考文献

[1] Goodfellow, I., Pouget-Abadie, J., Mirza, M., et al. (2014) Generative Adversarial Nets. Proceedings of the 27th International Conference on Neural Information Processing Systems, 2, 2672-2680.
[2] 朱立忠, 佟昕. 基于深度学习的图像双分支修复算法[J]. 通信与信息技术, 2024(5): 14-18+55.
[3] 徐嘉悦, 赵建平, 李冠男, 等. 级联式生成对抗网络的全景图像修复[J]. 重庆理工大学学报(自然科学), 2024, 38(8): 154-163.
[4] Nazeri, K., Ng, E., Joseph, T., Qureshi, F. and Ebrahimi, M. (2019) EdgeConnect: Structure Guided Image Inpainting Using Edge Prediction. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, 27-28 October 2019, 3265-3274. [Google Scholar] [CrossRef
[5] Xiang, H., Min, W., Han, Q., Zha, C., Liu, Q. and Zhu, M. (2024) Structure-Aware Multi-View Image Inpainting Using Dual Consistency Attention. Information Fusion, 104, Article ID: 102174. [Google Scholar] [CrossRef
[6] Szegedy, C., Ioffe, S., Vanhoucke, V. and Alemi, A. (2017) Inception-v4, Inception-Resnet and the Impact of Residual Connections on Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 31, 4278-4284. [Google Scholar] [CrossRef
[7] Wang, T., Liu, M., Zhu, J., Tao, A., Kautz, J. and Catanzaro, B. (2018) High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, 18-23 June 2018, 8798-8807. [Google Scholar] [CrossRef
[8] Zhang, R., Isola, P., Efros, A.A., Shechtman, E. and Wang, O. (2018) The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, 18-23 June 2018, 586-595. [Google Scholar] [CrossRef
[9] Song, Y., Yang, C., Shen, Y., et al. (2018) SPG-Net: Segmentation Prediction and Guidance Network for Image Inpainting.
[10] Zhang, R., Quan, W., Zhang, Y., Wang, J. and Yan, D. (2023) W-Net: Structure and Texture Interaction for Image Inpainting. IEEE Transactions on Multimedia, 25, 7299-7310. [Google Scholar] [CrossRef