显著性目标检测轮廓增强技术研究
Refinement-Based Approach of Saliency Detection
DOI: 10.12677/CSA.2018.81014, PDF,   
作者: 庞珍珍*:四川大学计算机系,四川 成都
关键词: 显著性目标深度学习图像Saliency Object Deep Learning Image
摘要: 显著性区域检测即针对一张图像,找出其中最显著的目标。传统方法大多基于先验知识以及根据人工提取的特征来计算对比度,继而得到显著性目标。这普遍存在适应性差的问题,即对于某些场景效果比较好,对于别的场景效果则差很多。近年来深度学习算法的兴起开始应用于显著性检测,优点在于只要数据集覆盖比较全面,对于各种场景,都可以得到优于传统方法的结果。本文在原有模型的基础上,结合粗提取到细精炼两个过程,提出了新的深度学习模型。粗提取过程由两个子网络组合,以原图像作为输入,第一个子网络以超像素为单位,结合局部上下文关系得到的特征与第二个子网络在VGG中提取的高层次特征串联得到原图的显著性粗略预测。细精炼过程由一系列循环卷积层组成的,从粗糙尺度到精细尺度精炼这个粗糙的预测,最终端对端输出精度高的显著性目标区域。
Abstract: The aim of saliency detection is to find out significant regions of an image. Traditional salient ob-ject detection methods often use various prior knowledge and hand-crafted features to formulate contrast to get the saliency object, which have poor adaptability. Recently, deep learning is more and more popular in saliency detection. With a comprehensive learning set, the result will be much better than the traditional methods, especially for complex scenes. In this paper, a new deep learning model is proposed with a coarse-extracting process and fine-refining process. The coarse-extracting process contains two subnetworks. The first subnetwork’s output feature map with the local context of superpixels cascade to the second subnetwork’s high feature map extracted by VGG, then generating a coarse saliency prediction map. The fine-refining process composed of a series of recurrent convolution layers refining the coarse prediction map from coarse scales to fine scales, finally generating a fine saliency map.
文章引用:庞珍珍. 显著性目标检测轮廓增强技术研究[J]. 计算机科学与应用, 2018, 8(1): 107-113. https://doi.org/10.12677/CSA.2018.81014

参考文献

[1] Cheng, M., Mitra, N.J., Huang, X., Torr, P.H. and Hu, S. (2015) Global Contrast Based Salient Region Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37, 569-582. [Google Scholar] [CrossRef
[2] Guo, C. and Zhang, L. (2010) A Novel Multiresolution Spatiotemporal Saliency Detection Model and Its Applications in Image and Video Compression. IEEE TIP, 19, 185-198.
[3] Sun, J., Xie, J., Liu, J. and Sikora, T. (2013) Image Adaptation and Dynamic Browsing Based on Two-Layer Saliency Combination. IEEE Transactions on Broadcasting, 59, 602-613. [Google Scholar] [CrossRef
[4] Margolin, R., Zelnik-Manor, L. and Tal, A. (2013) Saliency for Image Manipulation. The Visual Computer, 29, 1-12. [Google Scholar] [CrossRef
[5] Itti, L., Koch, C. and Niebur, E. (1998) A Model of Saliency-Based Visual Attention for Rapid Scene Analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20, 1254-1259. [Google Scholar] [CrossRef
[6] Harel, J., Koch, C. and Perona, P. (2006) Graph-Based Visual Saliency. NIPS, 545-552.
[7] Achanta, R., Hemami, S., Estrada, F. and Susstrunk, S. (2009) Frequency-Tuned Salient Region Detection. 2009 IEEE Conference on Computer Vision and Pattern Recognition, June 20-25 2009, Miami, 1597-1604. [Google Scholar] [CrossRef
[8] LeCun, Y., Bottou, L., Bengio, Y. and Haffner, P. (1998) Gradient Based Learning Applied to Document Recognition. Proceedings of the IEEE, 86, 2278-2324. [Google Scholar] [CrossRef
[9] He, S., Lau, R., Liu, W., Huang, Z. and Yang, Q. (2015) Supercnn: A Superpixelwise Convolutional Neural Network for Salient Object Detection. International Journal of Computer Vision, 115, 330-344. [Google Scholar] [CrossRef
[10] Wang, L., Lu, H., Ruan, X. and Yang, M.-H. (2015) Deep Networks for Saliency Detection via Local Estimation and Global Search. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 7-12 2015, Boston. [Google Scholar] [CrossRef
[11] Liu, N. and Han, J. (2016) Dhsnet: Deep Hierarchical Saliency Network for Salient Object Detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition, June 27-30 2016, Las Vegas, 678-686. [Google Scholar] [CrossRef
[12] Simonyan, K. and Zisserman, A. (2014) Very Deep Convolutional Networks for Large-Scale Image Recognition. Computer Vision and Pattern Recognition. arXiv:1409.1556
[13] Liang, M. and Hu, X. (2015) Recurrent Convolutional Neural Network for Object Recognition. 2015 IEEE Conference on Computer Vision and Pattern Recognition, June 7-12 2015, Boston. [Google Scholar] [CrossRef
[14] Comaniciu, D. and Meer, P. (2002) Mean Shift: A Robust Approach toward Feature Space Analysis. IEEE Transactions on Pattern Analysis & Machine Intelligence, 24, 603-619. [Google Scholar] [CrossRef
[15] Krizhevsky, A., Sutskever, I. and Hinton, G.E. (2012) Imagenet Classification with Deep Convolutional Neural Networks. NIPS, 1097-1105.
[16] Deng, L.Y. (2006) The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning. Technometrics, 48. [Google Scholar] [CrossRef