基于YOLOv8的江豚性行为检测与识别模型
Yangtze Finless Porpoise Sexual Behavior Object Detection Model Based on YOLOv8
DOI: 10.12677/mos.2025.144270, PDF,   
作者: 陈慧妍, 张荣福*:上海理工大学光电信息与计算机工程学院,上海;戈潘缘元, 郝玉江:中国科学院水生生物研究所,湖北 武汉
关键词: 江豚目标检测深度学习Finless Porpoise Object Detection Deep Learning
摘要: 对江豚性行为的研究有助于此濒危物种的保育工作。传统方式依靠人眼观察,存在效率低、成本高、易疲劳等问题。随着深度学习技术的发展,中国科学院水生生物研究所配备的视觉监控设备可借助目标检测与识别技术提高观测效率。本文首先创建YFPSB数据集,包含2901张图像,涵盖五种行为类别。针对水下画质模糊、江豚与背景混淆等问题,本文建立了YOLOv8n-DBTA模型,首先通过自动伽马校正和暗通道去雾对图像预处理,接着设计双主干特征融合框架,增加更多特征融合路线,保留更多底层细节和特征以提升检测能力,然后设计任务对齐检测头,将位置与类别信息交互,学习江豚行为与位置的关系,最后进行模型实验分析,结果表明:模型在mAP0.50和mAP0.50:0.95指标分别达到97.9%和79.1%,优于其他主流模型。
Abstract: Research on the sexual behavior of Yangtze finless porpoises contributes to the conservation of this endangered species. Traditional methods rely on human observation, which is inefficient, costly, and prone to observer fatigue. With advancements in deep learning technology, the visual monitoring system at The Finless Porpoise Pavilion at the Institute of Hydrobiology, Chinese Academy of Sciences, can leverage object detection and recognition techniques to improve observation efficiency. In this study, we first constructed the YFPSB dataset, consisting of 2901 images covering five behavioral categories. To address challenges such as underwater image blurriness and the confusion between finless porpoises and the background, we establish the YOLOv8n-DBTA model. First, we apply automatic gamma correction and dark channel dehazing for image preprocessing. Then, we design a dual-branch feature fusion framework that introduces additional feature fusion paths, preserving more low-level details and features to enhance detection capability. Additionally, we develop a task-aligned detection head to enable interaction between positional and categorical information, allowing the model to learn the relationship between porpoise behavior and location. Finally, we conduct model experiments and analysis. The results demonstrate that our model achieves 97.9% and 79.1% on mAP0.50 and mAP0.50:0.95, respectively, outperforming other mainstream models.
文章引用:陈慧妍, 戈潘缘元, 郝玉江, 张荣福. 基于YOLOv8的江豚性行为检测与识别模型[J]. 建模与仿真, 2025, 14(4): 118-128. https://doi.org/10.12677/mos.2025.144270

参考文献

[1] 陈炳耀, 信誉, 路方婷, 等. 长江江豚监测现状及展望[J]. 中国环境监测, 2023, 39(2): 1-10.
[2] 于道平. 长江江豚保护生物学研究进展[J]. 安徽大学学报: 自然科学版, 2003(4): 98-103.
[3] 孙月平, 刘勇, 郭佩璇, 等. 基于改进YOLOv8n-seg的蟹塘水草区域分割与定位方法[J]. 农业工程学报, 2024, 40(17): 224-233.
[4] Ji, W., Peng, J., Xu, B. and Zhang, T. (2023) Real-Time Detection of Underwater River Crab Based on Multi-Scale Pyramid Fusion Image Enhancement and MobileCenterNet Model. Computers and Electronics in Agriculture, 204, Article 107522. [Google Scholar] [CrossRef
[5] 张铮, 鲁祥, 胡庆松. 基于图像增强与GC-YOLO v5s的水下环境河蟹识别轻量化模型研究[J]. 农业机械学报, 2024, 55(11): 124-131.
[6] 张长群, 郑杨, Platto Sara, 等. 人工饲养长江江豚的同性性行为和自慰行为[J]. 兽类学报, 2015, 35(3): 241-252.
[7] 尚志军, 张华, 曾成, 等. 基于机器视觉的枳壳自动定向方法与试验[J]. 中国农机化学报, 2019, 40(7): 119-124.
[8] Varghese, R. and M., S. (2024) YoLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, 18-19 April 2024, 1-6. [Google Scholar] [CrossRef
[9] Babakhani, P. and Zarei, P. (2015) Automatic Gamma Correction Based on Average of Brightness. Advances in Computer Science, 4, 156-159.
[10] He, K., Sun, J. and Tang, X. (2009) Single Image Haze Removal Using Dark Channel Prior. 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, 20-25 June 2009, 1956-1963. [Google Scholar] [CrossRef
[11] Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., et al. (2024) DETRs Beat YOLOs on Real-Time Object Detection. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 16-22 June 2024, 16965-16974. [Google Scholar] [CrossRef
[12] Feng, C., Zhong, Y., Gao, Y., Scott, M.R. and Huang, W. (2021) TOOD: Task-Aligned One-Stage Object Detection. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, 10-17 October 2021, 3490-3499. [Google Scholar] [CrossRef
[13] Tian, Z., Shen, C., Chen, H. and He, T. (2019) FCOS: Fully Convolutional One-Stage Object Detection. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, 27 October-2 November 2019, 9626-9635. [Google Scholar] [CrossRef
[14] Zhu, X., Hu, H., Lin, S. and Dai, J. (2019) Deformable ConvNets V2: More Deformable, Better Results. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, 15-20 June 2019, 9300-9308. [Google Scholar] [CrossRef
[15] Mohapatra, D., Das, N. and Mohanty, K.K. (2022) Deep Neural Network Based Fruit Identification and Grading System for Precision Agriculture. Proceedings of the Indian National Science Academy, 88, 228-239. [Google Scholar] [CrossRef