高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

UWF-YOLO: 冗余信息优化的轻量化水下目标检测

侯国家 马佳琦 王岳川 黄宝香 李坤乾

侯国家, 马佳琦, 王岳川, 黄宝香, 李坤乾. UWF-YOLO: 冗余信息优化的轻量化水下目标检测[J]. 电子与信息学报. doi: 10.11999/JEIT251129
引用本文: 侯国家, 马佳琦, 王岳川, 黄宝香, 李坤乾. UWF-YOLO: 冗余信息优化的轻量化水下目标检测[J]. 电子与信息学报. doi: 10.11999/JEIT251129
HOU Guojia, MA Jiaqi, WANG Yuechuan, HUANG Baoxiang, LI Kunqian. UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251129
Citation: HOU Guojia, MA Jiaqi, WANG Yuechuan, HUANG Baoxiang, LI Kunqian. UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251129

UWF-YOLO: 冗余信息优化的轻量化水下目标检测

doi: 10.11999/JEIT251129 cstr: 32379.14.JEIT251129
基金项目: 国家自然科学基金 (62371431, 61901240),青岛市自然科学基金 (24-4-4-zrjj-122-jch),山东省自然科学基金 (ZR2024MF125, ZR2025QB60)
详细信息
    作者简介:

    侯国家:男,博士,副教授,研究方向为图像/视频处理、海洋视觉感知与理解

    马佳琦:女,博士生,研究方向为计算机视觉

    王岳川:男,硕士生,研究方向为图像处理

    黄宝香:女,博士,教授,研究方向为大数据海洋学、遥感图像处理、人工智能

    李坤乾:男,博士,副教授,研究方向为计算机视觉,海洋装备人工智能技术

    通讯作者:

    李坤乾 likunqian@ouc.edu.cn

  • 中图分类号: TP391.41

UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization

Funds: National Natural Science Foundation of China (62371431, 61901240), Qingdao Natural Science Foundation (24-4-4-zrjj-122-jch), Natural Science Foundation of Shandong Province, China (ZR2024MF125, ZR2025QB60)
  • 摘要: 针对现有水下目标检测方法在成像退化类型多样与背景干扰等复杂场景中鲁棒性差以及在设备资源受限条件下难以兼顾检测精度与模型轻量化的问题,本文提出基于冗余信息优化的轻量化水下目标检测网络(Underwater Faster YOLO Network Based on Redundancy Information Optimization, UWF-YOLO),并进一步构建了复杂场景水下目标检测数据集(Underwater Object Detection Dataset with Complex Scene, CSUOD)。UWF-YOLO采用FasterNet Block重构C2f模块优化主干和颈部网络,通过特征通道选择机制减少冗余特征,并引入Ghost卷积增强颈部网络的多尺度特征融合能力;同时,通过基于分组卷积的参数共享检测头降低计算开销;最后,应用结构化通道剪枝技术进一步压缩网络规模。CSUOD数据集通过收集真实水下图像标注并进行分辨率标准化处理,覆盖雾化、色偏、非均匀照明等各种退化类型,可用于复杂场景下水下目标检测模型的鲁棒性训练与性能评测。在DUO,RUOD和TrashCan数据集上进行实验表明,相较于YOLOv8s,所提方法在计算量、权重大小与参数量三个指标上的分别降低了60.4%、77.3%和78.4%;与参数量相当的YOLOv9-tiny相比,mAP指标在三个数据集上分别提升了0.3%、2.3%和3.4%。同时,在自建CSUOD数据集上的主客观对比实验,进一步证实所提模型在实现显著轻量化的同时,能够有效避免背景干扰导致的误检、漏检等问题,特别在复杂水下环境中展现出优异的检测性能。此外,本文构建的复杂场景水下数据集将有助于推动水下目标检测方法的发展。
  • 图  1  UWF-YOLO网络整体架构

    图  2  主干网络与高效跨阶段特征优化模块(ECFO)

    图  3  颈部网络与Ghost卷积

    图  4  冗余优化分组检测头(RRG-Head)

    图  5  CSUOD数据集部分样例

    图  6  数据集目标数量与类别比例分布

    图  7  各目标检测网络在不同数据集上的可视化对比结果

    表  1  DUO、RUOD、TrashCan数据集客观指标对比

    网络mAP50(%)FLOPs(G)Weight(M)Params(M)
    DUORUODTrashCanDUORUODTrashCanDUORUODTrashCanDUORUODTrashCan
    YOLOv5s82.385.989.215.815.815.914.414.514.57.07.07.1
    Deformable DETR80.183.381.451.151.151.1480.0480.0480.039.839.839.8
    YOLOv7-tiny83.684.684.713.013.113.212.412.312.46.06.06.1
    YOLOv8s83.186.589.628.428.528.522.522.622.511.111.111.1
    YOLOv9-tiny82.884.386.010.710.710.76.16.16.12.62.62.6
    UWF-YOLO
    (剪枝前)
    84.287.090.119.519.519.514.714.714.77.27.27.2
    UWF-YOLO83.186.689.411.311.411.35.15.05.12.42.42.4
    下载: 导出CSV

    表  2  自建CSUOD数据集指标对比

    网络 mAP50_all (%) AP_j (%) AP_c (%) AP_f (%) AP_s (%) AP_d (%) AP_t (%)
    YOLOv5s 78.4 86.7 81.5 76.9 50.8 87.1 87.8
    YOLOv7-tiny 65.3 56.6 70.2 74.0 52.6 72.6 65.6
    YOLOv7 79.6 89.6 82.5 80.5 61.6 81.6 81.8
    YOLOv8s 78.0 85.9 72 79.6 64.2 76.4 90
    YOLOv9-tiny 65.1 76.5 63.1 64.9 45.6 65.9 74.4
    UWF-YOLO 79.6 84.2 79.8 75.2 70.5 84.2 84.1
    下载: 导出CSV

    表  3  消融实验

    数据集 方案 基线模型 ECFO Ghost RRG-Head 通道剪枝 mAP50 (%) Params (M) FLOPs (G) Weight (M)
    RUOD 1 86.5 11.1 28.5 22.6
    2 86.7 9.4 27.0 19.0
    3 86.7 9.0 26.6 18.3
    4 87.0 7.2 19.5 14.7
    5 86.6 2.4 11.4 5.0
    TrashCan 1 89.6 11.1 28.5 22.5
    2 90.0 9.4 27.1 19.0
    3 90.2 9.0 26.6 18.3
    4 90.1 7.2 19.5 14.7
    5 89.4 2.4 11.3 5.1
    下载: 导出CSV
  • [1] 黄海宁, 李宝奇, 刘纪元, 等. 声呐图像水下目标识别综述与展望[J]. 电子与信息学报, 2024, 46(5): 1742–1760. doi: 10.11999/JEIT231207.

    HUANG Haining, LI Baoqi, LIU Jiyuan, et al. Sonar image underwater target recognition: A comprehensive overview and prospects[J]. Journal of Electronics & Information Technology, 2024, 46(5): 1742–1760. doi: 10.11999/JEIT231207.
    [2] WANG Hao, ZHANG Weibo, XU Yinghao, et al. WaterCycleDiffusion: Visual-textual fusion empowered underwater image enhancement[J]. Information Fusion, 2025, 127: 103693. doi: 10.1016/j.inffus.2025.103693.
    [3] ZHANG Dehua, YU Changcheng, LI Zhen, et al. A lightweight network enhanced by attention-guided cross-scale interaction for underwater object detection[J]. Applied Soft Computing, 2025, 184: 113811. doi: 10.1016/j.asoc.2025.113811.
    [4] CHEW A L, TONG P B, and CHIA C S. Automatic detection and classification of man-made targets in side scan sonar images[C]. 2007 Symposium on Underwater Technology and Workshop on Scientific Use of Submarine Cables and Related Technologies, Tokyo, Japan, 2007: 126–132. doi: 10.1109/UT.2007.370841.
    [5] BEIJBOM O, EDMUNDS P J, KLINE D I, et al. Automated annotation of coral reef survey images[C]. 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, USA, 2012: 1170–1177. doi: 10.1109/CVPR.2012.6247798.
    [6] LI Xiu, SHANG Min, QIN Hongwei, et al. Fast accurate fish detection and recognition of underwater images with fast R-CNN[C]. OCEANS 2015-MTS/IEEE Washington, Washington, USA, 2015: 1–5. doi: 10.23919/OCEANS.2015.7404464.
    [7] SONG Pinhao, LI Pengteng, DAI Linhui, et al. Boosting R-CNN: Reweighting R-CNN samples by RPN’s error for underwater object detection[J]. Neurocomputing, 2023, 530: 150–164. doi: 10.1016/j.neucom.2023.01.088.
    [8] 王非, 王欣宇, 周景春, 等. 一种基于YOLOv3的水下声呐图像目标检测方法[J]. 电子与信息学报, 2022, 44(10): 3419–3426. doi: 10.11999/JEIT220260.

    WANG Fei, WANG Xinyu, ZHOU Jingchun, et al. An underwater object detection method for sonar image based on YOLOv3 model[J]. Journal of Electronics & Information Technology, 2022, 44(10): 3419–3426. doi: 10.11999/JEIT220260.
    [9] DAI Linhui, LIU Hong, SONG Pinhao, et al. A gated cross-domain collaborative network for underwater object detection[J]. Pattern Recognition, 2024, 149: 110222. doi: 10.1016/j.patcog.2023.110222.
    [10] YUAN Jieyu, CAI Zhanchuan, and CAO Wei. A novel underwater detection method for ambiguous object finding via distraction mining[J]. IEEE Transactions on Industrial Informatics, 2024, 20(7): 9215–9224. doi: 10.1109/TII.2024.3383537.
    [11] 沈学利, 李东峰. 频域重标定与自适应稀疏金字塔水下实时目标检测[J/OL]. 激光与光电子学进展. https://link.cnki.net/urlid/31.1690.TN.20260121.1736.048, 2026.

    SHEN Xueli and LI Dongfeng. Real-time underwater object detection with frequency-domain recalibration and an adaptive sparse pyramid[J/OL]. Laser & Optoelectronics Progress. https://link.cnki.net/urlid/31.1690.TN.20260121.1736.048, 2026.
    [12] WANG Junzhe, CHEN Xinke, DAI Anbang, et al. LS-DETR: Lightweight transformer for object detection in forward-looking sonar images[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 1502805. doi: 10.1109/LGRS.2025.3575615.
    [13] JOCHER G, QIU Jing, and CHAURASIA A. Ultralytics YOLO[EB/OL]. https://github.com/ultralytics/ultralytics, 2025.
    [14] CHEN Jierun, KAO S H, HE Hao, et al. Run, don't walk: Chasing higher FLOPS for faster neural networks[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 12021–12031. doi: 10.1109/CVPR52729.2023.01157.
    [15] HAN Kai, WANG Yunhe, TIAN Qi, et al. GhostNet: More features from cheap operations[C]. Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2020: 1577–1586. doi: 10.1109/CVPR42600.2020.00165.
    [16] LEE J, PARK S, MO S, et al. Layer-adaptive sparsity for the magnitude-based pruning[C]. 9th International Conference on Learning Representations, 2021. (查阅网上资料, 未找到出版地信息, 请补充).
    [17] FANG Gongfan, MA Xinyin, SONG Mingli, et al. DepGraph: Towards any structural pruning[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 16091–16101. doi: 10.1109/CVPR52729.2023.01544.
    [18] LIU Chongwei, LI Haojie, WANG Shuchang, et al. A dataset and benchmark of underwater object detection for robot picking[C]. 2021 IEEE International Conference on Multimedia & Expo Workshops, Shenzhen, China, 2021: 1–6. doi: 10.1109/ICMEW53276.2021.9455997.
    [19] FU Chenping, LIU Risheng, FAN Xin, et al. Rethinking general underwater object detection: Datasets, challenges, and solutions[J]. Neurocomputing, 2023, 517: 243–256. doi: 10.1016/j.neucom.2022.10.039.
    [20] HONG J, FULTON M, and SATTAR J. TrashCan: A semantically-segmented dataset towards visual detection of marine debris[EB/OL]. arXiv: 2007.08097. https://doi.org/10.48550/arXiv.2007.08097, 2020.
    [21] ZHU Xizhou, SU Weijie, LU Lewei, et al. Deformable DETR: Deformable transformers for end-to-end object detection[C]. 9th International Conference on Learning Representations, 2021. (查阅网上资料, 未找到出版地信息, 请补充).
    [22] JOCHER G. YOLOv5 by ultralytics[EB/OL]. https://github.com/ultralytics/yolov5, 2025.
    [23] WANG C Y, BOCHKOVSKIY A, and MARK LIAO H Y. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors[C]. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 7464–7475. doi: 10.1109/CVPR52729.2023.00721.
    [24] WANG C Y, YEH I H, and MARK LIAO H Y. Yolov9: Learning what you want to learn using programmable gradient information[C]. 18th European Conference on Computer Vision, Milan, Italy, 2025: 1–21. doi: 10.1007/978-3-031-72751-1_1.
  • 加载中
图(7) / 表(3)
计量
  • 文章访问数:  16
  • HTML全文浏览量:  11
  • PDF下载量:  2
  • 被引次数: 0
出版历程
  • 修回日期:  2026-02-13
  • 录用日期:  2026-02-13
  • 网络出版日期:  2026-03-01

目录

    /

    返回文章
    返回