Advanced Search
Turn off MathJax
Article Contents
HOU Guojia, MA Jiaqi, WANG Yuechuan, HUANG Baoxiang, LI Kunqian. UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251129
Citation: HOU Guojia, MA Jiaqi, WANG Yuechuan, HUANG Baoxiang, LI Kunqian. UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251129

UWF-YOLO: A Lightweight Framework for Underwater Object Detection via Redundant Information Optimization

doi: 10.11999/JEIT251129 cstr: 32379.14.JEIT251129
Funds:  National Natural Science Foundation of China (62371431, 61901240), Qingdao Natural Science Foundation (24-4-4-zrjj-122-jch), Natural Science Foundation of Shandong Province, China (ZR2024MF125, ZR2025QB60)
  • Accepted Date: 2026-02-13
  • Rev Recd Date: 2026-02-13
  • Available Online: 2026-03-01
  •   Objective  The rapid development of underwater imaging technology has significantly elevated the importance of underwater object detection for resource exploration and environmental monitoring applications. Generally, complex underwater environments yield various degradations of image quality such as color casts, haze-like effects, and non-uniform illumination. Unfortunately, existing vision-based object detection algorithms always suffer from unpleasing performance and notable limitations especially for detecting small objects, resulting in missed detections and false positives. Moreover, existing deep learning based underwater detection models also face substantial challenges in striking an optimal balance between accuracy and lightweight design under the condition of limited equipment resources. To address these issues, it is of great importance to design efficient underwater object detection methods in view of water-related vision tasks, which play a crucial role in marine resource exploration, ecological monitoring, underwater robotics, and intelligent perception systems for autonomous underwater vehicles.  Methods  In this paper, we propose a novel lightweight framework based on redundant information optimization for underwater object detection. Technically, we propose a lightweight underwater object detection network called UWF-YOLO based on redundancy information optimization. First, the C2f module is reconstructed by FasterNet Block to optimize both the backbone and neck networks, and a feature channel selection mechanism is incorporated to reduce the redundant features. On other hand, due to the redundant traditional convolutional features in the YOLO neck, it is difficult to adapt to the underwater environment. Ghost Convolution is also introduced to generate the Ghost feature map for enhancing the multi-scale feature fusion capability of the neck network. Next, our proposed method achieves parameter sharing by replacing the original detection head with a redundant optimization group detection head (RRG-Head) based on group convolution, thereby reducing computational costs. Finally, the structured channel pruning technique is applied to identify the inter-layer dependencies of the graph and bind the pruning units. Combined with the LAMP weight magnitude score normalization for evaluating the importance of channels, the low-contributing groups are pruned and fine-tuned to achieve network size compression. In addition, since the scene of underwater detection datasets are typically monotonous and the underwater objects contained in the available datasets are usually small and clustered. We also construct an underwater object detection dataset with complex scene, namely CSUOD, by collecting real-world underwater images from different websites and platforms to ensure both its diversity and authenticity, followed by manual annotation and resolution normalization preprocessing. CSUOD is specifically designed for various challenging underwater environments characterized by color casts, haze-like effects, and non-uniform illumination. In our CSUOD, we manually select 1135 images containing 6 different types, and perform the manual annotation and resolution standardization operations.  Results and Discussions  Extensive experiments are conducted on three public underwater object detection datasets (i.e., DUO, RUOD, and TrashCan) by comparing several popular and widely used object detection methods. The proposed model is evaluated against mainstream detectors, including YOLOv5s, YOLOv7-tiny, YOLOv8s, YOLOv9-tiny, and Deformable DETR. In computational complexity assessment, experimental results show that the proposed method has reduced the FLOPs, model size, and parameters by 60.4%, 77.3%, and 78.4%, respectively, compared to the baseline. In addition, our method has outperformed YOLOv9-tiny with comparable parameters by 0.3%, 2.3%, and 3.4% in mAP across the three datasets. Also, some comparative results on our established CSUOD dataset also indicate that our proposed model has a good improvement and stability even in complex underwater environments. Qualitative visualization results further illustrate the model’s robustness and detection stability under various underwater degradations, such as haze-like effects and non-uniform illumination.  Conclusions  Quantitative and qualitative experiments on different datasets have validated the effectiveness and robustness of the proposed method. In addition, our method achieves superior detection performance in complex underwater environments, effectively solving missed detections and false positives caused by background interference. A large number of experimental results show that our designed UWF-YOLO can not only achieve significant light weighting, but also maintain the comparable detection accuracy comparing with the benchmark model. This balance between the detection accuracy and low computational cost makes it particularly suitable for underwater devices with limited resources. Besides, the proposed method has great potential in practical scenarios such as marine ecological monitoring, underwater resource exploration, and autonomous underwater vehicle perception systems. It also provides a reliable and efficient technical foundation for real-time applications, with strong adaptability to different underwater conditions, efficient integration into embedded platforms, and support for real-time perception and decision-making. Our constructed dataset CSUOD in this study will help address the limitations of existing underwater object detection datasets and promote the development of underwater object detection. In the future, this work can be further extended to multi-modal perception systems and larger-scale datasets. These efforts will enable adaptive models for more dynamic underwater scenarios and support broader applications in intelligent ocean observation and autonomous navigation.
  • loading
  • [1]
    黄海宁, 李宝奇, 刘纪元, 等. 声呐图像水下目标识别综述与展望[J]. 电子与信息学报, 2024, 46(5): 1742–1760. doi: 10.11999/JEIT231207.

    HUANG Haining, LI Baoqi, LIU Jiyuan, et al. Sonar image underwater target recognition: A comprehensive overview and prospects[J]. Journal of Electronics & Information Technology, 2024, 46(5): 1742–1760. doi: 10.11999/JEIT231207.
    [2]
    WANG Hao, ZHANG Weibo, XU Yinghao, et al. WaterCycleDiffusion: Visual-textual fusion empowered underwater image enhancement[J]. Information Fusion, 2025, 127: 103693. doi: 10.1016/j.inffus.2025.103693.
    [3]
    ZHANG Dehua, YU Changcheng, LI Zhen, et al. A lightweight network enhanced by attention-guided cross-scale interaction for underwater object detection[J]. Applied Soft Computing, 2025, 184: 113811. doi: 10.1016/j.asoc.2025.113811.
    [4]
    CHEW A L, TONG P B, and CHIA C S. Automatic detection and classification of man-made targets in side scan sonar images[C]. 2007 Symposium on Underwater Technology and Workshop on Scientific Use of Submarine Cables and Related Technologies, Tokyo, Japan, 2007: 126–132. doi: 10.1109/UT.2007.370841.
    [5]
    BEIJBOM O, EDMUNDS P J, KLINE D I, et al. Automated annotation of coral reef survey images[C]. 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, USA, 2012: 1170–1177. doi: 10.1109/CVPR.2012.6247798.
    [6]
    LI Xiu, SHANG Min, QIN Hongwei, et al. Fast accurate fish detection and recognition of underwater images with fast R-CNN[C]. OCEANS 2015-MTS/IEEE Washington, Washington, USA, 2015: 1–5. doi: 10.23919/OCEANS.2015.7404464.
    [7]
    SONG Pinhao, LI Pengteng, DAI Linhui, et al. Boosting R-CNN: Reweighting R-CNN samples by RPN’s error for underwater object detection[J]. Neurocomputing, 2023, 530: 150–164. doi: 10.1016/j.neucom.2023.01.088.
    [8]
    王非, 王欣宇, 周景春, 等. 一种基于YOLOv3的水下声呐图像目标检测方法[J]. 电子与信息学报, 2022, 44(10): 3419–3426. doi: 10.11999/JEIT220260.

    WANG Fei, WANG Xinyu, ZHOU Jingchun, et al. An underwater object detection method for sonar image based on YOLOv3 model[J]. Journal of Electronics & Information Technology, 2022, 44(10): 3419–3426. doi: 10.11999/JEIT220260.
    [9]
    DAI Linhui, LIU Hong, SONG Pinhao, et al. A gated cross-domain collaborative network for underwater object detection[J]. Pattern Recognition, 2024, 149: 110222. doi: 10.1016/j.patcog.2023.110222.
    [10]
    YUAN Jieyu, CAI Zhanchuan, and CAO Wei. A novel underwater detection method for ambiguous object finding via distraction mining[J]. IEEE Transactions on Industrial Informatics, 2024, 20(7): 9215–9224. doi: 10.1109/TII.2024.3383537.
    [11]
    沈学利, 李东峰. 频域重标定与自适应稀疏金字塔水下实时目标检测[J/OL]. 激光与光电子学进展. https://link.cnki.net/urlid/31.1690.TN.20260121.1736.048, 2026.

    SHEN Xueli and LI Dongfeng. Real-time underwater object detection with frequency-domain recalibration and an adaptive sparse pyramid[J/OL]. Laser & Optoelectronics Progress. https://link.cnki.net/urlid/31.1690.TN.20260121.1736.048, 2026.
    [12]
    WANG Junzhe, CHEN Xinke, DAI Anbang, et al. LS-DETR: Lightweight transformer for object detection in forward-looking sonar images[J]. IEEE Geoscience and Remote Sensing Letters, 2025, 22: 1502805. doi: 10.1109/LGRS.2025.3575615.
    [13]
    JOCHER G, QIU Jing, and CHAURASIA A. Ultralytics YOLO[EB/OL]. https://github.com/ultralytics/ultralytics, 2025.
    [14]
    CHEN Jierun, KAO S H, HE Hao, et al. Run, don't walk: Chasing higher FLOPS for faster neural networks[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 12021–12031. doi: 10.1109/CVPR52729.2023.01157.
    [15]
    HAN Kai, WANG Yunhe, TIAN Qi, et al. GhostNet: More features from cheap operations[C]. Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2020: 1577–1586. doi: 10.1109/CVPR42600.2020.00165.
    [16]
    LEE J, PARK S, MO S, et al. Layer-adaptive sparsity for the magnitude-based pruning[C]. 9th International Conference on Learning Representations, 2021. (查阅网上资料, 未找到出版地信息, 请补充).
    [17]
    FANG Gongfan, MA Xinyin, SONG Mingli, et al. DepGraph: Towards any structural pruning[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 16091–16101. doi: 10.1109/CVPR52729.2023.01544.
    [18]
    LIU Chongwei, LI Haojie, WANG Shuchang, et al. A dataset and benchmark of underwater object detection for robot picking[C]. 2021 IEEE International Conference on Multimedia & Expo Workshops, Shenzhen, China, 2021: 1–6. doi: 10.1109/ICMEW53276.2021.9455997.
    [19]
    FU Chenping, LIU Risheng, FAN Xin, et al. Rethinking general underwater object detection: Datasets, challenges, and solutions[J]. Neurocomputing, 2023, 517: 243–256. doi: 10.1016/j.neucom.2022.10.039.
    [20]
    HONG J, FULTON M, and SATTAR J. TrashCan: A semantically-segmented dataset towards visual detection of marine debris[EB/OL]. arXiv: 2007.08097. https://doi.org/10.48550/arXiv.2007.08097, 2020.
    [21]
    ZHU Xizhou, SU Weijie, LU Lewei, et al. Deformable DETR: Deformable transformers for end-to-end object detection[C]. 9th International Conference on Learning Representations, 2021. (查阅网上资料, 未找到出版地信息, 请补充).
    [22]
    JOCHER G. YOLOv5 by ultralytics[EB/OL]. https://github.com/ultralytics/yolov5, 2025.
    [23]
    WANG C Y, BOCHKOVSKIY A, and MARK LIAO H Y. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors[C]. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 7464–7475. doi: 10.1109/CVPR52729.2023.00721.
    [24]
    WANG C Y, YEH I H, and MARK LIAO H Y. Yolov9: Learning what you want to learn using programmable gradient information[C]. 18th European Conference on Computer Vision, Milan, Italy, 2025: 1–21. doi: 10.1007/978-3-031-72751-1_1.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(7)  / Tables(3)

    Article Metrics

    Article views (56) PDF downloads(14) Cited by()
    Proportional views
    Related

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return