BG-YOLO: A Bidirectional-Guided Method for Underwater Object Detection.

Ruicheng Cao, Ruiteng Zhang, Xinyue Yan, Jian Zhang
Author Information
  1. Ruicheng Cao: School of Cybersecurity, Northwestern Polytechnical University, Xi'an 710072, China. ORCID
  2. Ruiteng Zhang: College of Computer Science and Technology, Zhejiang University, Hangzhou 310027, China. ORCID
  3. Xinyue Yan: College of Computer Science, Chongqing University, Chongqin 400044, China. ORCID
  4. Jian Zhang: School of Tropical Agriculture and Forestry, Hainan University, Haikou 571158, China. ORCID

Abstract

Degraded underwater images decrease the accuracy of underwater object detection. Existing research uses image enhancement methods to improve the visual quality of images, which may not be beneficial in underwater image detection and lead to serious degradation in detector performance. To alleviate this problem, we proposed a bidirectional guided method for underwater object detection, referred to as BG-YOLO. In the proposed method, a network is organized by constructing an image enhancement branch and an object detection branch in a parallel manner. The image enhancement branch consists of a cascade of an image enhancement subnet and object detection subnet. The object detection branch only consists of a detection subnet. A feature-guided module connects the shallow convolution layers of the two branches. When training the image enhancement branch, the object detection subnet in the enhancement branch guides the image enhancement subnet to be optimized towards the direction that is most conducive to the detection task. The shallow feature map of the trained image enhancement branch is output to the feature-guided module, constraining the optimization of the object detection branch through consistency loss and prompting the object detection branch to learn more detailed information about the objects. This enhances the detection performance. During the detection tasks, only the object detection branch is reserved so that no additional computational cost is introduced. Extensive experiments demonstrate that the proposed method significantly improves the detection performance of the YOLOv5s object detection network (the mAP is increased by up to 2.9%) and maintains the same inference speed as YOLOv5s (132 fps).

Keywords

References

  1. IEEE Trans Pattern Anal Mach Intell. 2021 Aug;43(8):2623-2633 [PMID: 32149681]
  2. IEEE Trans Image Process. 2022;31:4922-4936 [PMID: 35849672]
  3. IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6129-6143 [PMID: 33900925]
  4. J Opt Soc Am. 1971 Jan;61(1):1-11 [PMID: 5541571]
  5. Sensors (Basel). 2021 Oct 29;21(21): [PMID: 34770509]
  6. IEEE Trans Pattern Anal Mach Intell. 2011 Dec;33(12):2341-53 [PMID: 20820075]
  7. IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14164-14177 [PMID: 37220059]

Grants

  1. 623RC449./Natural Science Foundation of Hainan province

Word Cloud

Created with Highcharts 10.0.0detectionobjectbranchimageenhancementunderwatersubnetperformanceproposedmethodimagesguidednetworkconsistsfeature-guidedmoduleshallowfeatureoptimizationYOLOv5sDegradeddecreaseaccuracyExistingresearchusesmethodsimprovevisualqualitymaybeneficialleadseriousdegradationdetectoralleviateproblembidirectionalreferredBG-YOLOorganizedconstructingparallelmannercascadeconnectsconvolutionlayerstwobranchestrainingguidesoptimizedtowardsdirectionconducivetaskmaptrainedoutputconstrainingconsistencylosspromptinglearndetailedinformationobjectsenhancestasksreservedadditionalcomputationalcostintroducedExtensiveexperimentsdemonstratesignificantlyimprovesmAPincreased29%maintainsinferencespeed132fpsBG-YOLO:Bidirectional-GuidedMethodUnderwaterObjectDetectionjoint

Similar Articles

Cited By (1)