Underwater Target Detection Based on Parallel High-Resolution Networks
Next Article in Journal
SUGAN: A Stable U-Net Based Generative Adversarial Network
Next Article in Special Issue
Fast Automatic Fuzzy C-Means Knitting Pattern Color-Separation Algorithm Based on Superpixels
Previous Article in Journal
Rapid and Green Classification Method of Bacteria Using Machine Learning and NIR Spectroscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Target Detection Based on Parallel High-Resolution Networks

1
College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
2
Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(17), 7337; https://doi.org/10.3390/s23177337
Submission received: 17 July 2023 / Revised: 2 August 2023 / Accepted: 20 August 2023 / Published: 23 August 2023
(This article belongs to the Special Issue Image/Video Segmentation Based on Sensor Fusion)

Abstract

:
A parallel high-resolution underwater target detection network is proposed to address the problems of complex underwater scenes and limited target feature extraction capability. First, a high-resolution network (HRNet), a lighter high-resolution human posture estimation network, is used to improve the target feature representation and effectively reduce the semantic information lost in the image during sampling. Then, the attention module (A-CBAM) is improved to capture complex feature distributions by modeling the two-dimensional space in the activation function stage through the introduction of the flexible rectified linear units (FReLU) activation function to achieve pixel-level spatial information modeling capability. Feature enhancement in the spatial and channel dimensions is performed to improve understanding of fuzzy targets and small target objects and to better capture irregular and detailed object layouts. Finally, a receptive field augmentation module (RFAM) is constructed to obtain sufficient semantic information and rich detail information to further enhance the robustness and discrimination of features and improve the detection capability of the model for multi-scale underwater targets. Experimental results show that the method achieves 81.17%, 77.02%, and 82.9% mean average precision (mAP) on three publicly available datasets, specifically underwater robot professional contest (URPC2020, URPC2018) and pattern analysis, statistical modeling, and computational learning visual object classes (PASCAL VOC2007), respectively, demonstrating the effectiveness of the proposed network.

1. Introduction

With the development of computer vision and the exploitation of marine resources, biological detection in underwater environments has become a hot research topic with applications in underwater robotics [1], underwater exploration [2], and marine research [3]. However, compared to terrestrial scenes [4], underwater images suffer from problems such as color bias [5], low contrast [6], blur [7], and noise [8], resulting in the loss of clear contour and texture information in images, which limits the accuracy of target detection. Therefore, accurate detection of marine organisms in complex underwater environments remains a challenge. To this end, a parallel high-resolution underwater target detection network is proposed to improve the accuracy and efficiency of target detection given the complexity of underwater scenes, limited target feature extraction capability, fuzzy targets, and small targets.
Currently existing target detection models can be classified into two types: anchor-free and anchor-based. Anchor-based detectors include the region-based convolutional network (R-CNN) series [9,10,11,12], the You Only Look Once (YOLO) series [13,14,15], and the Single Shot MultiBox detector (SSD) series [16,17], which provide a priori knowledge of the model by using predefined anchor boxes and have achieved better performance. However, the anchor box introduces redundant negative samples, leading to a reduction in inference speed. Anchor-free detectors such as CenterNet [18], CornerNet [19], and fully convolutional one-stage (FCOS) [20] and end-to-end object detection with transformers (DETR) [21] discard the anchor box and treat the target as a key point for prediction. With effective training, the anchor-free detector can compete with the anchor-based detector in terms of performance and has a better generalization capability.
Recent research on underwater objects, for example, Victor et al. [22] provided an overview of various applications of federated learning in the internet of underwater things (IoUT), as well as its challenges and unsolved issues, have indicated future research prospects. The IoT framework emphasizes the significance of data security and privacy. Meng et al. [23] argued that in order to discover the valuable hidden data detected by an automatic underwater vehicle (AUV) swarm, it is necessary to aggregate the data detected by the AUV swarm to generate a powerful machine learning model and introduced federated learning into the collaboration of an AUV swarm. To further reduce the constraints of scarce underwater communication resources on federated learning and to mitigate the lag effect, an asynchronous federated learning method was developed. Bhattacharya et al. [24] first discussed blockchain technology and the Internet of Things (IoT), pointing out the benefits of integrating blockchain technology with IoT systems and providing an overview of various applications of blockchain-enabled IoUT systems, their respective challenges, and possible future directions. Finally, this work identifies key aspects of IoUT systems and enables researchers to address challenges using blockchain technology to improve the accuracy and efficiency of underwater target detection and to promote the development of underwater target detection by integrating research results and improvement strategies from different fields.
In the field of underwater environments, researchers have made important contributions by proposing various target detection algorithms [25,26,27]. For example, Zhou et al. [28] combined image enhancement techniques with an extended visual geometry group (VGG16) feature extraction network in 2017, using a faster region-based convolutional neural network (Faster R-CNN) and feature mapping to detect and identify underwater biological targets in a URPC dataset. Villon et al. [29] combined a histogram of oriented gradient (HOG) and a support vector machine (SVM) to detect coral reef fish in captured images, solving the problem of underwater accuracy degradation caused by target occlusion and overlap in the images. Chen et al. [30] introduced a novel sample distribution-based weighted loss function called the Invert Multi-Class AdaBoost (IMA) in 2020 to mitigate the adverse effects of noise on detection performance. Ma et al. [31,32] developed lightweight models by pruning existing target detection models (e.g., SDD) or integrating lightweight backbones (e.g., MobileNet). Li et al. [33] added an extra detection head to the YOLOv5 model to improve multi-scale detection and small target detection accuracy. Zhang et al. [34] combined an attention mechanism to learn target features and increase the receptive field to improve the detection accuracy and robustness of underwater small target detection. Yao et al. [35,36] used residual networks as a backbone network to improve the efficiency of sea cucumber feature extraction. Jiang et al. [37] also proposed a method called the Channel Sharpening Attention Module (CSAM) for further fusion of advanced image information. A summary of related work methods is shown in Table 1.
The task of target detection becomes particularly difficult in the underwater environment due to blurred feature information, severe color distortion, and low visibility, requiring simultaneous localization and classification. In order to improve the accuracy and portability of underwater target detection capabilities, we have chosen to improve the classical CenterNet model in the anchor-free family and propose a parallel high-resolution network model to address the challenges faced in this domain. The effectiveness of our proposed model is demonstrated by conducting experiments on underwater images.
The innovations in this study are as follows: (1) aiming at the problem of complex underwater scenes and limited target feature extraction capability, a lighter high-resolution human posture estimation network (HRNet) [38] is used to enhance the target feature representation and effectively reduce the semantic information lost in the image sampling process; (2) for the problem of fuzzy and small targets, by introducing the FReLU [39] activation function, the improved convolutional block attention module (CBAM) [40] structure captures the complex feature distribution by modeling the two-dimensional space in the activation function stage to achieve the ability to model spatial information at the pixel level while performing feature enhancement in the spatial dimension and channel dimension; (3) we constructed a receptive field augmentation module (RFAB) to further obtain sufficient semantic information and rich detail information to further enhance the robustness of the features and discriminative ability of the model for underwater multi-scale target detection.

2. Related Work

From the literature, it can be seen that by having a multi-resolution network parallel connection structure and fusing feature maps of different resolutions so that each resolution feature map can accept information from other resolution feature maps many times, output feature maps can be obtained that contain both high-resolution positional information and low-resolution semantic information, which can enhance the expression of target features and reduce the semantic information lost in the sampling process. However, the separate use of low- and high-resolution features still cannot meet our needs for high-quality recognition. If the attention mechanism can be properly used to enhance the target information in the feature map from both spatial and channel dimensions, it can guide the model to increase its attention to key information. It also improves its activation function to model the 2D space in the activation function stage to capture the complex feature distribution. In addition, the receptive field module demonstrates the model’s ability to obtain rich semantic information about the target context.

2.1. High-Resolution Network

In convolutional neural networks, high-resolution images contain richer detail information and play a crucial role in target detection. However, high-resolution feature maps also have a higher computational cost, resulting in more operations. In addition, in the deeper layers of the model, low-resolution feature maps contain rich semantic information that helps improve the model’s ability to detect targets.
HRNet fuses low-resolution feature maps with high-resolution feature maps by connecting networks of different resolutions in parallel, so that the feature extraction network outputs high-resolution feature maps that contain rich semantic information. The feature maps of different resolutions are obtained through different branches of the image, and the feature maps of each branch are fused with the feature maps of all scales in the previous stage, so that the feature maps of each branch take into account both high- and low-resolution feature information, thus retaining more target feature information.
The key feature of HRNet is that it maintains high resolution throughout the model’s processing of the image and generates multiple low-resolution feature maps simultaneously, which are used to augment the feature information of the high-resolution feature maps. Unlike previous networks, HRNet does not use extended convolution or upsampling to pass low-resolution features to recover the high-resolution representation. Instead, HRNet maintains the high-resolution representation throughout the modeling process and continuously performs cross-scale fusion to learn richer feature representations. HRNet is designed using a stage-by-stage incremental strategy, starting with a high-resolution subnet and then gradually adding new subnets in parallel to form multiple stages, with the resolution of the current subnet being 1/2 the resolution of the previous subnet. This process is repeated multiple times, eventually forming four stages. Based on these considerations, HRNet is a lighter weight network for high-resolution human posture estimation.

2.2. Attention Mechanism

Attention mechanisms are becoming popular in various fields and are often applied in a weighted manner. In computer vision tasks, where different regions of an image or video frame may have different levels of importance, attention mechanisms have been found to allow more flexible processing of visual information. The basis of this mechanism is similar to the human selective attention mechanism. Humans tend to ignore irrelevant information when observing objects and focus on information that is useful for decision making. Attention mechanisms are based on the idea of acquiring task-critical information. Therefore, an attention mechanism is often used to automatically select features or regions of interest during target detection.
Models that process target images contain a large number of “meaningless” regions that do not contain information about the target, which interferes with feature extraction of valid targets. To improve feature extraction, researchers have proposed attention mechanisms to guide models to focus on target regions. The design of attention mechanisms is inspired by the unique brain signal processing mechanism of human vision. When the human eye observes an image, it generates a focus of attention that concentrates on certain key regions to eliminate the interference of irrelevant information, thus improving the accuracy and efficiency of information processing.
Attention mechanisms in neural networks are plug-and-play learnable modules that work by assigning weight coefficients to values in the feature map tensor to enhance information in the focus region. They mainly include the channel attention mechanism and the spatial attention mechanism. The channel attention mechanism achieves dynamic enhancement of the features in each channel by assigning weights to the channel dimension of the feature map, while the spatial attention mechanism assigns weights in the spatial dimension and learns the dependencies of different locations on the image to dynamically enhance the features in the spatial dimension. Self-attention mechanisms, on the other hand, are a branch of attention mechanisms developed from the field of natural language processing and are unique in that they do not rely on the guidance of external information but rely on their own input to establish global dependencies to generate weight coefficients. By combining the structure of these two mechanisms, the target information in the feature map can be strengthened in terms of spatial and channel dimensions. In addition, the activation function can be enhanced to capture complex feature distributions by modeling the 2D space using the activation function stage. The FReLU activation function, which is specifically designed for vision tasks, is chosen to replace the ReLU activation function in the CBAM module. The FReLU activation extends ReLU to a two-dimensional activation function by adding a negligible spatial conditioning overhead, which improves target detection accuracy by capturing complex visual layouts in two-dimensional space. In addition, spatial conditioning implements pixel-level modeling capabilities in a simple way to capture complex visual layouts through regular convolution. FReLU, on the other hand, is a spatial context-dependent 2D condition that helps to extract the spatial layout of target features.

2.3. Receptive Field Module

Small targets occupy fewer pixels in images, resulting in insufficient information about their characteristics. In order to optimize the recognition effect, an optimization idea can be adopted, which is to use the information of the target’s environment to support the recognition, called the target contextual semantic information. Usually, the objects in the image do not exist independently, and there is some relationship between them and the environment. Therefore, by making full use of the target contextual semantic information, the feature representation of the target can be effectively enriched. One of the commonly used optimization methods is to expand the receptive field of the neural network by using null convolution.
The receptive field module takes advantage of the population receptive field properties of human vision and constructs a multi-branch structure using hollow convolutional kernels of different sizes and eccentricities. This receptive field module on the input image is usually first processed with three ordinary convolution kernels of different sizes (1 × 1, 3 × 3, 5 × 5). Then, three feature maps with different receptive fields are separately obtained using hollow convolution with different centrifugal rates. By fusing these feature maps, feature enhancement results can be obtained. To reduce the computational effort, alternative convolution kernels can be used, e.g., the 5 × 5 convolution layer can be replaced by a 3 × 3 convolution layer, and the 3 × 3 convolution layer can be replaced by 1 × 3 and 3 × 1 convolution layers.
Compared to the traditional convolutional module, the feature maps produced by the lower centration branches of the module are able to condense feature information that includes parts or the target as a whole. The higher centrifugal branches can condense the feature information including the target itself and the surrounding environment. The advantage of fusing multiple branches is that the higher centrifugal branch can provide contextual semantic information to the lower centrifugal branch, while the lower centrifugal branch can compensate for the loss of detail caused by the convolutional kernel diffusion of the higher centrifugal branch.
The structure of the receptive field is borrowed and both scale and centrifugal rate are taken into account. Compared to the use of a computationally intensive and deeply layered backbone network, the module uses a lightweight backbone network. By combining the receptive field module, the receptive field augmentation module proposed in this paper achieves a stronger feature representation capability, which enables the detector to have a faster inference speed while maintaining excellent performance. In addition, the training process does not require too many complex tricks, so the training resources of the model are less.
In the existing one-stage and two-stage detectors, they usually face the dilemma of good performance but slow speed or fast speed but unsatisfactory performance. To address this problem, this paper proposes a scheme that uses a lightweight backbone and enhanced receptive field module to achieve faster inference while maintaining performance. The design elegantly combines feature fusion with different centrifugal velocities, bringing significant performance gains to the target detection task.

3. Methods

3.1. Improved Network Structure

In order to effectively integrate high-resolution shallow feature information and low-resolution deep feature information, this paper proposes a network structure based on the parallel connection of multi-resolution subnetworks designed by CenterNet. The underwater object detection capability is enhanced by using a backbone network (HRNet), improving the attention module (A-CBAM), and constructing a receptive field augmentation module (RFAB). It consists of two parts: the AR-HRNet backbone network and the recognition module. The improved network structure is shown in Figure 1.
The input image is first passed through four parallel subnetworks of the HRNet, allowing for the rapid extraction of features at four different resolutions. The four different resolution features are then fed into the attention module (A-CBAM), which improves model performance by increasing the spatial and channel dimensions. In the CBAM structure, complex feature distributions are captured in the activation function stage by modeling the two-dimensional space through the introduction of the FReLU activation function to achieve the ability to model spatial information at the pixel level. It then passes through the receptive field augmentation module (RFAB) to further capture rich semantic and positional information to enhance target detection at multiple scales underwater. Finally, the fused features are passed to three separate branches to generate key point heatmaps, position offsets and target width and height, and the final detection results are obtained through decoding operations.

3.1.1. Parallel High-Resolution Network

The benchmark network CenterNet adopts Hourglass-104 as the backbone network; although it can achieve high detection accuracy, the network has a complex structure and a larger number of parameters, which greatly increases the computational complexity of the network and reduces the detection speed of the network. Therefore, in order to improve the detection speed of the algorithm while maintaining high detection accuracy, this paper changes the backbone network to a HRNet network from the perspective of lightweight network structure.
Originally used for human posture estimation, HRNet is a feature extraction network designed to overcome the information loss and ambiguity problems that occur with traditional deep learning networks when processing high-resolution images. HRNet effectively improves the network’s ability to perceive detailed information by parallelizing multiple branches to process feature maps at different resolutions and by interacting and fusing between them.
The core idea of HRNet is to maintain the transfer and use of high-resolution features. It enables the network to maintain high-resolution features at different resolutions by keeping information fluid between branches. Specifically, HRNet allows the network to maintain both global receptive field and detail information by performing resolution-by-resolution upsampling and feature fusion multiple times. This multi-level feature fusion design allows HRNet to better capture targets at different scales in target detection tasks and provide more accurate and detailed detection results. The structure of the HRNet network is shown in Figure 2.
HRNet is divided into four stages: each stage is a parallel connection of multi-resolution subnetworks and consists of a series of standard convolutions; the feature map resolution of the same subnetwork does not change with the depth of the network, while the feature map resolution of the parallel subnetworks decreases in turn by 1/2, while the number of channels increases by a factor of 2. Starting from the backbone network, the image is processed by a 3 × 3 convolution with a step size of 2. The feature map size obtained is 1/4 of the original image and the number of channels is increased from 3 to 64. A high-resolution subnetwork is constructed and the first stage consists of a subnetwork using four bottleneck [41] modules to extract features and adjusting the channels to 32. The multi-resolution module is shown in Figure 3. The second, third, and fourth stages then consist of multi-resolution modules containing 1, 4, and 3 multi-resolution modules, respectively, where the multi-resolution module consists of 4 BasicBlocks at each resolution, followed by a cross-resolution fusion module. The specific fusion method is shown in Figure 4. For the upsampling operation, the number of channels is first aligned using a 1 × 1 convolutional layer and the resolution is expanded using a neighborhood interpolation method. For the downsampling operation, a 3 × 3 convolution with a step size of 2 is used for 2-fold downsampling and two 3 × 3 convolutions with a step size of 2 are used for 4-fold downsampling.
The four subnetworks of HRNet downsample the images by 4, 8, 16, and 32 times to extract feature maps of different resolutions. The original HRNet only uses the output of the high-resolution subnetwork, but to make full use of the multi-scale feature information, this paper uses the feature maps generated by the four parallel subnetworks as the output of the backbone network. This can effectively fuse features at different scales to improve the performance and accuracy of target detection.

3.1.2. Improved Attention Mechanism

The CBAM consists of a linear stack of the channel attention module (CAM) and the spatial attention module (SAM). This paper improves the ReLU activation function in the CBAM module by replacing it with the FReLU activation function, which is specifically designed for visual tasks. FReLU activation extends ReLU to a two-dimensional activation function by adding negligible spatial conditioning overhead, which captures complex visual layouts in two-dimensional space, improving target detection accuracy. In addition, spatial conditioning implements pixel-level modeling capability in a simple way to capture complex visual layouts through regular convolution. FReLU, on the other hand, is a spatial context-dependent 2D condition that helps to extract the spatial layout of target features. Based on this, we have designed a new attention mechanism called the A-CBAM attention mechanism. By applying the A-CBAM attention mechanism, we expect to further improve the target detection performance. Its structure is shown in Figure 5.
The function of the aforementioned channel attention module is to compress the spatial dimensionality of the feature map by applying average pooling and max pooling operations based on width and height, transforming it into a one-dimensional vector and then processing it. Average pooling and max pooling can be used to aggregate the spatial information of the feature map. The feature vectors are then passed to the fully connected layer, where the spatial dimensions of the input feature map are compressed and then summed and merged element by element. Finally, the channel attention map is generated by applying a sigmoid activation function. The structure is shown in Figure 6.
Its formula is:
M c ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) )
where M c ( F ) is the channel attention feature, σ is the sigmoid activation function, M L P is a 2-layer neural network, F is the input, AvgPool is the average pooling, and MaxPool is the maximum pooling.
To obtain the attentional features of the feature map in the spatial dimension, we performed global maximum pooling and global average pooling on the feature map output using the channel attention module to rescale the dimension of the feature map from H × W to 1 × 1. Next, we performed convolution using a 7 × 7 convolution kernel and applied the sigmoid activation function to reduce the dimensionality of the feature map. We then merged the feature map normalized by the sigmoid activation function with the feature map output by the channel attention module to complete the rescaling of the feature map in both spatial and channel dimensions. The structure is shown in Figure 7.
Its formula is:
M S ( F ) = σ f 7 × 7 ( [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] )
where M S ( F ) is the spatial attention feature and f 7 × 7 is the convolution of the 7 × 7 convolution kernel.
A key component of CBAM is the Multilayer Perceptron (MLP), a module that uses ReLU as the activation function for implementing non-linear activation. The introduction of the ReLU function solves the gradient disappearance problem in the backpropagation algorithm. However, in order to achieve pixel-level spatial information modeling capability and improve accuracy in the activation function stage, this paper replaces the original ReLU activation function with a FReLU activation function specifically designed for vision tasks. The ReLU function is represented as y = max ( x , 0 ) , while the FReLU function is of the form y = max ( x , T ( x ) ) , where T ( ) is the two-dimensional spatial representation. By using two-dimensional spatial information, FReLU can better capture complex visual layouts and improve the accuracy of target detection. Unlike the fixed zero-value condition of ReLU, FReLU is a two-dimensional condition that depends on the spatial context and helps to extract the spatial layout of the target features. Specifically, the two-dimensional condition depends on the spatial context of each pixel and uses the max ( ) function to obtain the maximum value between input x and the condition. The FReLU activation function is defined as shown below:
f x c , i , j = max x c , i , j , T x c , i , j
In the above formulation, T x c , i , j denotes the two-dimensional condition and we use a parametric pooling window to introduce spatial dependence. To achieve spatial dependence, we use a highly optimized depth-separable convolution operator and a batch normalization layer. The specific unfolding T x c , i , j is computed as shown below:
T x c , i , j = x c , i , j w p c w
In the above equation, ( i , j ) denotes the pixel position in two-dimensional space, c denotes the c-th channel, and x c , i , j w denotes the input pixel as a non-linear activation function for the c-th channel at the two-dimensional spatial position ( i , j ) . A parameterized pool window is used here, where p c w represents the coefficients shared by this window in the same channel. At the same time, the ReLU activation function was compared with the FReLU activation function, as shown in Figure 8.
The CBAM structure with the addition of the FReLU activation function is called A-CBAM. FReLU has a stronger context capture capability, resulting in a better understanding of fuzzy targets and small target objects. In complex cases, the A-CBAM structure is able to capture irregular and detailed object layouts better than the CBAM structure.

3.1.3. Improved Receptive Field Augmentation Module

The receptive field is the size of the region in which the pixel points on the feature map output by each layer of the convolutional neural network are mapped onto the input image. Neurological studies have shown that in the human visual cortex, the size of the population receptive field correlates with retinal eccentricity and that the size of the population receptive field varies in different retinal regions. To simulate this finding and to improve the deep features learned by lightweight convolutional neural network models, Liu et al. [42] proposed the receptive field block (RFB). This block can improve the ability of networks to learn deep features by simulating the properties of population receptive fields in the human visual system.
The receptive field module combines the parallel convolution of dilated convolution [43] and InceptionNet [44] to expand the receptive field of the feature layers. Null convolution is proposed to achieve receptive field expansion without reducing the size (resolution) of the feature map, allowing the output of each convolution to contain a greater range of information. A comparison of the dilated convolution with the ordinary convolution is shown in Figure 9.
The original RFB has three branches, each with a different sized convolution kernel; the sizes of the convolution kernels are 1 × 1, 3 × 3, and 5 × 5, and the end of the branch has a dilated convolution layer; the dilated rates are 1, 3, and 5, respectively, which determines the spacing of the convolution kernel processing data. The larger the dilated rate, the larger the spacing of the convolution kernel processing data. To adapt to different situations, we propose the RFAM structure, which has more branches than RFM and contains four branches, with different branches corresponding to the respective convolutional and dilated layers. RFAM uses 1 × 1 convolutional kernels in each branch to adjust the number of channels and uses a series of 3 × 3 convolutions instead of 5 × 5 convolutions to reduce the number of parameters in the model to increase the detection speed while increasing the model’s nonlinearity. In addition, we replaced the original 3 × 3 convolution with a 1 × 3 convolution and a 3 × 1 convolution. The results of the 4 branches are concatenated and passed through a 1 × 1 convolution and then fed into the activation function with a residual edge as input. The effect of the residual edge is to allow the input image signal to propagate directly from bottom to top, mitigating the gradient divergence caused by an overly deep network. Its calculation is shown below:
X out   = τ X in     ϵ ( B r 1 + B r 2 + B r 3 )   ×   s c a l e )
where X in   denotes the input features and Br1, Br2, and Br3 denote the outputs of the three branches. The + denotes the feature fusion operation. We use the symbol ϵ to denote the process of adjusting the number of channels by 1 × 1 convolution. scale denotes the weight of the residual edge linear operation; here, we take 0.1 as the weight value. The operator denotes summation by element. Finally, the symbol τ denotes the ReLU activation function. A comparison of the receptive modules is shown in Figure 10.

3.1.4. Loss Function

The total loss function in this paper’s algorithm can be divided into three components: heat map loss, center bias loss, and width–height loss, and the three loss functions are weighted and summed.
Heat map loss L k is a function with a focal loss penalty mechanism, defined as:
L k = 1 N x , y , c 1 Y ^ x , y , c α log Y ^ x , y , c , Y x , y , c = 1 1 N x , y , c 1 Y x , y , c β Y ^ x , y , c α log 1 Y ^ x , y , c ,   otherwise  
In CenterNet, the hyperparameters a and β of the focal losses were considered as well as the number of key points in the image N and the predicted value Y ^ x , y , c . According to the original method, we chose α = 2 and β = 4 and normalized all the focal losses. In addition, when dealing with easily classifiable samples, we reduced their training weight appropriately.
During the downsampling process of the model, the resolution of the feature maps extracted by the backbone network decreases due to the output span, which leads to the loss of details in the images and, thus, causes a bias in the prediction results. Therefore, we introduced the centroid bias value O ^ R W R H R 2 , used the L 1 loss function for training, and defined the centroid bias loss L off as:
L off = 1 N p O ^ p ˜ p R p ˜
where R is the sampling factor, typically R = 4, p is the target frame centroid, and p R p ˜ is the bias value.
Let the target bounding box coordinates be x 1 ( k ) , y 1 ( k ) , x 2 ( k ) , y 2 ( k ) ; then the centroid can be denoted as x 1 ( k ) + x 2 ( k ) 2 , y 1 ( k ) + y 2 ( k ) 2 , and the same L 1 loss function is used for training, with the width–height loss L s i z e defined as:
L size   = 1 N k = 1 N S ^ p k s k
Combining the above types of losses gives the total loss L det   , which is expressed as:
L det   = L k + λ size   L size + λ off   L off  
where the weights λ off   and L size are taken as 1 and 0.1, respectively.

4. Experiments and Analysis

In this section, we validate the effectiveness of the proposed method through several sets of experiments. First, we conducted experiments on four types of underwater image datasets, including comparison experiments with popular target detection algorithms, ablation studies, and comparison experiments with existing underwater target detection algorithms. Then, we conducted experiments on the PASCAL VOC2007 dataset to verify that our method not only performs well on underwater datasets but is also applicable to standard datasets.

4.1. Training Details

The experimental environment is Python 3.8, Pytorch 1.7.0, Torchvision 0.7.0, and the rest of the Python modules can be downloaded and installed from third parties. The hardware consists of an Intel® Core™ i9-10900X CPU and an NVIDIA RTX308Ti graphics card. The Intel® Core™ i9-10900X CPU is manufactured by Intel, while the RTX3080Ti graphics card is manufactured by NVIDIA. Intel is headquartered in Santa Clara, CA, USA, and NVIDIA also has its headquarters there. For training, the input image size was 416 × 416. A total of 200 epochs were used to train the network. The learning rate is 0.001 for the first 25 epochs, 0.0001 for epochs 25 to 100, and 0.00001 for epochs 100 to 200. A total of 1000 iterations are performed for each of the first 25 epochs. After 25 epochs, 2000 iterations are performed for each epoch. To achieve higher accuracy and generalization performance, data enhancement is performed on underwater data. By introducing color dithering, horizontal flipping, random rotation, and random cropping, it adapts to the color state changes of underwater targets under different shooting conditions.

4.2. Evaluation Metrics for Model Algorithms

The performance of the network is objectively evaluated using three metrics: frames per second (FPS), average precision (AP), and mean average precision (mAP) as follows:
(1)
The speed of network detection is evaluated using the FPS metric, with higher values indicating faster network detection;
(2)
The AP metric is used to evaluate the detection accuracy of each category, which can be calculated as follows:
A P = 0 1 P ( R ) d R
where P is precision, the probability that all detected targets are correctly detected, and R is recall, the probability that all true positives are correctly detected, which can be calculated as follows:
P = T P T P + F P ,   R = T P T P + F N
where TP is the number of positive samples correctly recognized, FP is the number of samples recognized as positive but actually negative, and FN is the number of samples recognized as negative but actually positive;
(3)
The combined detection accuracy of the model is evaluated using mAP, and the calculation can be expressed as follows:
m A P = i = 1 N A P i / N
where N is the number of categories and APi is the average precision of category i;
(4)
The model complexity uses parameters, and the specific calculation formula is as follows:
P a r a m s = C o   ×   ( k w   ×   k h   ×   C i + 1 )
where Co represents the number of output channels, Ci represents the number of input channels, and kw and kh represent the width and height of the convolution kernel, respectively.

4.3. Experiments on Underwater Image Datasets

The underwater image datasets URPC2020 and URPC2018 contain 5543 and 4000 images, respectively, covering four classes of marine life, namely holothurian, starfish, echinus, and scallop. Due to the realism of the marine environment, these images typically suffer from low resolution, low signal-to-noise ratio, low contrast, color aberration, and blurred feature information, resulting in the loss of clear contour and texture information, limiting the accuracy of target identification and greatly reducing the effectiveness of our method. In addition, targets are dense and often obscured, presenting significant challenges to target detection. In underwater imagery, the scale of marine life is often small. Of particular note is the fact that some marine organisms, such as sea cucumbers and scallops, have protective colors that blend in with their environment and can hide. Also, due to the aggregation behavior of marine life, there is often a high density of targets in the captured images. All these characteristics make the task of underwater target detection difficult. In order to validate our proposed algorithm, we conducted a series of experiments on an underwater image dataset.

4.3.1. Analyze the Effect of Image Resolution and Signal-to-Noise Ratio on Our Algorithm

In underwater environments, light propagation is affected by water absorption and scattering, resulting in distorted and blurred images. With low resolution, target details may not be clearly visible in the image, affecting the accuracy of target detection. Higher-resolution images can provide more detailed information that can help identify and locate underwater targets. In addition, the scattering and absorption of light introduces noise that can degrade image quality and obscure the target. If the signal-to-noise ratio is low, the target is drowned out by the noise, making it difficult for the detection algorithm to identify it correctly.
Looking at the experimental results in Table 2, the algorithm achieves an average detection accuracy of 80.55% when the input size is 384 × 384. When the input size is increased to 416 × 416, the average detection accuracy reaches 81.17%, which is a 0.62% improvement in detection accuracy. Experiments on the URPC2020 underwater image dataset show that a larger input size can provide finer target information, thus improving target detection accuracy.
Observing the experimental results in Table 3, we add Gaussian noise, which has a greater impact on the underwater image (“+” indicates the addition of Gaussian noise); by adding Gaussian noise, the clarity of the image decreases, reducing the quality and legibility of the image. Gaussian noise reduces the overall quality of the image, thus limiting the effective resolution of the image. Although the image has a certain number of pixels, some details and features cannot be clearly represented due to the presence of noise. Experiments on the URPC2020 underwater image dataset show that the proposed algorithm achieves a detection accuracy of 81.17%, whereas after adding noise to the image, the accuracy rate of the algorithm in this paper decreases by 1.39%.
As can be seen in Figure 11, our algorithm can correctly detect and identify underwater organisms. The introduction of Gaussian noise in the underwater image reduces the effective resolution of the image, which affects the accuracy of the target detection algorithm. Due to the presence of noise, some details of the target features are missed, resulting in a decrease in target detection accuracy.
In summary, underwater target detection is affected by resolution and the signal-to-noise ratio, making detection difficult. Reducing the input resolution can lead to distortion or loss of some small targets, reducing the detection accuracy of small targets. Increasing the input resolution provides more target detail and helps to detect small targets. Higher-resolution images provide more accurate target localization, allowing the position of the target to be determined more accurately. Higher-resolution images reduce noise and interference in the image, reducing the likelihood of false detections and improving the reliability of the algorithm. In addition, a lower signal-to-noise ratio makes the noise in the image more prominent, blurring the features of the target in the image and, thus, reducing the visibility of the target. A low signal-to-noise ratio can lead to an increase in false alarms. Noise can be mistaken for a target, which can lead to false alarms in the algorithm, affecting the accuracy and reliability of detection. This has an impact on the ability to detect and locate targets in the underwater environment, which ultimately affects underwater target acquisition results.

4.3.2. Comparison with Popular Object Detection Algorithms

We compared several popular target detection methods on the underwater image dataset URPC2020 and recorded the results as shown in Table 4 Our method is compared with Faster-RCNN, YOLOv5, YOLOv7, SSD, DSSD, SWIPENet, CornerNet, and CenterNet. Faster-RCNN is a two-stage target detection algorithm, while YOLOv5, SSD, and RetinaNet are classical one-stage target detection algorithms. SWIPNet is the most commonly used underwater target detection algorithm. We evaluate the effectiveness of target detection in terms of mean average precision (mAP) and frames per second (FPS). As can be seen in Table 4, Faster-RCNN achieves 71.45% mAP when implemented with VGGNet and 74.82% mAP when implemented with ResNet-101 + FPN, while the FPS decreases. However, YOLOv5 achieved 78.34% accuracy on mAP and had an FPS of 14.6, while YOLOv7 achieved 79.49% accuracy on mAP and an FPS of 17.5. The YOLOv5 and YOLOv7 methods showed high performance in terms of detection accuracy and speed. SSDs and DSSDs are almost equal in performance. SWIPTNet was used as a novel re-sampling weighted underwater algorithm to reduce the effect of underwater noise, but performance was average. The detection accuracy of the anchor-free detector CornerNet was significantly lower. This difference can be attributed to the fact that the above algorithms require several key corner points to be defined for underwater target detection and grouped according to the distance between these corner points. However, due to the presence of strong occlusion and a large number of small targets in the underwater environment, the problem of misgrouping can easily occur, which significantly reduces the detection accuracy. The method proposed in this paper achieves a detection accuracy of 81.17% with an FPS of 7.4, which meets the real-time requirement. In the following, we further analyze the advantages of the method.
In order to evaluate the effectiveness of the algorithm proposed in this paper, experiments were performed on the URPU2020 underwater object detection dataset, and we compared and analyzed its model complexity with that of the CenterNet algorithm. The experimental results are presented in Table 5. Using identical experimental settings, the proposed algorithm has a much lower model size, fewer model parameters, and fewer floating-point operations compared to the CenterNet algorithm. Specifically, the proposed algorithm’s model size is reduced by 638.5 M (i.e., compressed by 84.4%), the number of model parameters is reduced by 159.6 M (i.e., compressed by 84.3%), and the floating-point operation volume is reduced by 138.3 G (i.e., scaled down by 81.6%). These results demonstrate that the algorithm proposed in this paper significantly reduces the model complexity, making it more favorable for practical applications.
In addition, this paper performs experiments on the underwater image dataset URPC2018 to compare with the baseline network. The results are presented in Table 6. From Table 6, it can be seen that the algorithm in this paper outperforms the baseline network with faster detection speed, obtaining 77.02% mAP on URPC2018. This further demonstrates the effectiveness of the proposed parallel high-resolution network.
Table 7 displays detection results on the underwater image dataset URPC2018. SWIPTNet is a novel re-sampling weighted underwater algorithm used to reduce the effect of underwater noise; its detection accuracy was only 68.0%. The RoIMix [45] data enhancement is used to improve the detection of overlapping, occluded, and blurred targets; its detection accuracy reached 74.9%. It can be seen that the algorithm in this paper has a great advantage in detection accuracy, surpassing existing underwater target detection algorithms.

4.3.3. Ablation Experiments

To validate the algorithms in this paper, we perform ablation experiments to compare and analyze the detection accuracy and speed. Table 8 shows the results of the ablation experiments on the URPC2020 underwater dataset. CenterNet was used as the baseline for the experiments on the underwater dataset.
The first row represents the baseline network with Hourglass-104 as the backbone network. The second row replaces the backbone network with HRNet, which slightly reduces the average accuracy but improves the detection speed, further demonstrating that HRNet is a lighter network that reduces model complexity and improves detection speed. With the addition of the enhanced attention module A-CBAM in the third row, excellent spatial information modeling capability is achieved by modeling the two-dimensional space in the activation function stage, capturing complex feature distributions and demonstrating feature enhancement in the spatial and channel dimensions, which improves the quality of the feature maps extracted by the network and increases the detection accuracy as well as the speed of inference. The fourth line introduces a receptive field augmentation module, which makes full use of the semantic and positional information of the network to further improve accuracy while maintaining detection speed.
In the evaluation of detection accuracy presented in Table 8, the categories corresponding to echinus and starfishes exhibit a more substantial number of targets and more distinctly delineated feature information. Therefore, under different experimental settings, the algorithm in this paper can achieve 91.32% and 85.32% accuracy for echinus and starfishes, respectively, because the model can easily extract target features and achieve high detection accuracy. However, the total number of targets in the two categories of holothurians and scallops was low. As these two target types are usually small in size and have ambiguous feature information, their detection accuracy is low. The detection accuracy of the algorithm in this paper for these two types of targets reached 76.40% and 71.67%, respectively, which is significantly higher than that of other networks. This indicates that the feature extraction capability of the network is significantly improved by the combination of different enhancement strategies proposed in this paper. This alleviates the limitations imposed by the lack of target data samples and improves the detection of small and fuzzy targets.

4.4. Experiments on PASCAL VOC Datasets

To further validate the performance of the algorithms in this paper, we conducted experiments on the PASCAL VOC dataset. We used the VOC2007 and VOC2012 training sets to train the network model and tested the network model on the VOC2007 test set. Table 9 shows the results of our method in comparison to currently popular target detection algorithms. The methods compared include one-stage and two-stage target detection algorithms. In general, two-stage target detection algorithms have higher detection accuracy but slower detection speed. For example, the Faster R-CNN using ResNet-101 improved the mAP from 73.2% to 76.4% compared to the VGGNet backbone. However, the single-stage target detection algorithm has the advantage of fast detection. For example, SSD512 has a detection speed of 19 FPS and YOLOv5 has a detection speed of 38 FPS. In addition, we can sacrifice some detection speed for higher detection accuracy. When the input size is 416 × 416, the mAP of the algorithm in this paper reaches 82.5%, which is higher and faster than the Faster R-CNN two-stage target detection algorithms; when the input size is increased to 512 × 512, the average detection accuracy reaches 82.9%, which is higher than that of the current mainstream single-stage target detection algorithms SSD, YOLOv5, and RetinaNet; the detection speed is reduced but the detection accuracy is higher. This shows that the algorithm in this paper still has advantages in the detection of general-purpose objects, which proves the reasonableness of the network structure design in this paper.

4.5. Visualization Analysis of Detection Results Output

In order to show the detection results of the algorithm in a more intuitive way, four images from the more complex dataset were randomly selected and compared with the original model, as shown in Figure 12. Echinus, scallops, starfishes, and holothurians are still relatively blurred underwater and highly integrated with the background, making detection difficult. The algorithm in this paper was able to correctly detect the target underwater marine organisms with high accuracy, proving the effectiveness of the improved model.
As shown in Figure 13. By analyzing the above four images, we can conclude that the algorithm in this paper has good detection capability in the face of seagrass clutter, complex environments, and blurred overlapping targets. The comprehensive evaluation of the experimental results shows that the surface algorithm is a detection model with high generalization ability and can accurately detect targets of different states. It can be concluded that the algorithm in this paper has excellent comprehensive performance in underwater target detection. Based on the above results, further comparisons with the state of the art are made.

4.6. Comparison of the Results Obtained with the State of the Art

Based on the above results, further analysis and comparison with existing techniques are performed. The overall performance of our method is improved, and we compare the results of experiments on the underwater dataset URPC2020; is the performance of our method is higher than that of other mainstream target detection algorithms in terms of detection accuracy and higher than the latest YOLOv7 model’s detection accuracy. This shows that the improved algorithm in this paper has more advantages in detecting underwater targets and can achieve higher detection accuracy. It is worth noting that in the detection of underwater images, anchor-free frame detector CornerNet’s detection accuracy is significantly lower than that of this paper’s algorithm and other algorithms. To analyze the reasons why, the above algorithm needs to define a number of key corner points and group them according to the distance between the key points; underwater targets with severe occlusion are present; a wide range of small targets are present; and it is easy to misgroup, significantly reducing the detection accuracy from the point of view of detection speed. In terms of detection speed, the algorithm in this paper is faster than SWIPTNet, Faster R-CNN, CornerNet, and CenterNet, but slower than SSD, RetinaNet, YOLOv5, and YOLOv7 algorithms. The reason for this is that the backbone network of this paper is based on HRNet, which has fewer parameters and improves the detection speed, but due to the repeated multi-scale feature fusion, it is easier to misgroup underwater targets, which significantly reduces the detection accuracy. However, due to the repeated multi-scale feature fusion within the network, the network inference speed is reduced to some extent; in terms of model complexity, the model size, number of parameters, and floating-point operation volume in this paper are only 117.6 M, 29.7 M, and 31. 2 G, which are much lower than those of the benchmark network, indicating that the improved network in this paper is a relatively lightweight network, which can significantly reduce the number of parameters in the model, reduce the complexity of the model, and improve the network inference speed.
Replacing Hourglass-104 with HRNet as the backbone of the network results in a slight decrease in average accuracy and an increase in inference speed, further suggesting that HRNet is a lighter network that reduces model complexity and improves recognition speed. With the addition of the enhanced attention module A-CBAM, excellent spatial information modeling capability is achieved by modeling the two-dimensional space in the activation function stage, capturing complex feature distributions and demonstrating feature enhancement in the spatial and channel dimensions, which improves the quality of the feature maps extracted by the network and significantly improves both detection accuracy and inference speed. The receptive field augmentation module is introduced to fully exploit the semantic and positional information of the network to further improve the accuracy while maintaining detection speed.
For each type of detection accuracy, both echinus and starfishes have a larger number of targets and clearer feature information. Therefore, under different experimental settings, the model can easily extract target features and achieve high detection accuracy, and the algorithm in this paper can achieve 91.32% and 85.32% accuracy for echinus and starfishes, respectively. However, the total number of targets in the two categories of holothurians and scallops is small. As these two target types are usually small and have fuzzy feature information, their detection accuracies are lower. However, the algorithms in this paper achieve detection accuracies of 76.40% and 71.67% for these two types of targets, which is significantly higher than that of other networks. This indicates that the feature extraction capability of the network is significantly improved by the combination of different enhancement strategies proposed in this paper. To a certain extent, this alleviates the limitation of the scarcity of target data samples and improves the detection ability for small and fuzzy targets.
Comprehensively analyzed, the algorithm in this paper achieves the highest detection accuracy compared to other mainstream algorithms, while maintaining a faster detection speed and a lower number of model parameters, which has a significant advantage in detecting underwater targets.

5. Conclusions

To solve a number of problems such as complex underwater scenes and difficult target feature extraction, this paper makes targeted improvements to the CenterNet benchmark framework and proposes a parallel high-resolution underwater target detection algorithm. First, HRNet, a lighter high-resolution human posture estimation network, is used to improve the target feature representation and effectively reduce the semantic information lost in the image sampling process. Then, the attention module (A-CBAM) is enhanced to achieve pixel-level spatial information modeling capability by introducing the FReLU activation function, which captures complex feature distributions by modeling the two-dimensional space in the activation function stage. Feature enhancement in the spatial and channel dimensions is performed to improve understanding of fuzzy targets and small target objects and to better capture irregular and detailed object layouts. Finally, a receptive field augmentation module (RFAM) is constructed to obtain sufficient semantic information and rich detail information to further enhance the robustness and discrimination of features and improve the detection capability of the model for multi-scale underwater targets. The method improves the mAP by 4.96%, 2.62%, and 2.2% on the three publicly available datasets URPC2020, URPC2018, and PASCAL VOC2007, respectively. The experiments on the underwater target detection dataset URPU2020 show that the number of model parameters of the model in this paper is reduced by 159.6 M and compressed by 84.3% compared with the benchmark network. These results show that the algorithm proposed in this paper improves the accuracy and reduces the model complexity, which makes it more conducive to practical applications. Meanwhile, the surface algorithm is not only applicable to the underwater environment, but also to the surface environment. The strong generalization ability of the algorithm is verified.
There are several threats to the study of underwater object detection that affect the validity of the work in this paper. At the model level, the use of parallel high-resolution networks and feature extraction processes leads to problems of information loss and redundant parameters. The feature extraction capability of the high-resolution network is also somewhat lacking. A receptive field augmentation module was introduced but, at the same time, it increased the complexity of the model and reduced the speed of model inference. From an image point of view, the number of two types of targets, holothurians and scallops, is low. Since these two types of targets are usually small in size and have ambiguous feature information, their detection accuracy is low.
Future research will focus on improving model performance, increasing underwater image datasets, further exploring the potential of convolutional neural networks for marine target detection, and improving detection of fuzzy and small targets, as well as performance in turbid and crowded environments. Methodologically, we will investigate techniques such as model pruning and knowledge distillation and continue to explore lightweight processing approaches to improve the detection speed of the model.

Author Contributions

Conceptualization, Z.B. and Y.G.; methodology, Z.B.; software, Z.B.; validation, Z.B.; formal analysis, Y.G.; investigation, J.W. and S.Y.; resources, Y.G.; data curation, J.H.; writing—original draft preparation, Z.B.; writing—review and editing, Y.G.; visualization, Z.B.; supervision, L.Z.; project administration, L.Z.; funding acquisition, Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61971229).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, M.F.R.; Chen, Y.C. Artificial intelligence based object detection and tracking for a small underwater robot. Processes 2023, 11, 312. [Google Scholar] [CrossRef]
  2. Wang, Y.; Fu, B.; Fu, L.; Xia, C. In situ sea cucumber detection across multiple underwater scenes based on convolutional neural networks and image enhancements. Sensors 2023, 23, 2037. [Google Scholar] [CrossRef]
  3. Li, J.; Lian, Z.; Wu, Z.; Zeng, L.; Mu, L.; Yuan, Y.; Bai, H.; Guo, Z.; Mai, K.; Tu, X. Artificial intelligence–based method for the rapid detection of fish parasites (Ichthyophthirius multifiliis, Gyrodactylus kobayashii, and Argulus japonicus). Aquaculture 2023, 563, 738790. [Google Scholar] [CrossRef]
  4. Lv, J.; Zhu, D.; Geng, Z.; Han, S.; Wang, Y.; Yang, W.; Ye, Z.; Zhou, T. Recognition of deformation military targets in the complex scenes via MiniSAR submeter images with FASAR-Net. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5209219. [Google Scholar] [CrossRef]
  5. Zhou, J.; Sun, J.; Zhang, W.; Lin, Z. Multi-view underwater image enhancement method via embedded fusion mechanism. Eng. Appl. Artif. Intell. 2023, 121, 105946. [Google Scholar] [CrossRef]
  6. Zhou, J.; Pang, L.; Zhang, D.; Zhang, W. Underwater image enhancement method via multi-interval subhistogram perspective equalization. IEEE J. Ocean. Eng. 2023, 48, 474–488. [Google Scholar] [CrossRef]
  7. Priya, D.S.; Balakrishna, P.; Sai, R.C.; siva Bhavana, C.N.; Uday, Y.; Joyel, Y. Underwater image enhancement via minimal color loss and locally adaptive contrast enhancement. J. Eng. Sci. 2023, 14, 49–55. [Google Scholar]
  8. Mehranian, A.; Wollenweber, S.D.; Walker, M.D.; Bradley, K.M.; Fielding, P.A.; Su, K.H.; Johnsen, R.; Kotasidis, F.; Jansen, F.P.; McGowan, D.R. Image enhancement of whole-body oncology [18F]-FDG PET scans using deep neural networks to reduce noise. Eur. J. Nucl. Med. Mol. Imaging 2022, 49, 539–549. [Google Scholar] [CrossRef]
  9. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  10. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  11. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 13–16 December 2015; pp. 1440–1448. [Google Scholar]
  12. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  13. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  15. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  16. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  17. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. Dssd: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  18. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  19. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  20. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  21. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  22. Victor, N.; Chengoden, R.; Alazab, M.; Bhattacharya, S.; Magnusson, S.; Maddikunta, P.K.R.; Ramana, K.; Gadekallu, T.R. Federated learning for iout: Concepts, applications, challenges and future directions. IEEE Intern. Things Mag. 2022, 5, 36–41. [Google Scholar] [CrossRef]
  23. Meng, Z.; Li, Z.; Hou, X.; Du, J.; Chen, J.; Wei, W. Efficient asynchronous federated learning for AUV swarm. Sensors 2022, 22, 8727. [Google Scholar] [CrossRef]
  24. Bhattacharya, S.; Victor, N.; Chengoden, R.; Ramalingam, M.; Selvi, G.C.; Maddikunta, P.K.R.; Donta, P.K.; Dustdar, S.; Jhaveri, R.H.; Gadekallu, T.R. Blockchain for internet of underwater things: State-of-the-art, applications, challenges, and future directions. Sustainability 2022, 14, 15659. [Google Scholar] [CrossRef]
  25. Shaik, A.S.; Karsh, R.K.; Islam, M.; Laskar, R.H. A review of hashing based image authentication techniques. Multimed. Tools Appl. 2022, 81, 2489–2516. [Google Scholar] [CrossRef]
  26. Shaik, A.S.; Karsh, R.K.; Suresh, M.; Gunjan, V.K. LWT-DCT based image hashing for tampering localization via blind geometric correction. In Proceedings of the ICDSMLA 2020: Proceedings of the 2nd International Conference on Data Science, Machine Learning and Applications, Singapore, 22 December 2022; pp. 1651–1663. [Google Scholar]
  27. Shaik, A.S.; Karsh, R.K.; Islam, M.; Singh, S.P. A secure and robust autoencoder-based perceptual image hashing for image authentication. Wirel. Commun. Mob. Comput. 2022, 2022, 1645658. [Google Scholar] [CrossRef]
  28. Zhou, H.; Huang, H.; Yang, X.; Zhang, L.; Qi, L. Faster R-CNN for marine organism detection and recognition using data augmentation. In Proceedings of the Faster R-CNN for Marine Organism Detection and Recognition Using Data Augmentation, Singapore, 27–29 December 2017; pp. 56–62. [Google Scholar]
  29. Villon, S.; Chaumont, M.; Subsol, G.; Villéger, S.; Claverie, T.; Mouillot, D. Coral reef fish detection and recognition in underwater videos by supervised machine learning: Comparison between Deep Learning and HOG+ SVM methods. In Proceedings of the 17th International Conference on Advanced Concepts for Intelligent Vision Systems (ACIVS 2016), Lecce, Italy, 24–27 October 2016; pp. 160–171. [Google Scholar]
  30. Chen, L.; Liu, Z.; Tong, L.; Jiang, Z.; Wang, S.; Dong, J.; Zhou, H. Underwater object detection using Invert Multi-Class Adaboost with deep learning. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  31. Ma, K.; Huang, B.; Yin, H. Underwater sea cucumbers detection based on improved SSD. In Proceedings of the 2019 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS), Shenyang, China, 12–14 July 2019; pp. 343–347. [Google Scholar]
  32. Qiu, Z.; Yao, Y.; Zhong, M. Underwater sea cucumbers detection based on pruned SSD. In Proceedings of the 2019 IEEE 3rd Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 11–13 October 2019; pp. 738–742. [Google Scholar]
  33. Li, Y.; Bai, X.; Xia, C. An improved YOLOV5 based on triplet attention and prediction head optimization for marine organism detection on underwater mobile platforms. J. Mar. Sci. Eng. 2022, 10, 1230. [Google Scholar] [CrossRef]
  34. Zhang, L.; Xing, B.; Wang, W.; Xu, J. Sea cucumber detection algorithm based on deep learning. Sensors 2022, 22, 5717. [Google Scholar] [CrossRef]
  35. Yao, Y.; Qiu, Z.; Zhong, M. Application of improved MobileNet-SSD on underwater sea cucumber detection robot. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019; pp. 402–407. [Google Scholar]
  36. Guo, X.; Zhao, X.; Liu, Y.; Li, D. Underwater sea cucumber identification via deep residual networks. Inform. Process. Agric. 2019, 6, 307–315. [Google Scholar] [CrossRef]
  37. Jiang, L.; Wang, Y.; Jia, Q.; Xu, S.; Liu, Y.; Fan, X.; Li, H.; Liu, R.; Xue, X.; Wang, R. Underwater species detection using channel sharpening attention. In Proceedings of the Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 4259–4267. [Google Scholar]
  38. Huang, J.; Zhu, Z.; Huang, G. Multi-stage HRNet: Multiple stage high-resolution network for human pose estimation. arXiv 2019, arXiv:1910.05901. [Google Scholar]
  39. Ma, N.; Zhang, X.; Sun, J. Funnel activation for visual recognition. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 351–368. [Google Scholar]
  40. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  42. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  43. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  44. Das, D.; Santosh, K.C.; Pal, U. Truncated inception net: COVID-19 outbreak screening using chest X-rays. Phys. Eng. Sci. Med. 2020, 43, 915–925. [Google Scholar] [CrossRef] [PubMed]
  45. Lin, W.H.; Zhong, J.X.; Liu, S.; Li, T.; Li, G. Roimix: Proposal-fusion among multiple images for underwater object detection. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 2588–2592. [Google Scholar]
Figure 1. Overall network structure.
Figure 1. Overall network structure.
Sensors 23 07337 g001
Figure 2. HRNet structure.
Figure 2. HRNet structure.
Sensors 23 07337 g002
Figure 3. Multi-resolution module.
Figure 3. Multi-resolution module.
Sensors 23 07337 g003
Figure 4. Specific fusion method.
Figure 4. Specific fusion method.
Sensors 23 07337 g004
Figure 5. A-CBAM module.
Figure 5. A-CBAM module.
Sensors 23 07337 g005
Figure 6. Channel attention module.
Figure 6. Channel attention module.
Sensors 23 07337 g006
Figure 7. Spatial attention module.
Figure 7. Spatial attention module.
Sensors 23 07337 g007
Figure 8. Comparison of ReLU activation function and FReLU activation function. (a) ReLU activation function. (b) FReLU activation function.
Figure 8. Comparison of ReLU activation function and FReLU activation function. (a) ReLU activation function. (b) FReLU activation function.
Sensors 23 07337 g008
Figure 9. Comparison of ordinary convolution and dilated convolution. (a) The 3 × 3 convolution kernel. (b) Schematic of the response to a 3 × 3 convolution operation. (c) The 3 × 3 dilated convolution kernel. (d) Schematic of the response to a 3 × 3 dilated convolution operation.
Figure 9. Comparison of ordinary convolution and dilated convolution. (a) The 3 × 3 convolution kernel. (b) Schematic of the response to a 3 × 3 convolution operation. (c) The 3 × 3 dilated convolution kernel. (d) Schematic of the response to a 3 × 3 dilated convolution operation.
Sensors 23 07337 g009
Figure 10. Comparison of receptive field module and receptive field augmentation module. (a) Receptive field module (RFM). (b) Receptive field augmentation module (RFAM).
Figure 10. Comparison of receptive field module and receptive field augmentation module. (a) Receptive field module (RFM). (b) Receptive field augmentation module (RFAM).
Sensors 23 07337 g010
Figure 11. The comparison result of adding Gaussian noise and not adding Gaussian noise. (a) Detection results with Gaussian noise added. (b) Detection results without Gaussian noise added.
Figure 11. The comparison result of adding Gaussian noise and not adding Gaussian noise. (a) Detection results with Gaussian noise added. (b) Detection results without Gaussian noise added.
Sensors 23 07337 g011
Figure 12. Comparison results between the algorithm in this paper and the baseline in the underwater image dataset URPC2020. (a) Baseline. (b) Ours.
Figure 12. Comparison results between the algorithm in this paper and the baseline in the underwater image dataset URPC2020. (a) Baseline. (b) Ours.
Sensors 23 07337 g012
Figure 13. Detection results of different scenarios. (a) Seagrass disturbance. (b) Target overlap. (c) Complex background. (d) Long range small target.
Figure 13. Detection results of different scenarios. (a) Seagrass disturbance. (b) Target overlap. (c) Complex background. (d) Long range small target.
Sensors 23 07337 g013
Table 1. Methods for summarizing related work.
Table 1. Methods for summarizing related work.
MethodYearKey Features
Zhou et al. [28]2017The combination of image enhancement techniques and an extended VGG16 feature extraction network is presented.
Villon et al. [29]2020Combined HOG and SVM.
Chen et al. [30]2020A novel sample distribution-based weighted loss function called IMA is introduced.
Ma et al. [31,32]2019Lightweight models were developed by pruning the SDD target detection model or integrating the MobileNet lightweight backbone.
Li et al. [33]2022Added an extra detection header to the YOLOv5 model.
Zhang et al. [34]2022Incorporated attentional mechanisms to learn about target features and ways to increase the receptive field.
Yao et al. [35,36]2019Used a residual network as a backbone.
Jiang et al. [37]2021A method called the Channel Sharpening Attention Module (CSAM) is proposed.
Table 2. Comparison results for different input sizes.
Table 2. Comparison results for different input sizes.
NetworkBackboneInput SizeAP (%)mAP (%)
EchinusHolothurianScallopStarfish
OursAR-HRNet416 × 41691.3276.4071.6785.3281.17
OursAR-HRNet384 × 38491.8975.3470.8684.1380.55
Table 3. Comparison of detection accuracy with and without the addition of Gaussian noise.
Table 3. Comparison of detection accuracy with and without the addition of Gaussian noise.
NetworkBackboneAP (%)mAP (%)
EchinusHolothurianScallopStarfish
OursAR-HRNet91.3276.4071.6785.3281.17
Ours+AR-HRNet89.4775.2369.5984.8379.78
Table 4. Comparison results with popular object detector algorithms on URPC2020.
Table 4. Comparison results with popular object detector algorithms on URPC2020.
NetworkBackboneAP (%)mAP (%)FPS
EchinusHolothurianScallopStarfish
Faster R-CNNResNet-101 + FPN86.3363.1769.2580.5674.823.4
Faster-RCNNVGG-1680.6364.0162.7878.3871.456.16
YOLOv5CSP-Darknet5387.7167.6975.3381.6378.3414.6
YOLOv7CSP-Darknet5386.6470.2877.8183.2579.4917.5
SSD512VGG-1684.5069.3066.9080.5075.3014.3
DSSDResidual-10183.6270.3565.6379.7274.8313.2
SWIPENet VGG-1684.3270.7549.6279.2370.989.8
RetinaNetResNet-5083.4262.8369.0380.9474.059.56
CornerNetHouglass-10465.7252.2157.2160.4858.903.1
CenterNetHouglass-10486.4369.7868.0780.5976.215.3
OursAR-HRNet91.3276.4071.6785.3281.177.4
Table 5. Comparison with CenterNet algorithm model complexity.
Table 5. Comparison with CenterNet algorithm model complexity.
NetworkBackboneSize/MParams/MGFLOPs
CenterNetHouglass-104756.1189.3169.5
OursAR-HRNet117.629.731.2
Table 6. Comparison results with popular object detection algorithms on URPC2018.
Table 6. Comparison results with popular object detection algorithms on URPC2018.
NetworkBackboneAP (%)mAP (%)FPS
EchinusHolothurianScallopStarfish
CornerNetHouglass-10461.4545.9651.2153.3750.903.1
CenterNetHouglass-10486.9469.9258.3782.3774.405.3
OursAR-HRNet87.7172.5467.2680.5977.027.4
Table 7. Comparison with existing underwater object detection algorithms on URPC2018.
Table 7. Comparison with existing underwater object detection algorithms on URPC2018.
NetworkBackbonemAP (%)
SWIPTENetSWIPTENet68.0
ROIMixResNet-10174.9
OursAR-HRNet77.02
Table 8. Ablation experiments on URPC2020.
Table 8. Ablation experiments on URPC2020.
HRNetA-CBAMRFABEchinusHolothurianScallopStarfishmAP (%)FPS
86.4369.7868.0780.5976.215.3
85.2469.2866.4679.3275.078.3
88.2673.3268.5282.5778.167.8
91.3276.4071.6785.3281.177.4
Table 9. Detection results on the PASCAL VOC2007 datasets.
Table 9. Detection results on the PASCAL VOC2007 datasets.
NetworkBackboneInput SizeFPSmAP (%)
Faster R-CNNResNet-1011000 × 600576.4
Faster R-CNNVGG-161000 × 6006.773.2
YOLOv5CSP-Darknet53416 × 4163878.9
SSD512VGG-16512 × 5121976.8
DSSDResNet-101312 × 3129.578.6
RetinaNetResNet-501000 × 6007.674.8
CenterNetResNet-101416 × 4162678.7
CenterNetHouglass-104416 × 416580.7
OursAR-HRNet416 × 416882.5
OursAR-HRNet512 × 5127.882.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bao, Z.; Guo, Y.; Wang, J.; Zhu, L.; Huang, J.; Yan, S. Underwater Target Detection Based on Parallel High-Resolution Networks. Sensors 2023, 23, 7337. https://doi.org/10.3390/s23177337

AMA Style

Bao Z, Guo Y, Wang J, Zhu L, Huang J, Yan S. Underwater Target Detection Based on Parallel High-Resolution Networks. Sensors. 2023; 23(17):7337. https://doi.org/10.3390/s23177337

Chicago/Turabian Style

Bao, Zhengwei, Ying Guo, Jiyu Wang, Linlin Zhu, Jun Huang, and Shu Yan. 2023. "Underwater Target Detection Based on Parallel High-Resolution Networks" Sensors 23, no. 17: 7337. https://doi.org/10.3390/s23177337

APA Style

Bao, Z., Guo, Y., Wang, J., Zhu, L., Huang, J., & Yan, S. (2023). Underwater Target Detection Based on Parallel High-Resolution Networks. Sensors, 23(17), 7337. https://doi.org/10.3390/s23177337

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop