Next Article in Journal
Modeling and Prediction of Daily Traffic Patterns—WASK and SIX Case Study
Next Article in Special Issue
Implementation of an Award-Winning Invasive Fish Recognition and Separation System
Previous Article in Journal
Classifier Performance Evaluation for Lightweight IDS Using Fog Computing in IoT Security
Previous Article in Special Issue
CNN Algorithm for Roof Detection and Material Classification in Satellite Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Target Recognition Based on Improved YOLOv4 Neural Network

College of Mechanical Engineering and Automation, Fuzhou University, Fuzhou 350108, China
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(14), 1634; https://doi.org/10.3390/electronics10141634
Submission received: 8 June 2021 / Revised: 5 July 2021 / Accepted: 7 July 2021 / Published: 9 July 2021
(This article belongs to the Special Issue Convolutional Neural Networks and Vision Applications)

Abstract

:
The YOLOv4 neural network is employed for underwater target recognition. To improve the accuracy and speed of recognition, the structure of YOLOv4 is modified by replacing the upsampling module with a deconvolution module and by incorporating depthwise separable convolution into the network. Moreover, the training set used in the YOLO network is preprocessed by using a modified mosaic augmentation, in which the gray world algorithm is used to derive two images when performing mosaic augmentation. The recognition results and the comparison with the other target detectors demonstrate the effectiveness of the proposed YOLOv4 structure and the method of data preprocessing. According to both subjective and objective evaluation, the proposed target recognition strategy can effectively improve the accuracy and speed of underwater target recognition and reduce the requirement of hardware performance as well.

1. Introduction

There are countless oil and gas, mineral, biological, and chemical resources in the vast ocean. With the continuous development of science and technology, we have a better understanding of the ocean. In the process of ocean exploration, there are many underwater tasks, such as target positioning [1], biological recognition [2], archaeology [3], object search [4], environment detection [5], device maintenance [6], etc. In accomplishing these tasks, underwater target recognition plays an important role. Underwater target recognition depends on an underwater vision system, which can be divided into acoustic vision systems and optical vision systems [7]. Based on the acquired acoustic information, one can perform image processing, feature extraction, classification, and recognition [8]. However, there are some defects in underwater acoustic vision images, such as noise, low resolution, and less image information [9]. In recent years, with the development of optical equipment, optical vision system-based underwater target recognition has received increasing attention. From the point of view of methodology, many target recognition algorithms have been put forward and applied in underwater scenarios.
Target recognition, also known as target detection, is a kind of computer vision task. It aims to identify specific types of visual objects (such as human, animal, or vehicle) in digital images. Most of the early target recognition algorithms are based on handcrafted features. Viola and Jones [10] proposed a VJ recognizer, which realized real-time face recognition for the first time. Dalal and Triggs [11] initially proposed the histogram of oriented gradient (HOG) feature descriptor. Felzenszwalb et al. proposed the deformable part model (DPM) as an extension of the HOG recognizer [12] and made various improvements [13]. DPM is regarded as the peak of traditional target recognition. However, as the performance of handcrafted features tends to be saturated, traditional target recognition hits a bottleneck. In recent years, convolutional neural networks (CNNs) have received attention and have been proven as an effective tool for classification [14]. In the application of CNNs to target recognition, Girshick et al. [15] initially proposed a CNN-based target recognition tool. He et al. [16] developed a spatial pyramid pool network (SPPnet) with a speed more than 20 times that of RCNN (Regions with CNN factures), without sacrificing recognition accuracy. In 2015, Redmon et al. [17] proposed the YOLO (You only look once) network, which is the first one-stage detector in the area of deep learning. Liu et al. [18] proposed the second one-stage detector, SSD (single shot multibox detector). Compared to YOLO, the contribution of SSD is the introduction of multi-reference and multi-resolution recognition technology, which significantly improves the recognition accuracy of the one-stage detector, especially for some small objects.
Many researchers have applied target recognition methods to underwater images. In the past, most people used traditional target recognition methods for underwater target recognition. Xu et al. [19] proposed an inhomogeneous illumination field to reduce the backward scattered background noise in underwater environments. Xu et al. [20] proposed an underwater target feature extraction method based on the singular value of a generalized S-Transform module time–frequency matrix. Ma [21] analyzed and extracted the polarization feature, edge feature, and line feature, which are more suitable for underwater environment target detection, and then used the Itti model to generate a saliency map to detect underwater targets. Wang et al. [22] adjusted the regional saliency by simulating the eye movement to change the position of the attention focus and formed the saliency region to achieve underwater object detection. Oliver et al. [23] studied the influence of different underwater point spread functions on the detection of image features using many different feature detectors, and the influence of these functions on the feature detection ability when they are used for matching and target detection. Zhang et al. [24] extracted new combined invariant moments of underwater images as the recognition features of the system, and used an artificial fish swarm algorithm (AFSA) improved neural network as the underwater target classifier. Yang et al. [25] presented a classification algorithm based on multi-feature fusion for moving target recognition. Dubreuil et al. [26] investigated underwater target detection by combining active polarization imaging and optical correlation-based approaches. Yahya et al. [27] proposed a robust target recognition algorithm using bounding box partition to overcome the problem of failure of recognition due to unmatched targets. Liu et al. [28] proposed a feature matching algorithm based on the Hough transform [29] and geometrical features for target detection in special underwater environments. Li et al. [30] proposed an algorithm for the recognition of small underwater targets based on shape characteristics.
In 2015, researchers began to apply deep learning to underwater target recognition. Li et al. [31] first applied deep CNN to underwater detection and constructed an ImageCLEF dataset, which involves 24,277 fish images belonging to 12 classes. Sun et al. [32] proposed a CNN knowledge transfer framework for underwater object recognition and tackled the problem of extracting discriminative features from relatively low-contrast images. Zhou et al. [33] proposed three data augmentation methods for a Faster R-CNN network. Park and Kang [34] improved the YOLO network and proposed a method to accurately classify objects and count them in sequential video images. Arain et al. [35] presented two novel approaches for improving image-based underwater obstacle detection by combining sparse stereo point clouds with monocular semantic image segmentation. Jia and Liu [36] designed a rapid detection method for marine animals based on MSRCR (multiscale Retinex with color restore) and YOLOv3 to solve the problems of insufficient illumination of the underwater environment and slow detection speed in the detection of marine animals. Qiang et al. [37] proposed a target recognition algorithm based on improved SSD to improve the target detection accuracy and speed in the complex underwater environment. Zang et al. [38] used a region-based fully convolutional network (R-FCN) model to identify marine organisms to realize a submersible tool for video information recognition. Liu [39] solved the imbalance of underwater image datasets through a GAN network. Li et al. [40] employed a transfer learning strategy to train the underwater YOLO network and to alleviate the limitation of fish samples.
Generally, deep learning-based underwater target recognition has achieved some progress in recent years. Nevertheless, the powerful ability of deep learning has not been fully exploited. This paper introduces the latest version of YOLO, i.e., YOLOv4, to underwater target recognition. Moreover, in consideration of the particularity of underwater environments, YOLOv4 is modified to improve the speed and accuracy of target recognition. The original upsampling module is replaced with a deconvolution module. At the same time, the depthwise separable convolution is added to the YOLO network to reduce the calculation burden of network recognition and training. Another contribution of the study is that a new data preprocessing method is proposed to improve the accuracy of target recognition. In this method, the background environment factors are taken into account when extracting target features. Mosaic augmentation is employed and improved. It is noted from the literature that, in the process of underwater target recognition, enhancement or restoration algorithms for underwater images are often incorporated into target recognition to improve the accuracy of recognition. For example, Xie et al. [41] applied the dark channel prior model in underwater target recognition. Zhou et al. [42] adopted an adaptive underwater image enhancement algorithm to suppress noise and to improve edge clarity in recognizing moving underwater objects. Liang et al. [43] proposed an algorithm that combines the improved dark channel and MSR (multiscale Retinex) to improve the accuracy of target recognition. Yang et al. [44] proposed a novel underwater image enhancement method to improve the quality of underwater images through color compensation and correction, gamma correction, and brightness de-blurring. Note that, although some underwater image enhancement algorithms yield positive results, the processing speed is limited, which degrades their real-time performance. In this study, to improve the accuracy and speed of underwater target recognition, an improved YOLOv4 is combined with an improved mosaic augmentation in which an image enhancement algorithm, gray world algorithm, is incorporated.
The rest of the paper is organized as follows. In Section 2, the fundamentals of gray world algorithm, YOLO, mosaic augmentation, and the CIoU loss function are described. In Section 3, the improvement of YOLOv4 and modification of mosaic augmentation are explained. In Section 4, some examples are treated with the proposed method to verify its effectiveness. The final section contains the conclusions.

2. Fundamentals

2.1. Gray World Algorithm

As is known, underwater images are mostly affected by the refraction and absorption of light. Therefore, an image of objects in water presents a blue-green tone. At the same time, the scattering of light induced by particles in the water makes the image details blurred and the surface atomized. The temperature of water also affects the propagation of light, leading to light scattering [45]. In summary, most underwater images have problems of color deviation, blur, and atomization. As a result, it is difficult to extract the features of underwater images.
To alleviate the color deviation, the gray world algorithm is used in this study. This algorithm is based on the gray world hypothesis [46]. The hypothesis holds that the average reflection of light by natural objects is a fixed value in general, and this fixed value is approximately considered “gray”. The gray world algorithm is widely used owing to two advantages. Firstly, for a single-tone image—for example, a greenish underwater image—the enhancement effect of gray world algorithm is satisfactory. The second advantage of the gray world algorithm is the low computational cost [47]. The gray world algorithm can be described as follows.
First, the average values of the three channels are calculated:
R a v g = 1 m 0 m 1 R G a v g = 1 m 0 m 1 G B a v g = 1 m 0 m 1 B ,
where R, G, B, represent the red, green, and blue pixel values of each pixel; Ravg, Gavg, Bavg are the average of red, green, and blue channels of all pixels. m is the number of pixels in the image. Based on Equation (1), the mean value of RGB can be obtained as:
K = R a v g + G a v g + B a v g / 3 ,
Then, the gain of each channel relative to K can be determined:
K R = K / R a v g K G = K / G a v g K B = K / B a v g ,
The pixel values of pixels are adjusted one by one according to the obtained gains:
R n e w = K R R G n e w = K G G B n e w = K B B ,
where Rnew, Gnew, Bnew represent the new pixel value of each pixel.

2.2. YOLO Neural Network

The You only look once (YOLO) neural network was proposed by Redmon et al. in 2015 [18]. It is the first one-stage target detector in the era of neural networks. In this target detector, to improve the detection speed, the “proposal detection + verification” pattern in the two-stage detector (e.g., RCNN, Faster RCNN) is abandoned; only one neural network is used instead. The YOLO network divides the image into several regions with the same size. Then, the network predicts the kinds and probability of objects in bounding boxes in each region. Later, Redmon made a series of improvements on the basis of YOLO, by proposing v2 and v3 versions [48,49].
YOLOv3 was proposed by Redmon in 2018 [50]. Owing to the fast training speed and detection speed, it is often used in practice. YOLOv3 blends several excellent structures, including the Darknet53 network, anchor, and FPN network. The basic framework of the YOLOv3 neural network is shown in Figure 1.
YOLOv4 was proposed by Bochkovskiy [50] in 2020 on the basis of YOLOv3. Compared with YOLOv3, YOLOv4 makes many improvements, including changing the backbone network Darknet53 to CSPDarknet53, adding SPP and PANet network, taking CIoU as a part of the loss function, and using mosaic augmentation. The structure of the YOLOv4 neural network is shown in Figure 2. As can be seen, as with YOLOv3, in YOLOv4, the input image is processed as an image with the width and height of 416 × 416. For different sizes of targets, the final output layer of the YOLO network has three sizes of width and height, i.e., 13 × 13, 26 × 26, and 52 × 52. Note that in the output layers, as shown in Figure 1 and Figure 2, for example, 13 × 13 × (c + 5) × 3, c represents the number of kinds to be detected; 5 reflects whether there is a target in an anchor box and 4 adjustment parameters of the anchor box; 3 represents three different anchor boxes corresponding to each size of width and height. It can also be seen that the basic convolution blocks of the YOLOv4 network, i.e., CBL and CBM, are similar to YOLOv3. The main difference is that YOLOv4 takes the Mish function as the activation function in the backbone network, while the Leaky ReLU function is used in YOLOv3.
A basic convolution block in YOLOv4 is composed of normal convolution, batch normalization, and activation functions.
For the CBL block,
x′ = Leaky ReLU(Batch Normalization(Conv(x))).
For the CBM block,
x′ = Mish(Batch Normalization(Conv(x))),
where x is the input, x′ is the output, Conv() is the normal convolution operator, Batch Normalization() is the batch normalization operator, and Leaky ReLU() and Mish() are activation functions.
The Res unit is a basic residual component in both YOLOv3 and YOLOv4. The output feature of the unit is obtained by adding the input feature and the result of two convolutions. In the residual convolution block of YOLOv3, the input feature enters the Res unit after a basic convolution. Then, an output feature is obtained. In YOLOv4, the input feature travels along two routes after a basic convolution. In one route, the result of convolution goes through the Res unit and the basic convolution, while in the other route, the result of convolution only experiences the basic convolution. The final output feature is obtained by concatenating the results of two routes.
In addition to modifying the backbone network, YOLOv4 also adds the SPP network and modifies the FPN network to the PANet network. The SPP network performs maximum pooling of inputs with different sizes and concatenates the pooling results together to obtain an output. The PANet network can be regarded as an upgrade of the FPN network, in which the structure of downsampling is added.

2.3. Mosaic Augmentation

Mosaic augmentation [50] derives from CutMix augmentation [51], which stitches two pictures together into one picture to increase the number of targets and enhance the complexity of the background. It has been proven that such an operation is beneficial to the training of a target detection neural network.

2.4. CIoU

IoU (Intersection over Union) is a basic criterion in target recognition, being the ratio of the intersection and union of the predict box and ground truth. In the training of a neural network, sometimes, it is problematic to take IoU as a part of the loss function, especially for the case of IoU = 0 and the case when different intersections have the same IoU value. Therefore, in YOLOv4, CIoU (Complete IoU) [52] is selected to replace IoU as a part of the loss function. The lower the CIoU, the better a predicted box approximates the ground truth. The CIoU part of the loss function and its components are as follows:
L CIoU = 1 IoU + ρ 2 ( b , b gt ) c 2 + α ν ,
IoU = B B gt B B gt ,
ν = 4 π 2 arctan w gt h gt arctan w h 2 ,
α = ν 1 IoU + ν ,
where B, Bgt represent the predict box and the ground truth, respectively; ρ ( b , b gt ) is the distance between the center point of the predict box and the ground truth; c represents the diagonal distance of the smallest rectangle containing the predict box and the ground truth; w, h, wgt, hgt represent the width and height of the predict box and the ground truth, respectively; α is a positive trade-off parameter to measure the consistency of aspect ratio ν. The parameters of CIoU are shown in Figure 3.

3. Improvement of YOLOv4

3.1. E-Mosaic Augmentation

As is known, there are two main purposes of mosaic augmentation in target recognition. One is to increase the complexity of the image background. The other is to increase the number of objects in an image. Due to the particularity of the underwater environment, the image background is mostly blue-green for underwater images. The complexity of the background is too weak for mosaic augmentation. As a result, the accuracy of target recognition cannot be improved obviously, even if mosaic augmentation is carried out.
In this study, a novel data augmentation method, named after e-mosaic augmentation, is proposed. An image enhancement algorithm, gray world algorithm, is incorporated into mosaic augmentation. Similarly to normal mosaic augmentation, e-mosaic augmentation selects four images. The difference is that, before stitch, the e-mosaic augmentation first enhances two images on the diagonal by using gray world algorithm. The purpose is to increase the complexity of the background, which benefits the training of a target recognition neural network. Figure 4 presents the process of e-mosaic augmentation.

3.2. YOLOv4-uw

In this study, the structure of YOLOv4 is modified to increase the target recognition accuracy and speed. To improve the accuracy of recognition, the upsampling module in YOLOv4 is replaced with the deconvolution module. To increase the speed of recognition, depthwise separable convolution is used to replace the normal convolution in the Res unit and the normal convolution near the output layer.
Figure 5 shows the difference between upsampling and deconvolution. As can be seen, the upsampling structure can only enlarge the feature size and cannot restore the details. By contrast, the deconvolution structure can not only enlarge the feature size but also restore details. This is because deconvolution is the reverse operation of convolution, which can approximately restore the feature layer to what it was before convolution. Obviously, compared to upsampling, deconvolution is conducive to the restoration of image details.
Figure 6 presents the comparison between a normal three-dimensional convolution and a depthwise separable three-dimensional convolution. The left plot is the normal 3 × 3 convolution block, which is composed of 3 × 3 convolution, batch normalization (BN), and the activation function (AF). The right plot is a 3 × 3 depthwise separable convolution block, which is composed of depthwise convolution, point convolution (1 × 1 convolution), batch normalization, and the activation function. Depthwise separable convolution is the core of MobileNet [53]. The difference between depthwise separable convolution and normal three-dimensional convolution is that depthwise separable convolution divides convolution operation into two steps to reduce the amount of convolution calculation. Although the accuracy might be slightly reduced, the speed of training and detection can be usually improved.
Figure 7 presents the structure of the modified YOLOv4 in the study, named YOLOv4-uw since this version of YOLOv4 is constructed for the purpose of underwater target recognition. Notably, it is observed that in the used training set and test set, the objects to be recognized are of medium size and small size. To further improve the speed of target recognition, the large-sized output layer in YOLOv4, i.e., with the size of 52 × 52, is removed. Correspondingly, the SPP structure is deleted since SPP is introduced in YOLOv4 to deal with various sizes of objects.

4. Results and Discussion

4.1. Training and Test Set

The training environment of YOLO is: Windows10; Inter® Core™ i5-10600KF; NVIDIA GeForce RTX 3080; Python3.7; CUDA11.0; Cudnn8.0.4; tensorflow-gpu2.4.0.
The experimental dataset is from the Target Recognition Group of 2020 China Underwater Robot Professional Competition (URPC 2020). In total, it contains 4757 pictures, 3805 of which (80% data) are taken as the training set, while the rest, 952 pictures (20% data), are taken as the test set. The target types are divided into four categories: echinus, starfish, holothurian, and scallop.

4.2. Verification of E-Mosaic Augmentation

The proposed e-mosaic augmentation method is verified from objective and subjective aspects. The recognition precision is calculated as the objective evaluation index. The visual detection results are presented to subjectively evaluate the effect of e-mosaic augmentation.
Table 1 shows the precision of recognition results in terms of AP (average precision) and mAP (mean average precision), in which the threshold value of IoU is set to 0.5. The AP value is a comprehensive evaluation metric based on the precision and recall in a category. The precision and recall in a category can be calculated after setting the threshold of IoU. The mAP value is the mean AP of all categories. To verify the effectiveness of e-mosaic augmentation, three different preprocessing methods with respect to the original dataset are considered, including without data preprocessing, data preprocessing by mosaic augmentation, and data preprocessing by proposed e-mosaic augmentation. After the training set is determined, training is performed by using YOLOv4. As shown in Table 1, e-mosaic augmentation can effectively improve the detection accuracy. Compared with the YOLOv4 without any data preprocessing, in the four underwater animal categories, the recognition precision of starfish is improved the most by mosaic and e-mosaic augmentation. Compared with mosaic augmentation, the proposed e-mosaic augmentation leads to a general improvement in terms of both AP and mAP.
Figure 8 presents some recognition results of four underwater animals, in which recognition of scallop is labeled by a blue box; echinus by purple; starfish by yellow; holothurian by red. The plots in the first column of Figure 8a are obtained in the case of the training set without preprocessing. Presented in column Figure 8b are the results after using mosaic augmentation in training. Figure 8c shows the recognition results after using e-mosaic augmentation in training. It can be seen that the recognition results obtained through e-mosaic augmentation-assisted training are the best. Not only are the recognized targets the most, but also the predict box is the most accurate.

4.3. Verification of YOLOv4-uw

As aforementioned, the large-scale output layer is deleted from the proposed YOLOv4-uw in the study because of the lack of large-scale objects in the available training set. The SPP structure is also removed. Table 2 shows the comparison of YOLOv4-uw with and without the SPP structure. As can be seen from the comparison results, SPP does not improve the detection accuracy. It can be even found that the detection speed in terms of FPS (frame per second) is improved after the SPP structure is deleted.
The proposed YOLOv4-uw neural network is further verified via network performance and hardware requirement. The network performance is reflected by detection accuracy and speed while the hardware requirement refers to model size, parameters, and required calculation ability.
Table 3 shows the comparison between the proposed YOLOv4-uw and other classic networks. The mAP value, time spent in detection, and detection speed are used as evaluation criteria. Notably, when mAP is measured, the IoU threshold is set to 0.5. As can be seen from the comparison results, the target recognition by the YOLOv4-uw network yields a mAP index value of 75.34%, which is nearly 12% higher than YOLOv4. Moreover, YOLOv4-uw outperforms other target recognition networks. In terms of detection time, the proposed YOLOv4-uw takes 2.3 ms, which is 0.5 ms faster than YOLOv4. In terms of FPS, the proposed YOLOv4-uw reaches 44 FPS, which is 9 FPS more than YOLOv4.
Table 4 shows the comparison of the required network model size, parameters, and required calculation ability (represented by billion floating point operations per second, Bflop/s) among the proposed YOLOv4-uw and other one-step detectors. The smaller the network is, the lower the requirement of storage space is. As shown in Table 4, the size of YOLOv4-uw is 65 MB, around a quarter of that of YOLOv4. Moreover, the values of the parameters in the YOLOv4-uw are the lowest. BFLOP/s is usually used as a measure of the computing power required by a network. The larger the BFLOP/s is, the higher the requirement of the device is. As can be seen from Table 4, the BFLOP/s of YOLOv4-uw is the smallest by 19.6 Bflop /s, around one third of that of YOLOv4.

4.4. Verification of the Combination of E-Moasic and YOLOv4-uw

Table 5 shows the mAP of recognition results by YOLOv4, YOLOv4 with e-mosaic, YOLOv4-uw, and YOLOv4-uw with e-mosaic. Three cases with different IoU threshold values are considered. [email protected] indicates that the IoU threshold is set to 0.5. [email protected] means that the IoU threshold is set to 0.75. IoU@[0.5:0.95] means that when the IoU changes from 0.5 to 0.95, the mAP is tested every 0.05. Then, the average value of 10 measurement results is taken as the final mAP. As can be seen from the results in Table 5, when the IoU threshold is set to 0.5, the recognition precision by YOLOv4-uw associated with e-mosaic augmentation can reach 76.84%, which is nearly 13% higher than YOLOv4 without e-mosaic augmentation and around 8% higher than YOLOv4 with e-mosaic augmentation. When the IoU threshold is set to [0.5:0.95], it can be seen that under the same data augmentation conditions, the mAP of YOLOv4-uw is around 11% higher than that of YOLOv4. The mAP of the same network (YOLOv4 or YOLOv4-uw) can be increased by around 1.2% by using e-mosaic augmentation.

5. Conclusions

In this paper, an improved YOLOv4 is proposed to improve the accuracy and speed of underwater target recognition. A modified data augmentation is also proposed to improve the accuracy of recognition. Through the comparison results between the proposed YOLOv4-uw and other networks, including YOLOv4, it can be seen that the strategy proposed in this study can effectively improve the recognition accuracy and speed. Moreover, the requirement of hardware performance using the proposed YOLOv4-uw is lower than the others.
It should be noted that only one image enhancement method, i.e., gray world algorithm, is considered in performing e-mosaic augmentation. In the next work, other image enhancement methods will be considered and applied to the preprocessing of the training set. In addition, further optimization of YOLOv4-uw will be conducted to improve the recognition accuracy and speed. Due to the fact that the dataset in this paper consists of medium and small targets, the effectiveness of the neural network proposed here for large targets has not been confirmed. Future work will be concerned with the verification of the proposed YOLO-uw network when dealing with the detection of large-scale underwater objects.

Author Contributions

Conceptualization, L.C. and W.L.; methodology, L.C.; software, L.C., M.Z. and S.D.; validation, W.L. and L.Y.; formal analysis, L.C. and S.D.; investigation, L.C. and M.Z.; resources, W.L. and L.Y.; data curation, L.C., M.Z. and S.D.; writing—original draft preparation, L.C.; writing—review and editing, W.L.; visualization, L.C., M.Z. and S.D.; supervision, W.L. and L.Y.; project administration, W.L. and L.Y.; funding acquisition, W.L. and L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China Fujian Provincial Department of Ocean and Fisheries, Grant MHGX-16; Fujian Provincial Industrial Robot Basic Components Technology Research and Development Center, Grant number 2014H21010011.

Acknowledgments

The authors would like to thank the anonymous reviewers for their constructive suggestions, which comprehensively improved the quality of the paper. The authors also would like to thank the support from Fuzhou Ocean Research Institute and the support by Fujian Provincial College Marine Engineering Equipment Design and Manufacturing Engineering Research Center.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Boudhane, M.; Nsiri, B. Underwater image processing method for fish localization and detection in submarine environment. J. Vis. Commun. Image Represent. 2016, 39, 226–238. [Google Scholar] [CrossRef]
  2. Ahn, J.; Yasukawa, S.; Sonoda, T.; Ura, T.; Ishii, K. Enhancement of deep-sea floor images obtained by an underwater vehicle and its evaluation by crab recognition. J. Mar. Sci. Technol. 2017, 22, 758–770. [Google Scholar] [CrossRef]
  3. Singh, H.; Adams, J.; Mindell, D.; Foley, B. Imaging underwater for archaeology. J. Field Archaeol. 2000, 27, 319–328. [Google Scholar]
  4. Watanabe, J.-I.; Shao, Y.; Miura, N. Underwater and airborne monitoring of marine ecosystems and debris. J. Appl. Remote Sens. 2019, 13, 044509. [Google Scholar] [CrossRef]
  5. Gu, L.; Song, Q.; Yin, H.; Jia, J. An overview of the underwater search and salvage process based on ROV. Sci. Sin. Inf. 2018, 48, 1137–1151. [Google Scholar] [CrossRef]
  6. Hong, S.; Fang-Jian, S.; Bo, C.; Wei, Q. An underwater ship fault detection method based on sonar image processing. J. Phys. Conf. Ser. 2016, 679, 012036. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, R.S.; Yuan, X.H.; Hu, Z.; Qiu, C.; Liu, P.; Wu, W. Vision System Based on Identifying 3D Underwater Objects. Shipbuild. China. 1997, 4, 88–94. [Google Scholar]
  8. Zhang, J. Simultaneous Localization and Mapping for an AUV Using Forward-Looking Sonar. Master’s Thesis, Ocean University of China, Qingdao, China, 2008. [Google Scholar]
  9. Liu, B. Research on Feature Extraction and Target Identification in Machine Vision Underwater and Surface Image. Ph.D. Thesis, Dalian University of Technology, Dalian, China, 2013. [Google Scholar]
  10. Viola, P.; Jones, M. Rapid Object Detection using a Boosted Cascade of Simple Features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001; Volume 1, p. 1. [Google Scholar]
  11. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar]
  12. Felzenszwalb, P.; McAllester, D.; Ramanan, D. A discriminatively trained, multiscale, deformable part model. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  13. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D. Cascade object detection with deformable part models. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2241–2248. [Google Scholar]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc. (NIPS): Lake Tahoe, NV, USA, 2012; pp. 1097–1105. [Google Scholar]
  15. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  17. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single shot multibox detector. In European Conference on Computer Vision; Springer: Berlin, Germany, 2016; pp. 21–37. [Google Scholar]
  19. Xu, H.M.; Zhang, Z.G.; Zheng, B. Distribution Characteristics in the Underwater Inhomogeneous Illumination Field. Acta Photonic Sin. 2010, 39, 1606–1610. [Google Scholar]
  20. Xu, X.; Shi, M.; Zhang, J.L. Feature extraction of underwater targets using generalized S-transform. J. Comput. Appl. 2012, 32, 280–282. [Google Scholar]
  21. Ma, W.Y. Underwater Target Detection Based on The Polarization Features. Electron. Meas. Technol. 2014, 37, 90–92, 98. [Google Scholar]
  22. Wang, H.B.; Zhang, Q.; Wang, X.; Chen, Z. Object Detection Based on Regional Saliency and Underwater Optical Prior Knowledge. Chin. J. Sci. Instrum. 2014, 2, 387–397. [Google Scholar]
  23. Oliver, K.; Hou, W.; Wang, S. Image feature detection and matching in underwater conditions. In Ocean Sensing and Monitoring II; International Society for Optics and Photonics: Bellingham, WA, USA, 2010; Volume 7678. [Google Scholar]
  24. Zhang, H.; Wan, L.; Sun, Y. A new approach to underwater target recognition. In Proceedings of the 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009. [Google Scholar]
  25. Yang, J.; Xu, F.; Wei, Z.H.; Liu, J.; An, X.D. Multi-Feature Extraction and Fusion for the Underwater Moving Targets Classification. In Applied Mechanics and Materials; Trans Tech Publications Ltd: Stafa-Zurich, Switzerland, 2013; Volume 303. [Google Scholar]
  26. Dubreuil, M.; Delrot, P.; Leonard, I.; Alfalou, A.; Brosseau, C.; Dogariu, A. Exploring underwater target detection by imaging polarimetry and correlation techniques. Appl. Opt. 2013, 52, 997–1005. [Google Scholar] [CrossRef]
  27. Yahya, M.F.; Arshad, M.R. Robust recognition of targets for underwater docking of autonomous underwater vehicle. In 2016 IEEE/OES Autonomous Underwater Vehicles (AUV); IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
  28. Liu, L.X.; Jiao, S.H.; Chen, T. Detection and Recognition of Underwater Target Based on Feature Matching. Mod. Electron. Tech. 2011, 34, 73–76, 80. [Google Scholar]
  29. Hough, P.V.C. Method and Means for Recognizing. U.S. Patent 3,069,654, 25 March 1960. [Google Scholar]
  30. Li, K.; Liu, Z.; Mao, D. Algorithm for Recognition of Underwater Small Target Based on Shape Characteristic. Ship Sci. Technol. 2012, 34, 91–94. [Google Scholar]
  31. Li, X.; Shang, M.; Qin, H.; Chen, L. Fast accurate fish detection and recognition of underwater images with fast r-cnn. In Proceedings of the OCEANS 2015-MTS/IEEE Washington, Washington, DC, USA, 19–22 October 2015. [Google Scholar]
  32. Sun, X.; Shi, J.; Liu, L.; Dong, J.; Plant, C.; Wang, X.; Zhou, H. Transferring deep knowledge for object recognition in Low-quality underwater videos. Neurocomputing 2018, 275, 897–908. [Google Scholar] [CrossRef] [Green Version]
  33. Zhou, H.; Huang, H.; Yang, X.; Zhang, L.; Qi, L. Faster R-CNN for marine organism detection and recognition using data augmentation. In Proceedings of the International Conference on Video and Image Processing, Singapore, 27–29 December 2017. [Google Scholar]
  34. Park, J.-H.; Kang, C. A Study on Enhancement of Fish Recognition Using Cumulative Mean of YOLO Network in Underwater Video Images. J. Mar. Sci. Eng. 2020, 8, 952. [Google Scholar] [CrossRef]
  35. Arain, B.; McCool, C.; Rigby, P.; Cagara, D.; Dunbabin, M. Improving underwater obstacle detection using semantic image segmentation. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
  36. Jia, Z.Q.; Liu, X.F. Marine Animal Target Detection Based on YOLO and Image Enhancement. Electron. Meas. Technol. 2020, 43, 84–88. [Google Scholar]
  37. Qiang, W.; He, Y.Y.; Gou, Y.J.; Li, B.; He, L. Exploring Underwater Target Detection Algorithm Based on Improved SSD. J. Northwestern Polytech. Univ. 2020, 38, 747–754. [Google Scholar] [CrossRef]
  38. Zang, X.H.; Xu, Z.F.; Yu, Z.C.; Gao, M.K.; Zang, R.G. Deep Sea Biometric Recognition Technology Based on R-FCN. Electron. Meas. Technol. 2020, 43, 158–161. [Google Scholar]
  39. Liu, Y.Y.; Zhang, J.M.; Wang, K.P.; Feng, X.H.; Yang, X.H. Fast Underwater Target Recognition with Unbalanced Data Set. Comput. Eng. Appl. 2020, 56, 236–242. [Google Scholar]
  40. Li, Q.Z.; Li, Y.B.; Niu, J. Real-Time Detection of Underwater Fish Based on Improved YOLO and Transfer Learning. J. Pattern Recognit. Artif. Intell. 2019, 32, 193–203. [Google Scholar] [CrossRef]
  41. Kun, X.; Pan, W.; Xu, S. An underwater image enhancement algorithm for environment recognition and robot navigation. Robotics 2018, 7, 14. [Google Scholar] [CrossRef] [Green Version]
  42. Zhou, Y.; Li, Q.; Huo, G. Underwater moving target detection based on image enhancement. In International Symposium on Neural Networks; Springer: Cham, Switzerland, 2017. [Google Scholar]
  43. Liang, T.Q.; Zhang, X.Y.; Duan, P.; Yu, H.; Zhang, B.; Tang, Q. Underwater Target Detection Under Strong Scattering Medium Using Improved Dark Channel Method. Infrared Laser Eng. 2020, 49, 104–109. [Google Scholar]
  44. Yang, M.; Wang, H.W.; Hu, K.; Yin, G.; Hu, J. An Underwater Image Comprehensive Enhancement Algorithm Based on Color Compensation. J. Graph. 2021, 42, 59–64. [Google Scholar]
  45. Zhang, X.; Hu, L. Effects of temperature and salinity on light scattering by water. In Ocean Sensing and Monitoring II; International Society for Optics and Photonics: Bellingham, WA, USA, 2010; Volume 7678. [Google Scholar]
  46. Buchsbaum, G. A spatial processor model for object colour perception. J. Frankl. Inst. 1980, 310, 1–26. [Google Scholar] [CrossRef]
  47. Celik, T.; Yetgin, Z. Grey-wavelet: Unifying grey-world and grey-edge colour constancy algorithms. Signal Image Video Process. 2015, 9, 1889–1896. [Google Scholar] [CrossRef]
  48. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  49. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, preprint. arXiv:1804.02767. [Google Scholar]
  50. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, preprint. arXiv:2004.10934. [Google Scholar]
  51. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 3 November–27 October 2019. [Google Scholar]
  52. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34. No. 07. [Google Scholar]
  53. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, preprint. arXiv:1704.04861. [Google Scholar]
Figure 1. Framework of YOLOv3.
Figure 1. Framework of YOLOv3.
Electronics 10 01634 g001
Figure 2. Structure of YOLOv4.
Figure 2. Structure of YOLOv4.
Electronics 10 01634 g002
Figure 3. Description of CIoU.
Figure 3. Description of CIoU.
Electronics 10 01634 g003
Figure 4. E-mosaic augmentation.
Figure 4. E-mosaic augmentation.
Electronics 10 01634 g004
Figure 5. Upsampling and deconvolution: (a) upsampling; (b) deconvolution.
Figure 5. Upsampling and deconvolution: (a) upsampling; (b) deconvolution.
Electronics 10 01634 g005
Figure 6. A 3 × 3 convolution and corresponding depthwise separable convolution.
Figure 6. A 3 × 3 convolution and corresponding depthwise separable convolution.
Electronics 10 01634 g006
Figure 7. Structure of YOLO-uw.
Figure 7. Structure of YOLO-uw.
Electronics 10 01634 g007
Figure 8. Target recognition results: (a) without preprocessing training set; (b) preprocessing training set by mosaic augmentation; (c) preprocessing training set by e-mosaic augmentation.
Figure 8. Target recognition results: (a) without preprocessing training set; (b) preprocessing training set by mosaic augmentation; (c) preprocessing training set by e-mosaic augmentation.
Electronics 10 01634 g008
Table 1. Comparison of precision of recognition results with different preprocessing of training set.
Table 1. Comparison of precision of recognition results with different preprocessing of training set.
Preprocessing of Training SetAPmAP (%)
ScallopEchinusStarfishHolothurian
Without preprocessing0.660.850.510.5563.96
mosaic0.580.830.740.5467.20
e-mosaic0.610.840.720.5768.46
Table 2. Comparison of YOLOv4-uw with and without SPP.
Table 2. Comparison of YOLOv4-uw with and without SPP.
Preprocessing of Training SetAPmAP (%)Detection Speed (FPS)
ScallopEchinusStarfishHolothurian
YOLOv4-uw with SPP0.640.860.820.6273.4839
YOLOv4-uw without SPP0.640.860.830.6875.3444
Table 3. Comparison between YOLOv4-uw and other networks.
Table 3. Comparison between YOLOv4-uw and other networks.
NetworkmAP (%)Time Spent in Detection (ms)Detection Speed (FPS)
Faster RCNN41.985.717
SSD71.450.9105
CenterNet73.571.855
YOLOv331.522.148
YOLOv463.962.835
YOLOv4-uw75.342.344
Table 4. Comparison between YOLOv4-uw and other one-step detectors.
Table 4. Comparison between YOLOv4-uw and other one-step detectors.
NetworkModel Size (MB)Total Parameters (M)Bflop/s
SSD9224.263.2
CenterNet12532.750.5
YOLOv323561.665.4
YOLOv425064.059.7
YOLOv4-uw6516.719.6
Table 5. Comparison of YOLOv4 and YOLOv4-uw, with and without e-mosaic.
Table 5. Comparison of YOLOv4 and YOLOv4-uw, with and without e-mosaic.
ConditionmAP
YOLOv4YOLOv4-uw
e-mosaicNoYesNoYes
IoU @0.563.9668.4675.3476.84
IoU @0.7512.7712.2228.4130.08
IoU @[0.5:0.95]25.0326.3636.1237.35
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, L.; Zheng, M.; Duan, S.; Luo, W.; Yao, L. Underwater Target Recognition Based on Improved YOLOv4 Neural Network. Electronics 2021, 10, 1634. https://doi.org/10.3390/electronics10141634

AMA Style

Chen L, Zheng M, Duan S, Luo W, Yao L. Underwater Target Recognition Based on Improved YOLOv4 Neural Network. Electronics. 2021; 10(14):1634. https://doi.org/10.3390/electronics10141634

Chicago/Turabian Style

Chen, Lingyu, Meicheng Zheng, Shunqiang Duan, Weilin Luo, and Ligang Yao. 2021. "Underwater Target Recognition Based on Improved YOLOv4 Neural Network" Electronics 10, no. 14: 1634. https://doi.org/10.3390/electronics10141634

APA Style

Chen, L., Zheng, M., Duan, S., Luo, W., & Yao, L. (2021). Underwater Target Recognition Based on Improved YOLOv4 Neural Network. Electronics, 10(14), 1634. https://doi.org/10.3390/electronics10141634

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop