Next Article in Journal
Evaluation of Different Contact Assumptions in the Analysis of Friction-Induced Vibrations Using Dynamic Substructuring
Previous Article in Journal
Comprehensive and Simplified Fault Diagnosis for Three-Phase Induction Motor Using Parity Equation Approach in Stator Current Reference Frame
Previous Article in Special Issue
COVID-19 Pandemic Response Robot
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Object Tracking for an Autonomous Unmanned Surface Vehicle

1
Graduate Institute of Automation and Control, National Taiwan University of Science and Technology, Taipei 106335, Taiwan
2
Center for Cyber-Physical System Innovation, National Taiwan University of Science and Technology, Taipei 106335, Taiwan
*
Author to whom correspondence should be addressed.
Machines 2022, 10(5), 378; https://doi.org/10.3390/machines10050378
Submission received: 29 March 2022 / Revised: 10 May 2022 / Accepted: 11 May 2022 / Published: 16 May 2022
(This article belongs to the Special Issue Design and Control of Electrical Machines)

Abstract

:
The conventional algorithm used for target recognition and tracking suffers from the uncertainties of the environment, robot/sensors and object, such as variations in illumination and viewpoint, occlusion and seasonal change, etc. This paper proposes a deep-learning based surveillance and reconnaissance system for unmanned surface vehicles by adopting the Siamese network as the main neural network architecture to achieve target tracking. It aims to detect and track suspicious targets. The proposed system perceives the surrounding environment and avoids obstacles while tracking. The proposed system is evaluated with accuracy, precision, recall, P-R curve, and F1 score. The empirical results showed a robust target tracking for the unmanned surface vehicles. The proposed approach contributes to the intelligent management and control required by today’s ships, and also provides a new tracking network architecture for the unmanned surface vehicles.

1. Introduction

In recent years, due to experimental and communication difficulties, unmanned surface vehicles (USVs) have fallen far behind unmanned aerial vehicles (UAVs) and unmanned ground vehicles in the field of artificial intelligence (AI) research and development. However, USV not only has good future development, but also has a wide range of applications. In the military, it can be used for waypoint patrols [1,2], gathering intelligence, surveillance, and reconnaissance [3,4]. For civilian purposes, it can be used to assist in finding people who fall into the water [5], testing water quality [6,7], and so on. In the treacherous and ever-changing marine environment, collision avoidance and target tracking are the prerequisites for USV to perform tasks, and therefore become the key development direction in the USV research field [8]. The problems encountered in target tracking can be roughly divided into four categories: the shape change of the target, the scale change of the target, the occlusion and disappearance of the target, and the blurred image [9]. Target tracking methods are mainly divided into two categories, one is a filter algorithm, and the other is a deep learning algorithm. Using the Particle Filter method, based on particle distribution statistics [10]. First, the tracking target is modeled, and a similarity metric is defined to determine the degree of matching between the particles and the target. In the process of target search, it will sprinkle some particles according to a certain distribution (such as uniform distribution or Gaussian distribution), count the similarity of these particles, and determine the possible position of the target. Although this method is fast, it easily reduces the accuracy and stability of tracking. Using deep learning methods, and using deep learning to train the network model, the obtained convolution feature output performance is applied to the correlation filtering tracking framework, to obtain better tracking results [11]. This method obtains better eigenvalues, and improves the accuracy and stability of tracking, but at the same, time it also brings an increase in the amount of calculation.
Due to the floating and undercurrents of the water, many USV control systems have been proposed. Among the non-AI methods, [12] proposed the use of proportional-integral-derivative controller to control the motion of the USV. By establishing the three-degree-of-freedom kinematics and dynamics model of the USV, the heading tracking controller was designed based on the output feedback control method. In addition, [13] also proposed a sliding mode control method based on Kalman filter for the heading control of the water jet propelled USV in the horizontal plane. With Nomoto model, a heading controller based on sliding mode control is designed and the Sigmoid function is introduced to improve the traditional exponential approximation rate.
In the AI method, ref. [14] proposed an alternative navigation system when there is no global positioning system (GPS). USV uses the simultaneous localization and mapping framework to perform relative navigation with respect to the surrounding coastline and uses B-spline to parameterize coastline features for effective map management. A planning algorithm based on deep reinforcement learning was proposed to find the shortest collision avoidance path for USVs [15]. Solutions in target tracking against various certainties (such as light changes, different scenes, or occlusion of targets) have been proposed according to the surveyed literature.
A TensorFlow framework for moving object detection was proposed [16]. The proposed method is based on convolutional neural networks (CNN) target tracking algorithm for robust target detection. The dynamic frame rate optimization and selection of adaptive parameters according to the scene and content of the input video were proposed [17]. A general algorithm was proposed to estimate the perspective image area occluded by the object [18]. By connecting the real environment and the perspective space, the two coordinate spaces create a flexible object tracking environment.
When these algorithms are applied to vehicles, they face more challenges. An object tracking algorithm for UAVs using robust multi-collaborative tracker is proposed [19], which can provide additional object information and modify the short-term tracking model in time. In short-range maritime surveillance, X-band maritime radar is used to capture objects in an extended area with different intensities [20]. Combining the position, shape, and appearance of the target, multiple kernel correlation filters are proposed to track a single target in a real marine radar.
According to the above-surveyed literatures, the tracking methods developed today have insurmountable problems, such as the target of Kernelized Correlation Filter cannot be recovered after being completely occluded [21], there are many false positives in the target tracking of Tracking-Learning-Detection Tracker [22], and the Median Flow Tracker will fail in the case of moving drastically [23]. However, most of the current deep learning research has only a single neural network as the framework of artificial intelligence, and the use of a single neural network can easily cause target tracking and identification failure. This paper uses two parallel neural networks to share feature parameters to discuss improving the accuracy and recall of target tracking. In addition, in this paper, the use of self-made USVs combined with software systems were used to improve the current target tracking systems of USVs, which are mostly the shortcomings of traditional methods. This research predicts that the proposed Siamese architecture can improve the common target tracking problems in the previously proposed methods. Among them, the stability and accuracy of the system will be verified through changes in light, changes in viewpoint, and changes in obstructions. In terms of application to USV, this paper will use two different waters for testing: a large and bright swimming pool, and a narrow pond with a complex environment. Since the USV produced in this study is a catamaran, it is expected to overcome the environmental gap.
To meet the above requirements, the system will have three modes. The first mode is a 360-degree fixed-way cruise. The 360-degree platform of the camera mounted on this USV will make a circle in 60 s and take a photo every five seconds. Therefore, it will be used to stitch the panorama to ensure the entire domain is being monitored by the system. The second mode is to search for a suspicious target, grab the center of mass through the feature frame extracted by the algorithm, and calculate the gap to control the motor to feedback and follow the target. The third mode is to maintain a safe distance from the target and follow at a constant speed. After a while, the buzzer will give a warning and emit a laser light to warn the object. This research can be said to create a new hardware design and algorithm development for the USV field.

2. Materials and Methods

Deep learning is applied to improve the conventional object recognition and tracking for an autonomous USV. The hierarchical control architecture is shown in Figure 1. It can be divided into high-level control and low-level control. The environmental variables will be transmitted to the robot through wire or wireless via their respective sensors in the remote section. However, the wireless signal on the water is not stable enough for the AI model to deploy remotely to control the USV via wireless communication. The centralized architecture is adopted in this paper (wire communication). The remote AI is placed onboard the USV as edge computing to avoid abnormal communication between local and remote.
The high-level control of behavior is shown in Figure 2. The information obtained by the sensor can be planned by reasoning first, or it can be divided into four behaviors: goal-seeking, obstacle avoidance, trajectory tracking, and formation keeping. Perform low-level robot behavior control. The task of formation keeping is to maintain a constant pose between the USV and target for object recognition and tracking.

2.1. Robot System

According to previous studies, there is a huge difference between monohull and catamaran used for USV [24,25,26]. Due to the design of the wave-piercing body, the catamaran has the advantages of being relatively stable, lighter overall, high capacity, and not easy to capsize. The hulls on both sides are also relatively slender, and the resistance of the water flow is reduced during sailing, so the speed is increased, and the accuracy of target tracking can be greatly improved. This paper refers to the control method proposed by Qiang Zhu [27]. The hull adopts a catamaran design and uses twin propellers as the power source. Since the USV is manufactured by the author, the rest of the parts will use the Fusion 360 integrated 3-dimensions (3D) design. The tail of the fuselage is equipped with two sets of parallel power propellers, which can achieve movement by adjusting the running speed and direction of the twin propellers. The USV hull design is shown in Figure 3. The power components of this USV are placed at the bottom of the aft end of the hull, so the motor will be submerged underwater.

2.2. Algorithm

The target tracking and response system based on deep learning is applied to USV. In order to avoid image loss, cruise and image stitching are used to improve the accuracy of target tracking, and a new Siamese network method is adopted. After the system starts, it will automatically navigate to search and detect the most panoramic view. The overall system diagram is shown in Figure 4.

2.2.1. Feature-Based Panoramic Image Stitching

This system uses feature point detection as the benchmark for splicing. Feature point detection refers to the method of finding the feature points in the image based on the brightness, color, gradient, and other information of the image. In image alignment, feature point detection can be used to obtain feature points of two images, and then the alignment can be completed by matching these feature points. Common feature point detection methods such as: Harris Corner Detection, and Scale Invariant Feature Transformation (SIFT) [28,29,30]. As well, this system adopts SIFT method to carry on the detection, the flow chart is shown in Figure 5.
In the scale space extreme value detection, the Gaussian convolution kernel is applied due to its scale invariance. It detects the key points in the SIFT algorithm. The images are convolved using Gaussian filters at different scales, and then continuous Gaussian blurring of the image differences is used to find the key points. The key point is based on the maximum and minimum values of difference of Gaussians (DoG) at different scales:
L ( x , y , σ ) = G ( x , y , σ ) I ( x , y ) ,
G ( x , y , σ ) = 1 2 π σ 2 e ( x 2 + y 2 ) / 2 σ 2 ,
D ( x , y , σ ) = L ( x , y , k i σ ) L ( x , y , k j σ ) ,
where L(x, y, σ) is the image of the original image I(x, y) convoluted with Gaussian mask G(x, y, σ), L(x, y, σ) is the DoG image. The maximum and minimum values in the DoG image are defined as key points as
m ( x , y ) = ( L ( x + 1 , y ) L ( x 1 , y ) ) 2 + ( L ( x , y + 1 ) L ( x , y 1 ) ) 2 ,
θ ( x , y ) = tan 1 L ( x , y 1 ) L ( x , y + 1 ) L ( x 1 , y ) L ( x + 1 , y ) ,
where the magnitude and orientation of the key point are m (x, y) and θ (x, y) respectively. Each adjacent pixel is added to the histogram of the key point according to its magnitude and direction, and the direction of the maximum value in the final histogram is the direction of the key point. Each extracted point will have three pieces of information: scale, coordinates, and direction. To improve the stability of the registration of selected points, each point is represented by 4 × 4, a total of 16 seed points, each point contains 128 data, and the SIFT feature vector represented by the result is 128 dimensions. In this way, an image feature descriptor is generated for image feature matching, and the SIFT feature vector is no longer interfered by changes in direction and angle.
With these feature vectors, it is necessary to perform subsequent key point matching on the key points. The identification of the target is completed by the comparison of key point descriptors in the two-point set. The similarity index of the key point descriptor with 128 dimensions is used as follows:
R i = ( r i 1 , r i 2 , , r i 128 ) ,
S i = ( s i 1 , s i 2 , , s i 128 ) ,
d ( R i , S i ) = j 128 ( r i j s i j ) 2 ,
where Ri is the key point descriptor in the reference image, Si is the key point descriptor in the observation image, and d (Ri, Si) is the two similarity measures of the arbitrary reference image and observation image. The key point matching can be conducted by exhaustive method, but it will take too much time, so the data structure of the K-dimensional tree (K-d tree) is used to complete the search instead [31,32]. K-d tree is a binary tree in which each leaf node is a k-dimensional point. All non-leaf nodes can divide the space into two half spaces as a hyperplane. The content of the search is based on the key points of the target image, and the original image feature points that are closest to the feature points of the target image and the second adjacent original image feature points are searched.

2.2.2. Siamese-Based Target Tracking

Target tracking is to analyze the image sequence to calculate the position of the moving target in each frame of the image. Then, correlation matching is performed according to the characteristic values related to the moving target to obtain the complete trajectory of the target. The system uses a single target tracking method. The single target tracking (STT) method can predict the size and position of the target in subsequent frames. The basic task flow is shown in Figure 6 [33,34].
According to the above process, the first frame is selected in the movie, and then a lot of to-be-selected frames are generated in the next frame, and the feature values of these frames are extracted and scored. To adapt to changes in the appearance of the target and prevent drift in the tracking process, it is necessary to update the model approximately every frame. However, the past performance of the target is still important to the tracker. If it is continuously updated, it may lose the appearance information of the past performance and introduce too much noise. A combination of long and short-term updates can solve this problem. Therefore, the three steps of forgetting, updating, and output are added to the neural unit. The method of selecting the prediction results is generally divided into two categories: selecting the best one among multiple prediction results, using all the predicted values to weight the average, and then selecting.
Since the framed object may not be the relevant training set, it is impossible to track each frame of detection. Therefore, this paper needs to use deep learning to solve the above STT problem. This paper uses the Siamese region proposal network (SiamRPN) method to solve [35,36,37]. It can use large-scale images for offline end-to-end training. In general, this structure includes a Siamese subnetwork for feature extraction and a region proposal subnetwork. The candidate region generation network includes classification and regression. This network architecture is shown in Figure 7.
First, SiamRPN uses multi-scale testing to predict the change of scale to solve the problem that the previous algorithm cannot estimate the size of the target. Because the sliding window method is time-consuming, the system uses RPN to directly generate the detection frame, which can greatly increase the generation speed. In addition, the anchor technology is used to determine whether there is a recognized target in the fixed reference frame, and how far the target frame deviates from the reference frame, so there is no need for multi-scale traversal of sliding windows. In the network architecture of Figure 7, the Siamese network uses the AlexNet network structures [38,39,40]. The Siam feature extraction subnet has two branches, and the two branches share the same parameters in the CNN. One is called the template branch, which receives the target patch as input in the previous frame. The other is called the detection branch, which receives the target in the current frame as input. φ(z) is expressed as the output feature map of the template branch in the twin sub-network 6 × 6 × 256. φ(x) is expressed as the output feature map 22 × 22 × 256 of the detection branches in the Siamese network. The connection network structure can be found that the template image and the search image are respectively obtained by the Siamese network with 6 × 6 × 256 and 22 × 22 × 256 features, and then the template image features are respectively generated by a 3 × 3 convolution kernel features 4 × 4 × (2k × 256) and 4 × 4 × (4k × 256). In particular, the feature channel has been increased from the original 256 to 2k × 256 and 4k × 256. The reason why the number of channels has increased by 2k times is that k anchors are generated at each point of the feature map, and each anchor can be classified into the foreground or the background, so the classification branch has increased by 2k times. Similarly, each anchor can use four parameters as described, so the regression branch has increased by 4k times. At the same time, the search image also obtains two features through a 3 × 3 convolution kernel, where the number of feature channels remains unchanged. The RPN network architecture is shown in Figure 8, and the anchor method diagram is shown in Figure 9.
The RPN network is divided into two lines [41] in comparison to the Kalman filter approach with a self-learning RBFNN (Radial Basis Function Neural Network) [42]. The upper one uses softmax classification to make the anchors obtain positive and negative classifications, and the lower one is used to calculate the bounding box regression offset for the anchors to obtain an accurate proposal. The final Proposal layer is responsible for synthesizing positive anchors and the corresponding bounding box regression offset to obtain proposals, and at the same time eliminate proposals that are too small and beyond the boundary. In Figure 9, there are nine rectangles in the anchor method and there are three shapes in total. The aspect ratio of the schematic diagram in Figure 9 is {1:1,2:3,3:2}, and the multi-scale method commonly used in detection is introduced. For the classification branch, the 4 × 4 × 256 features of the 2k template image anchors are used as the convolution kernel and the search image feature is convolved to generate the classification branch response map 17 × 17 × 2k. The same is used in the regression branch, and the response map generated after the convolution operation is 17 × 17 × 4k. Each point represents a vector of size 4k, which is dx, dy, dw, and dh. The deviation between the anchor and ground truth is measured, and the calculation formula of the response graph is as follows:
A w × h × 2 k c l s = [ φ ( x ) ] c l s [ φ ( z ) ] c l s ,
A w × h × 2 k r e g = [ φ ( x ) ] r e g [ φ ( z ) ] r e g ,
where [φ(z)]cls and [φ(z)]reg means that φ(x) adds 2k classification channels and 4k regression channels, respectively, and ● represents the calculation of the correlation on the classification branch and the regression branch. Acls contains 2k channel vectors, each point in it represents positive and negative excitation, which is classified by softmax loss. Areg contains 4k channel vectors, each point represents the dx, dy, dw, and dh between anchor and ground truth. The above is normalized by the Smooth L1 loss function as follows:
L r e g = i = 0 3 s m o o t h L 1 ( δ [ i ] , σ ) ,
δ [ 0 ] = T x A x A w , δ [ 1 ] = T y A y A h δ [ 2 ] = ln T w A w , δ [ 3 ] = ln T h A h
s m o o t h L 1 ( x , σ ) = 0.5 σ 2 x 2 , x 1 σ 2 x 1 2 σ 2 , x 1 σ 2 ,
where Ax, Ay, Aw, Ah are the center point coordinates, length, and width of the anchor boxes; Tx, Ty, Tw, and Th are ground truth boxes. Lreg is the final regression loss, δ is the coordinate standardization of the anchor, and smoothL1(x, σ) is the Smooth L1 loss function.
Algorithm 1 is a key frame marking method, which can be used in the detection and monitoring process of target tracking. In this paper, the tracking task is planned as a one-shot detection task. This method limits the input structure and automatically discovers features that can be generalized from new samples. That is to learn a learner net, which corresponds to the similarity function in this paper, is trained through a supervised Siamese network-based metric learning, and then reuses the features extracted by that network for one-shot learning. This research regards the target tracking task as a combination of one-shot object detection and few-shot instance classification. The former is a class-level subtask used to find candidate frames similar to the target, and the latter is an instance-level task used to distinguish targets and distractors. Target-guidance module distinguishes the characteristics of the target and search area and their interaction with the subject. Although the detector is focused on objects related to the target, the surrounding background interference is ignored. To compensate for this, a few-shot instance classifier is proposed. However, training directly from scratch is time-consuming and easily leads to overfitting. Therefore, few-shot finetune is performed through model-agnostic meta-learning, which enhances discrimination and further eliminates distractors. SiamRPN elaborated on it from a theoretical point of view, in which the tracking framework is shown in Figure 10.
Algorithm 1 Keyframe detection
Inputs:f, the frame of the input video stream;
MAMf, the motion appearance mask of f;
MAMf–1, the motion appearance mask of f − 1;
tstop, the temporal threshold for detecting stop;
|MAMf∣ denotes the total number of ls in MAMf;
Outputs:lf, the label of the keyframe;
1.if |MAMf∣ > |MAMf−1
2.lf = SPLIT
3.else if |MAMf∣ < |MAMf−1
4.lf = JOIN
5.else if MAMf ^ MAMf−1 ≠ 0
6.lf = MOVE
7.else /*MAMf = MAMf−1 */
8.stop-count ← stop-count + 1
9.if stop-count > tstop
10.lf = SPLIT
Figure 10. The tracking framework in SiamRPN.
Figure 10. The tracking framework in SiamRPN.
Machines 10 00378 g010
In the detection part, according to this network architecture, the single sample detection task can be shown as the red box. The initial frame of the template frame passes through the convolutional layer in the RPN, and φ(x)reg and φ(x)cls are used. In the detection part, the average loss function, the definition of the classification feature map and the regression feature map formula are as follows:
min W 1 n i = 1 n L ( ζ ( φ ( x i ; W ) ; ( φ ( z i ; W ) ) , l i ) ,
A w × h × 2 k c l s = { ( x i c l s , y j c l s , c l c l s ) } ,
A w × h × 4 k r e g = { ( x i r e g , y j r e g , d x p r e g , d y p r e g , d w p r e g , d h p r e g ) } ,
where (14) is the average loss function L, li is the label, W is the weight of two networks, and ζ is the RPN operation. In (15), where i ∊ [0, w), j ∊ [0, h), l ∊ [0, 2k). In (16), where i ∊ [0, w), j ∊ [0, h), p ∊ [0, k). With the defined equation, the following relational equation, and the basis for selecting the best frame can be calculated:
C L S = { ( x i c l s , y j c l s , c l c l s ) i I , j J , l L } ,
A N C = { ( x i a n , y j a n , w l a n , h l a n ) i I , j J , l L } ,
R E G = { ( x i r e g , y j r e g , d x l r e g , d y l r e g , d w l r e g , d h l r e g ) i I , j J , l L } ,
P R O = { ( x i p r o , y j p r o , w l p r o , h l p r o ) i I , j J , l L } ,
x p r o = x a n + d x l r e g w l a n ,
y p r o = y a n + d y l r e g h l a n ,
w l p r o = w l p r o e d w l ,
h l p r o = h l p r o e d h l ,
where the top k values in the positive score found in the classification score are CLS*. It is found that the anchor box of the corresponding box is found to be ANC* and the predicted regression value is REG*, and finally the regression value is converted to the regression box PRO*. In the system, an represents the orbit generated by the anchor, and pro is the bounding box that is finally returned. Then, the anchors that are too far from the center are discarded to remove outliers, and then non-maximum suppression (NMS) is used to remove all non-maximum frames to remove redundant overlapping frames. The intersection over union (IoU) and NMS formulas needed to select the best frame are as follows:
I o U = Area   of   Intersection Area   of   Union ,
( w + p ) × ( h + p ) = s 2 ,
where IoU is the union of the intersection ratio of the two box areas, which is used to determine the pixel distance of the two boxes. In (26), w and h represent the width and height of the target, and p represents the filling value equal to (w + h)/2. First, select a box with the highest credibility, the rest of the boxes, and their IoU are greater than a certain threshold, then remove them, continue to select a box with the highest score from the unprocessed box, and repeat the above process to obtain the best.

3. Results

The target tracking currently used on USVs rarely uses Siamese neural network architecture and most of them are single-hull structures. The overall effect of this research is to achieve system integration, which includes three main modes: (1) fixed-way cruise and 360-degree panoramic monitoring. (2) real-time target tracking and USV following. (3) in the range Internal launch feedback module. To achieve the above objectives, this section will mention all the specifications, circuit diagrams, and algorithm results adopted by the system. The final product diagram and specification of the overall system are shown in Figure 11.
In order to estimate the difference between the USV’s travel route and the preset route, as shown in Figure 12, this paper uses two different routes for comparison. Path one is an arc-shaped curve, in which the trajectory with a straight line at the beginning can be used to compare whether it can meet the preset value in a straight line. Path two is a route to avoid obstacles and currents and is used to judge whether the USV can sail as usual under different wind speeds. Among them, each path has four tests at different times. The record of different trajectories path one and path two is shown in Figure 13. The green path is the first test, the purple path is the second test, the light blue path is the third test, and the dark blue path is the fourth test. In the four-day test, there was no wind and no water flow on the first day, light breeze on the second day, weak water flow, strong wind, and strong water flow on the third day, and light breeze but strong water flow on the fourth day.
The route information obtained from the four tests on path one and path two will be evaluated and verified with absolute trajectory error (ATE), as shown in Table 1. It can be seen from this that although the catamaran has strong resistance to water currents, it still has some continuous drop compared to static waters (path two). In the turning part, the two routes will deviate slightly, which may be caused by the wind speed change at that time or other underwater biological activities. It can be seen that the USV can fit as accurately as possible under no wind or breeze, and it can be overcome even under strong water flow. However, the performance is not satisfactory when encountering strong winds. The second part of the algorithm part is a panoramic stitching method based on feature comparison. The general method is to capture the feature points of two photos, and then connect the corresponding points together. The following paper will show the detailed stitching process step by step. This paper compares two pictures, and finally will provide the final complete stitching picture. All the steps and derivative pictures of the splicing method are shown in Figure 14, Figure 15, Figure 16, Figure 17 and Figure 18.
Table 1. Detailed specifications of the feedback module.
Table 1. Detailed specifications of the feedback module.
TrajectoryPath1Path2
MetricATE
Test_13.7454.045
Test_23.7493.704
Test_37.34119.363
Test_46.9584.499
Avg. Err.5.4737.903
Figure 14. The original picture of feature-based image alignment. (a) Picture to the left after centering the camera. (b) Picture to the right after centering the camera.
Figure 14. The original picture of feature-based image alignment. (a) Picture to the left after centering the camera. (b) Picture to the right after centering the camera.
Machines 10 00378 g014
Figure 15. Key points with the descriptors detected as in (a) Figure 14a. (b) Figure 14b.
Figure 15. Key points with the descriptors detected as in (a) Figure 14a. (b) Figure 14b.
Machines 10 00378 g015
Figure 16. Feature points matched.
Figure 16. Feature points matched.
Machines 10 00378 g016
Figure 17. Homograph corresponding to the two photos, Figure 14b is rotated and deformed based on Figure 14a.
Figure 17. Homograph corresponding to the two photos, Figure 14b is rotated and deformed based on Figure 14a.
Machines 10 00378 g017
Figure 18. Final corrected wide-angle image.
Figure 18. Final corrected wide-angle image.
Machines 10 00378 g018
The first step is to find and filter the key points and corresponding points in the photo. One can see the red circle in Figure 15. A key point will have four circles. The larger the circle, the more obvious the feature. The second step is to perform feature matching. There are many matching blue lines in Figure 16. Among them, this study only uses the feature points that have the distribution ratio reaching the top 85% score. The third step is Homography, which is a reversible transformation from the real projective plane to the projective plane. The straight line is still mapped to a straight line under this transformation, that is to say, the method of expanding the 3D plane to the plane. In the end, it is to merge the pictures to achieve the goal. The picture obtained in actual navigation according to the feature method is shown in Figure 17, and the picture obtained by synchronous stitching in actual navigation is shown in Figure 18.
The wide-angle image obtained is shown in Figure 19. This paper takes up to 500 key points and only selects the top 15% points for matchmaking.
The third part of the algorithm part is target tracking. This paper compares and analyzes SiamRPN, efficient convolution operators for tracking (ECO), continuous convolution operators (C-COT), and DaiSiamRPN target trackers, and proves the reliability of using SiamRPN. In addition, this paper will use the visual object tracking 2018 (VOT2018) and object tracking benchmark 100 (OTB100) data sets for short-term target tracking comparison. The evaluation index comparison in this section is used to discuss which tracker method to choose. VOT2018 is divided into four items: accuracy, robustness, loss, and expected average overlap (EAO) as the basis for verification and evaluation. The detailed calculation standards are as follows:
Φ t ( i ) = 1 N r e p k = 1 N r e p Φ t ( i , k ) ,
ρ A ( i ) = 1 N v a l i d j = 1 N v a l i d Φ j ( i ) ,
ρ R ( i ) = 1 N r e p k = 1 N r e p F ( i , k ) ,
Φ ^ = 1 N h i N 1 o N s = N 1 o : N h i Φ ^ N S ,
Overlap   score   O S =   a     b     a     b   ,
where the accuracy rate refers to the average overlap rate of the tracker in the test, that is, the IoU algorithm is used to compare the overlapping area of two rectangular boxes divided by the total area of the two rectangular boxes. Φt(i) is the definition of the average accuracy of each frame, where Φt(i, k) represents the accuracy of the ith tracking in repeated k frames. ρA(i) is the average accuracy of the entire video. Robustness refers to the number of failures of the test tracker. When the overlap rate of the rectangular frame is zero, it is judged as a failure, so the higher the value, the better. In (29), the function F (i, k) is defined as the number of tracking failures, and the measurement is repeated at the kth algorithm. EAO is the expected value of the non-reset overlap of each tracker on the short-term image sequence and is the most important indicator for evaluating the accuracy of the VOT target tracking algorithm. Success means that if the conformance rate score is higher than a certain value, it is regarded as a success. The higher the value, the better. ΦNs is the average coverage of the Ns in the video, and Φ(i) is the accuracy between the predicted frame and the real frame. As the video frame increases, the average coverage value will decrease because Φ(i) ≤ 1. In (31), the bounding box obtained by the tracking algorithm is a, and the box given by the ground-truth is b. When the OS of a frame is greater than the set threshold, the frame is regarded as Success, and the percentage of all successful frames to all frames is the success rate. The values of VOT2018 and OTB100 in each tracker are shown in Table 2 and Table 3, SiamRPN, ECO, C-COT, and DaiSiamRPN are compared. Target tracking in actual navigation is shown in Figure 20.
In Figure 20, it can be seen that good results can be achieved in all stages of target tracking, which improves the previous SiamFC problem. The SiamRPN selected in the system is the best tracker under comprehensive comparison. The AlexNet training loss as the backbone is shown in Figure 21. The blue line represents training, and the red line represents validation.
Figure 20. Target tracking of different scales in actual system navigation.
Figure 20. Target tracking of different scales in actual system navigation.
Machines 10 00378 g020
Figure 21. Target tracking of different scales in actual system navigation.
Figure 21. Target tracking of different scales in actual system navigation.
Machines 10 00378 g021
This study collected two different data sets: daytime and night. In these data sets, two frames per second are used as tests, and several ships of different durations are used for recording. Figure 22 is a database of different USV data during daytime and night. In the daytime and nighttime data, it can be found that the red hull target tracking efficiency is higher in the daytime data set. It may be because when the sunlight is too strong and the hull is brighter, which will cause reflections, making the lens unable to obtain sufficient feature values. In the night data set, the tracking efficiency of the white hull target is higher. The USVs have searchlights that illuminated on the tracked object in low light conditions, and the brighter the hull, the higher the feature variance can be obtained in contrast to the background. The following will test and verify the daytime data set and night data set, respectively. Among them, the precision, recall, and F1 scores of the day data set and night data set are shown in Table 4 below.
Table 4. Daytime and night data set evaluation.
Table 4. Daytime and night data set evaluation.
Data SetPrecisionRecallF1 Scores
Daytime0.850.610.711
Night0.740.820.778
Figure 22. Different USV data sets during daytime and night.
Figure 22. Different USV data sets during daytime and night.
Machines 10 00378 g022
The P-R curve can be obtained after the sum of the information obtained from the data set as shown in Figure 23. The confusion matrix of the daytime data set and the nighttime confusion matrix are shown in Table 5 and Table 6, respectively. The receiver operating characteristic (ROC) curve obtained through the sum of the two confusion matrices is shown in Figure 24.
Figure 23. P-R curve of the sum of information obtained by the data set.
Figure 23. P-R curve of the sum of information obtained by the data set.
Machines 10 00378 g023
Table 5. Confusion matrix of daytime in SiamRPN.
Table 5. Confusion matrix of daytime in SiamRPN.
Day-Time Data SetPredicted
PositiveNegative
ActualPositive1250782
Negative221575
Table 6. Confusion matrix of night in SiamRPN.
Table 6. Confusion matrix of night in SiamRPN.
Night-Time Data SetPredicted
PositiveNegative
ActualPositive751166
Negative263323
Figure 24. ROC curve of the data set.
Figure 24. ROC curve of the data set.
Machines 10 00378 g024
According to the above result evaluation data, precision and recall can achieve good results when the data sets are mixed. The area under the ROC curve is 0.72, which shows the superior performance of this system. The difference between light and dark can be clearly seen. In the daytime data set, the precision is higher, and the recall is lower, and in the night data set, the precision is lower, and the recall is higher. As far as the hardware is concerned, this study has verified the stability and proved that the stability of the catamaran is better than that of the monohull. In terms of software, SiamRPN, which has better accuracy and real-time performance, was chosen.

4. Discussion

This research has achieved the use of a USV to navigate, avoid obstacles, and perform target tracking, while achieving precise navigation, high accuracy of algorithm tracking, and high vehicle following speed. It can be clearly seen from the results that the system has significant improvements and enhancements to the above parts. In this section, it will be divided into hardware, algorithm, and integration parts for detailed discussion.
In the hardware part, this paper uses the information of the three-axis accelerometer and the three-axis gyroscope to compensate for the sensor in the system, which can improve the reliability of the sensor on the wave surface. In Figure 12, the Z-axis movement of the path before the compensation is too abrupt in the rotation stage, and the Z-axis movement of the path after the compensation is relatively smooth in the rotation stage. In addition, it replaced the monohull with Catamaran to improve stability and speed. Catamaran is that the waterline area is small, the interference force of the waves is small, and it has superior resistance in waves. In Figure 13, after the same use of the compensator, the monohull will amplify the wave amplitude due to its own hardware shortcomings, causing the information obtained by the sensor to be too extreme and lack reliability.
In the algorithm part, first, the system uses feature-based image stitching to expand the original viewing angle of only 55 degrees to a wider area without missing important information. In target tracking, this algorithm uses RPN so that it can instantly change the circle frame and track accurately, which is the best method compared to the comprehensive performance of ECO, C-COT, and DaSiamRPN. In the comparison between VOT2018 and VOT2018-LT, the effectiveness of SiamRPN and DaSiamRPN in Accuracy, Robustness, and EAO is significantly higher than that of ECO and C-COT. For the indicators in OTB: Success and Precision, SiamRPN performs better than DaSiamRPN. In addition, the effect of this algorithm on the test set is excellent, and the AOU can reach 0.72.
In the integration part, the most challenging is to make immediate responses to the sensing components and feedback components. It was originally expected that all sensing components and feedback components were placed on Jetson Xavier NX, but in experiments, it was found that if the target tracking algorithm and motor were activated, the remaining GPIO pins could not provide enough current to drive. In this study, methods such as pull-up resistors have been tried, and in the end, the use of dual control boards to interact with each other was chosen to achieve the best efficiency. This system uses Arduino and Jetson Xavier NX to communicate via USB using Python. The sensing component obtains the value through Arduino and sends it back to the main control board Jetson Xavier NX for judgment, correction, compensation, and response.
Other than our approach, various different approaches have been proposed as the use of laser scanners as machine vision systems in Unmanned Aerial Vehicle (UAV) navigation when compared with camera-based systems [43], autonomous robotic group behavior optimization during the mission on a distributed area in a cluttered hazardous terrain [44] and the machine vision systems to determine physical values of near distanced objects for Unmanned Aerial Vehicle (UAV) navigation [45].

5. Conclusions

In this research, it is proposed to use SiamRPN as USV target tracking and IMU as feedback to accurately locate ships and navigate fixed routes. This research is achieved under a special USV and embedded system. Because the USV is lighter and faster, it can be applied to the pursuit and rescue of smugglers. The main contributions of this paper are: (1) Improve the slow and poor accuracy of target tracking on common vehicles. (2) Combine IMU for dual-hull vehicles to improve deviated trajectories and wave undulations. (3) Combining image stitching methods based on feature points to reduce blind angles of sight. The experimental results show that this research can achieve target tracking and automatic navigation in different waters. Among the scheduled routes, this paper uses Catamaran’s way to replace monohulls to improve stability and speed. The method of image stitching is used to improve the problem of the blind angle of the viewing angle, so that the USV will not lose important information. At present, due to the current problem of the embedded system, two control boards are needed to meet all the requirements. In the future, it can be towards adding multiple lenses or 360-degree lenses to reduce the burden of algorithms and reduce other feedback sensors to reduce the current burden of the control board. For a future work, (1) more experiments on system stability of USV should be conducted over various water surface environments, (2) arbitrary multiple obstacles experiments should be carried out, and (3) more specifications should be evaluated (e.g., the duration to performs the tasks of object recognition and tracking on the water surface.), (4) comparison between monohull and catamaran type vehicle and (5) experiment and implement other control techniques to compare with the presented results.

Author Contributions

Conceptualization, M.-F.R.L.; methodology, M.-F.R.L. and C.-Y.L.; software, C.-Y.L.; validation, M.-F.R.L. and C.-Y.L.; formal analysis, M.-F.R.L.; investigation, M.-F.R.L.; resources, M.-F.R.L.; data curation, C.-Y.L.; writing—original draft preparation, C.-Y.L.; writing—review and editing, M.-F.R.L.; visualization, C.-Y.L.; supervision, M.-F.R.L.; project administration, M.-F.R.L.; funding acquisition, M.-F.R.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [Ministry of Science and Technology (MOST) in Taiwan] grant number [108-2221-E-011-142-] and [Center for Cyber-physical System innovation from the Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education (MOE) in Taiwan].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Suzuki, N.; Kitajima, H.; Kaba, H.; Suzuki, T.; Suto, T.; Kobayashi, A.; Ochi, F. An Experiment of Real-Time Data Transmission of Sonar Images from Cruising UUV to Distant Support Vessel via USV: Development of Underwater Real-Time Communication System (URCS) by Parallel Cruising. In Proceedings of the OCEANS 2015—Genova, Genova, Italy, 18–21 May 2015; pp. 1–6. [Google Scholar]
  2. Simetti, E.; Turetta, A.; Casalino, G.; Cresta, M. Towards the Use of a Team of USVs for Civilian Harbour Protection: The Problem of Intercepting Detected Menaces. In Proceedings of the OCEANS’10 IEEE SYDNEY, Sydney, NSW, Australia, 24–28 May 2010; pp. 1–7. [Google Scholar]
  3. Kent, B.M.; Ehret, R.A. Rethinking Intelligence, Surveillance, and Reconnaissance in a Wireless Connected World. In Proceedings of the IEEE International Symposium on Antennas and Propagation, Chicago, IL, USA, 8–14 July 2012; pp. 1–2. [Google Scholar]
  4. Duan, L.; Luo, B.; Li, Q.-Y.; Yu, G.-H. Research on intelligence, surveillance and reconnaissance mission planning model and method for naval fleet. In Proceedings of the Chinese Control and Decision Conference, Yinchuan, China, 28–30 May 2016; pp. 2419–2424. [Google Scholar]
  5. Alshbatat, A.I.N.; Alhameli, S.; Almazrouei, S.; Alhameli, S.; Almarar, W. Automated Vision-based Surveillance System to Detect Drowning Incidents in Swimming Pools. In Proceedings of the Advances in Science and Engineering Technology International Conferences, Dubai, United Arab Emirates, 4 February–9 April 2020; pp. 1–5. [Google Scholar]
  6. Uz, S.S.; Ames, T.J.; Memarsadeghi, N.; McDonnell, S.M.; Blough, N.V.; Mehta, A.V.; McKay, J.R. Supporting Aquaculture in the Chesapeake Bay Using Artificial Intelligence to Detect Poor Water Quality with Remote Sensing. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 3629–3632. [Google Scholar]
  7. Lambrou, T.P.; Anastasiou, C.C.; Panayiotou, C.; Polycarpou, M.M. A Low-Cost Sensor Network for Real-Time Monitoring and Contamination Detection in Drinking Water Distribution Systems. IEEE Sens. J. 2014, 14, 2765–2772. [Google Scholar] [CrossRef]
  8. Yunsheng, F.; Yutong, S.; Guofeng, W. On Model Parameter Identification and Trajectory Tracking Control for USV Based on Backstepping. In Proceedings of the 36th Chinese Control Conference, Dalian, China, 26–28 July 2017; pp. 4757–4761. [Google Scholar]
  9. Wang, P.; Wang, J. A Tracking Method Based on Target Classification and Recognition. In Proceedings of the IEEE Advanced Information Technology, Electronic and Automation Control Conference, Chengdu, China, 20–22 December 2019; pp. 255–259. [Google Scholar]
  10. Xu, Y.; Xu, K.; Wan, J.; Xiong, Z.; Li, Y. Research on Particle Filter Tracking Method Based on Kalman Filter. In Proceedings of the IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference, Xi’an, China, 25–27 May 2018; pp. 1564–1568. [Google Scholar]
  11. Gundogdu, E.; Alatan, A.A. Method for learning deep features for correlation based visual tracking. In Proceedings of the Signal Processing and Communications Applications Conference, Antalya, Turkey, 9–11 June 2017; pp. 1–4. [Google Scholar]
  12. Ge, Y.; Zhong, L.; Qiang, Z.J. Research on Underactuated USV Path Following Algorithm. In Proceedings of the Information Technology, Networking, Electronic and Automation Control Conference, Chongqing, China, 12–14 June 2020; pp. 2141–2145. [Google Scholar]
  13. Ge, Y.; Zhong, L.; Qiang, Z.J. Research on USV Heading Control Method Based on Kalman Filter Sliding Mode Control. In Proceedings of the Cambridge Crystallographic Data Centre, Hefei, China, 29 April 2020; pp. 1547–1551. [Google Scholar]
  14. Han, J.; Cho, Y.; Kim, J. Coastal SLAM with Marine Radar for USV Operation in GPS-Restricted Situations. IEEE J. Ocean. Eng. 2019, 44, 300–309. [Google Scholar] [CrossRef]
  15. Zhou, X.; Wu, P.; Zhang, H.; Guo, W.; Liu, Y. Learn to Navigate: Cooperative Path Planning for Unmanned Surface Vehicles Using Deep Reinforcement Learning. IEEE Access 2019, 7, 165262–165278. [Google Scholar] [CrossRef]
  16. Mane, S.; Mangale, S. Moving Object Detection and Tracking Using Convolutional Neural Networks. In Proceedings of the International Confederation of Contamination Control Societies, Madurai, India, 21 September 2018; pp. 1809–1813. [Google Scholar]
  17. Inoue, Y.; Ono, T.; Inouer, K. Situation-Based Dynamic Frame-Rate Control for on-Line Object Tracking. In Proceedings of the International Japan-Africa Conference on Electronics, Communications and Computations, Alexandria, Egypt, 17–19 December 2018; pp. 119–122. [Google Scholar]
  18. Koskowich, B.J.; Rahnemoonfai, M.; Starek, M. Virtualot—A Framework Enabling Real-Time Coordinate Transformation & Occlusion Sensitive Tracking Using UAS Products, Deep Learning Object Detection & Traditional Object Tracking Techniques. In Proceedings of the IEEE International Symposium on Geoscience and Remote Sensing, Valencia, Spain, 23–27 July 2018; pp. 6416–6419. [Google Scholar]
  19. Liu, Y.; Meng, Z.; Zou, Y.; Cao, M. Visual Object Tracking and Servoing Control of a Nano-Scale Quadrotor: System, Algorithms, and Experiments. IEEE/CAA J. Autom. Sin. 2021, 8, 344–360. [Google Scholar] [CrossRef]
  20. Zhou, Y.; Wang, T.; Hu, R.; Su, H.; Liu, Y.; Liu, X.; Suo, J.; Snoussi, H. Multiple Kernelized Correlation Filters (MKCF) for Extended Object Tracking Using X-Band Marine Radar Data. IEEE Trans. Signal Process. 2019, 67, 3676–3688. [Google Scholar] [CrossRef]
  21. Varfolomieiev, A.; Lysenko, O. Modification of the KCF tracking method for implementation on embedded hardware platforms. In Proceedings of the International Conference Radio Electronics & Info Communications, Kiev, Ukraine, 11–16 September 2016; pp. 1–5. [Google Scholar]
  22. Kalal, Z.; Mikolajczyk, K.; Matas, J. Tracking-Learning-Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Dattathreya; Han, S.; Kim, M.-J.; Maik, V.; Paik, J. Keypoint-based object tracking using modified median flow. In Proceedings of the IEEE International Conference on Consumer Electronics-Asia, Seoul, Korea, 26–28 October 2016; pp. 1–2. [Google Scholar]
  24. Jiu-Cai, J.; Jie, Z.; Feng, S. Modelling, manoeuvring analysis and course following for two unmanned surface vehicles driven by a single propeller and double propellers. In Proceedings of the 27th Chinese Control and Decision Conference, Qingdao, China, 23–25 May 2015; pp. 4932–4937. [Google Scholar]
  25. Sun, S.; Wang, N.; Liu, Y.; Dai, B. Fuzzy heading control of a rotary electric propulsion ship with double propellers. In Proceedings of the Proceedings of the 33rd Chinese Control Conference, Nanjing, China, 28–30 July 2014; pp. 4598–4602. [Google Scholar] [CrossRef]
  26. Raimondi, F.M.; Trapanese, M.; Franzitta, V.; Viola, A.; Colucci, A. A innovative semi-immergible USV (SI-USV) drone for marine and lakes operations with instrumental telemetry and acoustic data acquisition capability. In Proceedings of the OCEANS 2015—Genova, Genova, Italy, 18–21 May 2015; pp. 1–10. [Google Scholar]
  27. Zhu, Q. Design of control system of USV based on double propellers. In Proceedings of the IEEE International Conference of IEEE Region 10 (TENCON 2013), Xi’an, China, 26–29 August 2013; pp. 1–4. [Google Scholar]
  28. Nuari, R.; Utami, E.; Raharjo, S. Comparison of Scale Invariant Feature Transform and Speed Up Robust Feature for Image Forgery Detection Copy Move. In Proceedings of the International Conference on Information Technology, Information Systems and Electrical Engineering, Yogyakarta, Indonesia, 20–21 November 2019; pp. 107–112. [Google Scholar]
  29. Uddin, M.Z.; Khaksar, W.; Torresen, J. Activity Recognition Using Deep Recurrent Neural Network on Translation and Scale-Invariant Features. In Proceedings of the IEEE International Conference on Image Processing, Yogyakarta, Indonesia, 8 October 2018; pp. 475–479. [Google Scholar]
  30. Al-Shuibi, A.; Aldarawani, A.; Al-Homaidi, H.; Al-Soswa, M. Survey on Image Retrieval Based on Rotation, Translation and Scaling Invariant Features. In Proceedings of the First International Conference of Intelligent Computing and Engineering, Hadhramout, Yemen, 15–16 December 2019; pp. 1–11. [Google Scholar]
  31. Hou, W.; Li, D.; Xu, C.; Zhang, H.; Li, T. An Advanced k Nearest Neighbor Classification Algorithm Based on KD-tree. In Proceedings of the IEEE International Conference of Safety Produce Informatization, Chongqing, China, 10–12 December 2018; pp. 902–905. [Google Scholar]
  32. Zhang, J.; Shi, H. Kd-Tree Based Efficient Ensemble Classification Algorithm for Imbalanced Learning. In Proceedings of the International Conference on Machine Learning, Big Data and Business Intelligence, Taiyuan, China, 8–10 November 2019; pp. 203–207. [Google Scholar]
  33. Wei, L.; Ding, M.; Zhang, X. Single Target Tracking Using Reliability Evaluation and Feature Selection. In Proceedings of the International Symposium on Computational Intelligence and Design, Hangzhou, China, 14–15 December 2019; pp. 228–231. [Google Scholar]
  34. Mitchell, A.E.; Smith, G.E.; Bell, K.L.; Rangaswamy, M. Single target tracking with distributed cognitive radar. In Proceedings of the IEEE Radar Conference, Seattle, WA, USA, 8–12 May 2017; pp. 285–288. [Google Scholar]
  35. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High Performance Visual Tracking with Siamese Region Proposal Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8971–8980. [Google Scholar]
  36. Zhou, H.; Ni, B. Tracking of drone flight by neural network Siamese-RPN. In Proceedings of the International Conference on Engineering, Applied Sciences and Technology, Chiang Mai, Thailand, 1–4 July 2020; pp. 1–3. [Google Scholar]
  37. Fan, H.; Ling, H. Siamese Cascaded Region Proposal Networks for Real-Time Visual Tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7944–7953. [Google Scholar]
  38. Huang, F.; Yu, L.; Shen, T.; Jin, L. Chinese Herbal Medicine Leaves Classification Based on Improved AlexNet Convolutional Neural Network. In Proceedings of the IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference, Chengdu, China, 20–22 December 2019; pp. 1006–1011. [Google Scholar]
  39. Beeharry, Y.; Bassoo, V. Performance of ANN and AlexNet for weed detection using UAV-based images. In Proceedings of the International Conference on Emerging Trends in Electrical, Electronic and Communications Engineering, Réduit, Mauritius, 28–30 November 2020; pp. 163–167. [Google Scholar]
  40. Fairuz, S.; Habaebi, M.H.; Elsheikh, E.M.A. Finger Vein Identification Based on Transfer Learning of AlexNet. In Proceedings of the International Conference on Computer and Communication Engineering, Hyderabad, India, 22–27 October 2018; pp. 465–469. [Google Scholar]
  41. Shih, K.-H.; Chiu, C.-T.; Lin, J.-A.; Bu, Y.-Y. Real-Time Object Detection with Reduced Region Proposal Network via Multi-Feature Concatenation. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 2164–2173. [Google Scholar] [CrossRef] [PubMed]
  42. Xu, S.S.-D.; Huang, H.-C.; Chiu, T.-C.; Lin, S.-K. Biologically-inspired learning and adaptation of self-evolving control for networked mobile robots. Appl. Sci. 2019, 9, 1034. [Google Scholar]
  43. Lindner, L.; Sergiyenko, O.; Rivas-Lopez, M.; Ivanov, M.; Rodriguez-Quinonez, J.C.; Hernandez-Balbuena, D.; Flores-Fuentes, W.; Tyrsa, V.; Muerrieta-Rico, F.N.; Mercorelli, P. Machine vision system errors for unmanned aerial vehicle navigation. In Proceedings of the 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), Edinburgh, UK, 19–21 June 2017; pp. 1615–1620. [Google Scholar] [CrossRef]
  44. Ivanov, M.; Sergyienko, O.; Tyrsa, V.; Lindner, L.; Flores-Fuentes, W.; Rodríguez-Quiñonez, J.; Hernandez, W.; Mercorelli, P. Influence of data clouds fusion from 3D real-time vision system on robotic group dead reckoning in unknown terrain. IEEE/CAA J. Autom. Sin. 2020, 7, 368–385. [Google Scholar] [CrossRef]
  45. Lindner, L.; Sergiyenko, O.; Rivas-Lopez, M.; Valdez-Salas, B.; Rodriguez-Quinonez, J.C.; Hernandez-Balbuena, D.; Flores-Fuentes, W.; Tyrsa, V.; Barrera, M.; Muerrieta-Rico, F.N.; et al. Machine vision system for UAV navigation. In Proceedings of the 2016 International Conference on Electrical Systems for Aircraft, Railway, Ship Propulsion and Road Vehicles & International Transportation Electrification Conference (ESARS-ITEC), Toulouse, France, 2–4 November 2016; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. The hierarchical control architecture diagram.
Figure 1. The hierarchical control architecture diagram.
Machines 10 00378 g001
Figure 2. Architecture of high-level control vehicle.
Figure 2. Architecture of high-level control vehicle.
Machines 10 00378 g002
Figure 3. USV envisioned design drawing.
Figure 3. USV envisioned design drawing.
Machines 10 00378 g003
Figure 4. Overall system diagram as (a) system module; (b) system data training and classification.
Figure 4. Overall system diagram as (a) system module; (b) system data training and classification.
Machines 10 00378 g004
Figure 5. SIFT feature vector generation steps.
Figure 5. SIFT feature vector generation steps.
Machines 10 00378 g005
Figure 6. Basic structure and steps of STT system.
Figure 6. Basic structure and steps of STT system.
Machines 10 00378 g006
Figure 7. The main neural network framework of SiamRPN: the blue part is the Siamese subnet used for feature extraction, the green part is the regional proposal subnet, and the orange part is the final output result. There are two branches in the figure, one for classification and the other for regression.
Figure 7. The main neural network framework of SiamRPN: the blue part is the Siamese subnet used for feature extraction, the green part is the regional proposal subnet, and the orange part is the final output result. There are two branches in the figure, one for classification and the other for regression.
Machines 10 00378 g007
Figure 8. The RPN architecture used in this SiamRPN.
Figure 8. The RPN architecture used in this SiamRPN.
Machines 10 00378 g008
Figure 9. The anchor method architecture used in this SiamRPN.
Figure 9. The anchor method architecture used in this SiamRPN.
Machines 10 00378 g009
Figure 11. The USV as: (a) side view; (b) front view.
Figure 11. The USV as: (a) side view; (b) front view.
Machines 10 00378 g011
Figure 12. Two paths tested in established waters.
Figure 12. Two paths tested in established waters.
Machines 10 00378 g012
Figure 13. T Four test routes and preset routes, listed as: (a) path 1; (b) path 2.
Figure 13. T Four test routes and preset routes, listed as: (a) path 1; (b) path 2.
Machines 10 00378 g013
Figure 19. The picture obtained by synchronous stitching in actual navigation according to the feature method.
Figure 19. The picture obtained by synchronous stitching in actual navigation according to the feature method.
Machines 10 00378 g019
Table 2. Evaluation of VOT2018 by the system.
Table 2. Evaluation of VOT2018 by the system.
TrackerAccuracy
(ρA(i))
Robustness
(ρR(i))
EAO
(Φ)
SiamRPN0.6010.3370.318
ECO0.4840.2760.281
C-COT0.5360.1840.378
DaiSiamRPN0.6010.3370.327
Table 3. Evaluation of OTB100 by the system.
Table 3. Evaluation of OTB100 by the system.
TrackerSuccess (OS)Precision
SiamRPN0.6940.914
ECO0.6910.910
C-COT0.6710.898
DaiSiamRPN0.6580.881
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, M.-F.R.; Lin, C.-Y. Object Tracking for an Autonomous Unmanned Surface Vehicle. Machines 2022, 10, 378. https://doi.org/10.3390/machines10050378

AMA Style

Lee M-FR, Lin C-Y. Object Tracking for an Autonomous Unmanned Surface Vehicle. Machines. 2022; 10(5):378. https://doi.org/10.3390/machines10050378

Chicago/Turabian Style

Lee, Min-Fan Ricky, and Chin-Yi Lin. 2022. "Object Tracking for an Autonomous Unmanned Surface Vehicle" Machines 10, no. 5: 378. https://doi.org/10.3390/machines10050378

APA Style

Lee, M. -F. R., & Lin, C. -Y. (2022). Object Tracking for an Autonomous Unmanned Surface Vehicle. Machines, 10(5), 378. https://doi.org/10.3390/machines10050378

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop