Next Article in Journal
A New Method for Measuring the Rotational Angles of a Precision Spherical Joint Using Eddy Current Sensors
Next Article in Special Issue
A Binarized Segmented ResNet Based on Edge Computing for Re-Identification
Previous Article in Journal
High-Efficiency Microsatellite-Using Super-Resolution Algorithm Based on the Multi-Modality Super-CMOS Sensor
Previous Article in Special Issue
Multiple Object Tracking for Dense Pedestrians by Markov Random Field Model with Improvement on Potentials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking

1
School of Computer Science and Engineering, Kyungpook National University, Daegu 41566, Korea
2
Department of Computer Science, Information Technology University, Lahore 54000, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(14), 4021; https://doi.org/10.3390/s20144021
Submission received: 17 June 2020 / Revised: 3 July 2020 / Accepted: 15 July 2020 / Published: 20 July 2020
(This article belongs to the Special Issue Visual Sensors for Object Tracking and Recognition)

Abstract

:
We propose to improve the visual object tracking by introducing a soft mask based low-level feature fusion technique. The proposed technique is further strengthened by integrating channel and spatial attention mechanisms. The proposed approach is integrated within a Siamese framework to demonstrate its effectiveness for visual object tracking. The proposed soft mask is used to give more importance to the target regions as compared to the other regions to enable effective target feature representation and to increase discriminative power. The low-level feature fusion improves the tracker robustness against distractors. The channel attention is used to identify more discriminative channels for better target representation. The spatial attention complements the soft mask based approach to better localize the target objects in challenging tracking scenarios. We evaluated our proposed approach over five publicly available benchmark datasets and performed extensive comparisons with 39 state-of-the-art tracking algorithms. The proposed tracker demonstrates excellent performance compared to the existing state-of-the-art trackers.

1. Introduction

Visual Object Tracking (VOT) is a promising, attractive, and challenging field in computer vision with a wide range of real-world applications including robotics [1], autonomous vehicles [2], video understanding [3], surveillance, and security [4]. Given the initial target location in the first frame of a video, the goal of an object tracker is to estimate the new target positions for subsequent frames. VOT is an active research area owing to the challenges such as occlusion, the presence of various types of noise, appearance and scale variations of the target, environmental changes, motion blur, illumination variations, and background clutter.
Correlation Filter-based Trackers (CFTs) have widely been used due to robustness and low computational complexity [5,6,7,8]. CFTs exploit circulant target structure to replace exhaustive spatial correlation operations with effective element-wise multiplication in the frequency domain to achieve high speed. Furthermore, CFTs use fairly simple image features such as a Histogram of Oriented Gradients (HOG), color-name [9] features, and a color-histogram. Despite these advantages, CFTs performance may drop under difficult and complex scenarios, due to these hand-crafted features. The performance of CFTs may be improved by using deep features, due to CFT’s inherent ability to learn discrimination between the target and the background.
Deep learning has widely been utilized in computer vision applications such as image classification [10,11], action recognition [12,13], semantic segmentation [14,15], and pose estimation [16,17]. Deep CFTs use deep features extracted from pre-trained networks over large benchmarks to compute discriminative features that result in improved performance compared to the hand-crafted CFTs [18]. Although many deep trackers [19,20,21,22,23,24,25] have demonstrated good performance, a limitation of pre-trained models is that they do not fully capture a specific target appearance. Therefore, many deep models [26,27,28,29] are trained end-to-end to further improve the performance. Another limitation is that during tracking, models update to adapt to new target appearances but may encounter an over-fitting problem. Moreover, online deep learning is computationally expensive and requires extra computational resources for feature extraction, update operations, and inference processes.
Recently, Siamese networks have been used for tracking to address some of these limitations. Siamese trackers are popular, owing to underlying properties such as competitive performance and computational efficiency. The basic principle behind a Siamese network is to find the similarity between input images. Siamese-based trackers can be simplified as a similarity learning problem: a similarity measure is learned in order to compute the similarity between the template image and the candidate image. Siamese trackers including [30,31,32,33,34] are computationally efficient but exhibit performance degradation under many scenarios. These trackers learn similarity by training offline on the large benchmarks and do not learn the most discriminative features for a specific target, which reduces the tracking performance. One approach is to assign different weights to the target parts on the basis of previous information [35,36], which is inaccurate. However, in the current manuscript, we introduce a soft-mask to highlight the target pixel information in the spatial domain during training. Our model estimates the new target information efficiently without computing the soft-mask during test time. Moreover, computation of the soft-mask for each frame requires extra computational cost, which reduces the tracking speed.
Recently, attention methods are studied to improve the feature representations in various computer vision applications [15,37,38,39,40]. Motivated by these applications, many trackers are utilizing different kinds of attentions in their tracking framework. For example, the tracker [41] is motivated by Selective Kernel Networks (SKNet) [39]. Authors computed separate feature weights for different convolutional layers. While tracker [42] is inspired by Convolutional Block Attention Module (CBAM) [37] and computes feature descriptors utilizing both global average and global maximum pooling layers for channel as well as spatial attention. The RASNet [43] model integrates three types of attentions such as residual attention, channel attention, and general attention to produce more discriminative target features. The authors of [15] utilized masks to modulate the features for video object segmentation.
We propose to integrate two different attention mechanisms in the Siamese tracking framework to emphasize discriminative channels and important spatial features in the latent space. The proposed attention mechanism boosts the underlying model to learn ’which’ and ’where’ target information should be highlighted. In this work, we emphasize to highlight target information in spatial as well as in latent space. We construct a soft-mask to highlight target pixel information in the spatial domain. However, in latent space, we integrate channel and spatial attention modules within the proposed tracking framework. Proposed channel attention computes two separate weight vectors and then fuses to highlight important features and reduces the less important features. Our proposed spatial attentional module computes weights for each feature map using a simple model without utilizing the global average and maximum pooling layers.
In particular, we propose a Soft-mask with Channel and Spatial attentional Siamese (SCS-Siam) tracking framework that learns both effective and discriminative features. We extend the underlying Siamese network architecture to exploit the target-specific information. The proposed SCS-Siam network encodes more target-specific information than contextual knowledge to learn the similarity between inputs and exhaustively exploits the template features to capture a specific target representation. A soft mask is computed to focus on the region of interest while suppressing background information in the image space. A spatial feature fusion operation is performed by combining the features from both the input image and the generated soft mask image. A channel attention mechanism is integrated for the template branch to produce robust discriminative target features by highlighting the most useful channels while suppressing the less important channels. A spatial attention mechanism is also utilized after channel attention to emphasize the important spatial regions within a channel for better target location identification. Both channel and spatial attentions are integrated within the Siamese framework using a skip connection. The proposed model is trained offline to enable generalized tracking and to enhance the robustness of the tracker. Extensive experiments are performed to evaluate the proposed SCS-Siam algorithm over five benchmark datasets including OTB2015 [44], TempleColor123 [45], UAV123 [46], VOT2016 [47], and VOT2017 [48].
The main contributions of the current manuscript are as follows:
  • We propose a soft mask feature fusion mechanism to highlight the full target region compared to the background region. It helps the network to efficiently learn the target representation.
  • A channel attentional mechanism is proposed to give the discriminative channels more importance.
  • A spatial attention mechanism is proposed to emphasize the discriminative spatial locations within the target.
  • Soft mask feature fusion with dual attention is integrated within a Siamese tracking framework using a skip connection to enhance the tracker ability to better discriminate target from the background.
  • The proposed SCS-Siam tracker has shown excellent performance compared to 39 existing trackers over five benchmark datasets.
The rest of the paper is organized as follows. Section 2 presents the related work, Section 3 explains our proposed framework, Section 4 describes the experiments and evaluations and finally, Section 5 presents a conclusion and future research directions.

2. Related Work

In this section, we explore deep feature-based, Siamese-based, and attention-based trackers. Detailed research on trackers can be found in [18,49,50].

2.1. Deep Feature-Based Trackers

Deep learning has demonstrated ground-breaking performance in the tracking field. One notable limitation in visual tracking, however, is the limited availability of training samples. Most deep trackers use pretrained models to extract deep features trained over a large benchmark for object classification. Deep trackers [19,22] compute complementary features from shallow and semantic layers to obtain promising tracking results into the correlation filter. Moreover, deep features are exploited at various convolutional layers to boost the performance of visual trackers [20,21,24,51]. In contrast, a combination of features from different layers does not always guarantee a performance gain, mainly due to the increasing number of dimensions, assorted resolutions, and unknown target information [51]. VITAL [52] explored adversarial learning to produce efficient sample features and used a cost-sensitive loss function to leverage from the class imbalance. Other deep trackers [28,52,53,54,55] based on decision-making approaches have been proposed, such as Support Vector Machines (SVM), regression, and classification networks. Hong et al. [53] proposed CNN-SVM to perform a classification task using SVM and CNN models with saliency maps. MDNet [28] captured the domain-dependent information and performed tracking task as classification in a particle framework. Spatial and temporal information was encoded using CNNs for classification by Teng et al. [54]. Wang et al. [55] introduced a features-selection procedure based on a regression framework. These trackers use rich feature representations from deep networks but are limited in tracking performance and may drift, due to noisy updates during online learning. Moreover, additional computational cost is required to update these networks to capture a new target appearances.

2.2. Siamese Network-Based Trackers

Recently, Siamese networks have been utilized to develop robust visual trackers, drawing significant attraction in the visual tracking community owing to the real-time inference [30,33,34,56,57,58,59]. Siamese trackers learn the similarity between input images and cast the tracking problem as a matching problem. Siamese trackers perform tracking by comparing the initial target template features with search region features for every incoming frame. Siamese networks share the benefits of offline learning on large benchmarks to yield generic object tracking. Bertinetto et al. [30] developed SiameseFC computed complementary features using embedded CNN models and fused them to produce a response map. CFnet [34] was proposed to introduce a correlation layer in the template branch of the SiameseFC to produce superior results. GOTURN [33] was proposed to compute the similarity between two consecutive frames using a simple feed-forward network. Re3 [56] was proposed to utilize the recurrent network and obtain a better target representation. Guo et al. [57] proposed DSiam to suppress background information and performed online learning to capture target appearance variations. Tianyu and Antoni proposed MeemTrack [58] and integrated a dynamic memory network within a Siamese architecture. These methods are pre-trained on large benchmarks to learn similarity from pair-wise inputs; however, over-fitting may occur from learning on similar benchmarks. Moreover, these Siamese methods do not fully exploit the target information. The authors of [35,36] compute non-overlapping patches and assign weights to reflect the patch importance. In contrast, we extend the underlying Siamese architecture and compute a soft mask feature fusion to exploit the target information to highlight the objectness information while suppressing contextual information for better target feature representation.

2.3. Attention Mechanism-Based Trackers

An attention mechanism is popular across computer vision fields, including activity recognition [60], image classification [37,61], pose estimation [62], and semantic segmentation [40,63]. The RTT algorithm [64] uses multi-directional recurrent neural networks to produce saliency maps and draws attention to possible targets. Discriminative spatial attention is used by DAVT [65]. The SA-Siam algorithm [66] enhances the discriminative ability of the semantic branch by incorporating a channel attention module. The RASNet [43] model integrates three types of attentions: residual attention, channel attention, and general attention to produce more discriminative target features. Abdelpakey et al. [67] proposed DenseSiam and focused on non-local target features by using self attention in a template branch. The CSRDCF [68] algorithm constrains correlation filter learning by introducing spatial reliability that uses a spatial binary mask. In the current work, we focus on critical information and re-calibrate the channels for better discrimination. Proposed channel attention learns which deep channels should be highlighted for better target feature discrimination. We also exploit the target and background location information and focus where the pixel information should be highlighted or suppressed within each deep channel. The proposed channel and spatial attention modules exploit the intermediate features effectively to learn the ‘which’ and ‘where’ target information to focus or suppress.

3. Proposed SCS-Siam Network

The overall framework of the proposed Soft-mask with Channel and Spatial attentional Siamese (SCS-Siam) architecture is shown in Figure 1. Compared to the previous deep trackers, the proposed SCS-Siam learns the target object information by highlighting it while suppressing the contextual information by using a soft mask. As illustrated in Figure 1, the SCS-Siam fuses the soft mask features to adapt the learned target model with the appearance variations. Moreover, channel and spatial attention mechanisms are introduced to better utilize the intra-channel and inter-channel features for inference. The main components of the proposed SCS-Siam tracker are discussed in detail in the following sections.

3.1. Baseline SiameseFC Tracker

The SiameseFC works on the principle of template-matching and is a building block of our proposed tracker. The SiameseFC performs tracking by formulating the tracking problem as a general matching function. Embedded CNN models are used to compute the deep features for the input patches. SiameseFC is composed of two parallel branches known as the exemplar z branch and the search x branch. The goal of SiameseFC is to find the maximum similarity between the input images indicating the new target location. A response map g ( z , x ) is computed using a cross-correlation function as:
g ( z , x ) = ψ ( z ) ψ ( x ) + b ,
where ψ ( · ) represents the embedding model, ∗ represents the cross-correlation function, and b means the offset. The Equation (1) performs both feature representation and discriminative learning simultaneously, which may lead to the problem of over-fitting. However, the superior feature representation that preserves the target object information while reducing the impact of background can increase the tracking accuracy. In the current work, we present soft mask based feature fusion and two attention modules to improve the tracking performance.

3.2. Soft-Mask Feature Fusion

To develop a generic robust tracker and overcome the limitations in the baseline tracker, we propose a soft-mask feature fusion technique to suppress the background information compared to the target information. The underlying principle is that in Siamese networks, all components do not participate equally in the cross-correlation. For example, an object within the green bounding box should ponder more to compute a cross-correlation than the outside region of target bounding box, as illustrated in Figure 2. Moreover, to discriminate the target from the background, the visual difference between the target and the background must be more distinct. Thus, at the training stage, we introduced a soft mask generation process to train the network to exploit more target information for cross-correlation. The proposed mask generation module creates a contrast between the target and the background to enhance the discriminative ability of the tracker. The soft mask suppresses the background information outside the target bounding box. We constructed a masked image by providing lower weights σ   1 to the background. To do so, we multiply the input pixel values by a σ outside the bounding box. Figure 3 presents three examples of the soft mask generation.
Suppose, there is an input image I and its target bounding box B B in the first row of Figure 3. A soft-mask m is constructed on the basis of B B , as shown in the second row. The soft-mask is applied over the input image, as displayed in the third row. Template T and template with soft-mask T s m are generated after data curation. Similarly, Search image S and search with soft-mask S s m image are generated. Usually, the target can be on arbitrary location in the frame. To fit the model, we crop and resize the soft-masked image such that the target is centered (similar to [30]). Note that at the test time, the target region is not known precisely, therefore σ = 1 is used. During training, the value of σ   =   0.90 is empirically found to be the best performer on OTB2015 and hence used for all datasets. The details of the empirical study are given in Table 9.
To get an effective and discriminative feature representation, we integrated the soft mask for both the template and search region in the proposed network. The proposed framework SCS-Siam takes four inputs including the template, soft mask template, search patch, and soft mask search patch, as illustrated in Figure 1. To obtain a better and efficient feature representation, we fused the features for soft masks at early convolutional layers. The soft mask features contain complementary localization information of the object, as shown in Figure 4. Feature fusion at early convolutional layers encodes the spatial target localization information and it increases the fidelity of the network to produce effective feature representation. The features for template and search branches are fused as:
Z = B ( T ) + B ( T s m ) ,
X = B ( S ) + B ( S s m ) ,
where B represents a convolutional block including a convolutional layer, a normalization layer, a rectifier layer, and a pooling layer. The proposed framework is defined as:
g ( T , S ) = ( Ω ( Δ ( ψ ( Z ) ) ) ψ ( Z ) ) ψ ( X ) + b ,
where ψ ( . ) denotes the backbone network parameters, Δ ( . ) represents the proposed channel attention module, and Ω ( . ) shows the proposed spatial attentional module. The output of ψ ( Z ) is fed to the proposed channel attention as Δ ( ψ ( Z ) ) . Then, this output is forwarded to the proposed spatial attention module as Ω ( Δ ( ψ ( Z ) ) ) . The output of spatial attentional module is element-wise added with ψ ( Z ) . Then, finally, a response map is calculated from Equation (4). Please see Figure 1 for more details.

3.3. Channel Attention Module

We exploit the inter-channel relationship by integrating a soft channel-attentional mechanism into the proposed tracking framework. A special kind of visual pattern is captured by each deep channel during convolution. Deep channels behave differently and play a vital role to compute inference. Therefore, the channel attentional mechanism is a process of selecting important visual patterns for better inference. The channel attention module focuses on the most useful information for better discrimination. The objective of channel attention is to increase the adaption ability of the tracker by strengthening the most useful channels and reducing the impact of less important channels. SENet [69], SA-Siam [66], and RASNet [43] learn the channel attention using a max pooling and multi perception layer. In contrast to these, rather then employing single pooling layer, our channel exploits the channel relationships explicitly for both global max pooling and global average pooling using two separate sub-networks. The global average pooling exhibits the overall knowledge, whereas global average pooling indicates the finer object information for feature channels. The channels’ weight coefficients from our sub-networks yield better descriptor weights.
To exploit the channel attention, we proposed the light weight deep channel attentional network illustrated in Figure 5. Given an input feature map P containing c deep channels, features are forwarded to two independent sub-networks. Global Average Pooling (GAP) and Global Maximum Pooling (GMP) operations are performed for each sub-network separately to compute 1 × 1 × c dimension descriptors. In each sub-network, the channels are decreased by a fully connected layer and then again increased using another fully connected layer. A Sigmoid activation is performed for each sub-network to compute descriptor weights. These descriptors are fused using an element-wise summation operation. The channel weights ω R c are computed as:
ω = Sigmoid ( F C 2 ( Relu ( F C 1 ( G A P ( P ) ) ) ) ) + Sigmoid ( F C 2 ( Relu ( F C 1 ( G M P ( P ) ) ) ) ) .
The weights ω are channel-wise multiplied with input features P to obtain S shown in Figure 5.

3.4. Spatial Attention Module

We also propose a spatial attention mechanism within our network to emphasize the important spatial locations of the target feature map. CBAM [37] computes spatial attention by computing global average and maximum pooling across channel to generate feature descriptor while DANet [40] exhibits more complexity to compute spatial attention. RASNet [43] utilizes a Guassian map to exploit spatial reliability. This approach has limitations, as an arbitrary object may contain contexual information and the network should give weights less to contextual information. On the other hand, the proposed spatial attention computes the global contextual descriptors using a simple model.
In addition to the proposed soft-mask, the spatial attention mechanism highlights the important target locations within each channel in the latent space. Complementary to channel attention, our spatial attention module focuses on the most informative target regions within each channel. The output from the channel attention module is forwarded to the spatial attention module. The proposed spatial attention framework has two components including context and transform models, as shown in Figure 6. The context mode is responsible for computing the same spatial attention for all the feature channels. To compute the context model feature maps, we first apply a convolution layer to reduce the large number of channels to a single channel and then apply the softmax layer. The output from the softmax layer is multiplied with input feature maps, as shown in Figure 6. Contrary to the context model, transform is responsible for computing the different spatial attentions across the channels. To do so, the output of the context model is forwarded to multiple layers to first reduce the number of channels and then increase the number of channels. Then a sigmoid activation function is applied and finally the output is achieved using a skip connection from the input feature maps.

3.5. Network Training

The proposed network is trained as a generalized object tracker by learning on a large benchmark GOT-10K dataset [70], which is composed of 1.5 million annotated frames and more than 10000 video sequences. Similar to SiameseFC [30], we define a template patch size of 127 × 127 × 3 and a search region size of 255 × 255 × 3. We also generate a template with a soft mask of size 127 × 127 × 3 and a search with a soft mask of size 255 × 255 × 3. The network is trained offline using Stochastic Gradient Descent (SGD) to yield a response map g ( T k , T k s m , S k , S k m ) for input images and a Y k ground truth map, where T k represents the template image, T k s m shows the template with soft mask, S k denotes the search, and S k s m refers to the search with a soft mask, as shown in Figure 1. The network parameters θ were trained to minimize the following loss function:
argmin θ 1 K k = 1 K L ( g ( T k , T k s m , S k , S k s m ) , Y k ) ,
where θ represents the network parameters, K represents the total number of training samples, and L ( . ) represents the logistic loss function computed as:
L ( g k , y k ) = 1 | δ | ( i , j ) δ l o g ( 1 + e x p ( g k ( i , j ) . y k ( i , j ) ) ,
where g k ( i , j ) and y k ( i , j ) { + 1 , 1 } represent the similarity value and ground truth label corresponding to the ( i , j ) th position on the score map, δ represents the set of positions in the search window on the score map.
We present the training and testing algorithm for our proposed framework in Algorithm 1 and Algorithm 2, respectively.
Algorithm 1: Offline Training of the proposed framework
Sensors 20 04021 i001
Algorithm 2: Tracking of the proposed method
Sensors 20 04021 i002

4. Experiments and Results

We propose three different versions of tracking framework including Soft mask with Channel and Spatial attention Siamese (SCS-Siam), Soft mask with Channel Attention Siamese (SCA-Siam), and Soft mask with Spatial Attention Siamese (SSA-Siam). We evaluate the proposed tracking versions on five benchmark datasets including OTB2015 [44], TempleColor128 [45], UAV123 [46], VOT2016 [47], and VOT2017 [48]. We compare its performance with 39 state-of-the-art methods including TRACA [71], SRDCF [72], UDT [73], SiamtTri [31], SiameseFC [30], Staple [74], CFNet [34], CNNSI [75], RNN [76], MLT [77], P2FNet [78], SiamFc-lu [79], DSiamM [57], Li et al. [80], Kuai et al. [81], SINT [82], HASiam [83], ECO [21], MEEM [84], SAMF [85], MUSTER [86], CSK [6], CMKCF [87], SATIN [88], ACT [89], MemTrack [58], DSiam [90], GradNet [91], SiameseRPN [92], ACFN [93], CSRDCF [68], SCT [94], KCF [7], SSKCF [47], DPT [95], DSST [96], CCOT [20], SiamDCF [97], and UCT [98].

4.1. Implementation Details

The proposed tracker was trained offline to attain generalization on GOT-10K [70]. The model was trained using the Stochastic Gradient Descent (SGD) method to minimize Equation (6). The network parameters are shown in Table 1. We set the weight decay to 10 5 and momentum to 0.9. The initial learning rate was set to 10 2 and decreased exponentially until reaching 10 5 . During training, the background information weight in the generation of the soft mask in Section 3.2 was set to 0.9, and the target bounding box information was set to 1. To improve the tracking performance, we used larger images during inference. The sizes for the input images of both the exemplar and search branches were set to 135 × 135 × 3 and 263 × 263 × 3 respectively. During testing, we did not compute the soft mask but instead cloned the template and search region to compute inference. To address the scale variations across consecutive video frames, we constructed a pyramid for the target consisting of three different scales { 1 . 0375 1 , 1 , 1.0375 } at the current frame based on the previous target position. During inference, we got three response maps and we selected the best response matching the target scale. Code was implemented in Python 3.7 using PyTorch. Experiments were performed on a machine having Intel i7 3.6 GHz processor, 32 GB of RAM, and an NVIDIA TITAN XP GPU card. The average tracking speed of the proposed tracker was 73 frames per second (FPS).

4.2. Datasets and Evaluation Metrics

OTB2015 [44] contains 100 fully annotated sequences with 11 different tracking challenges. Temple Color-128 (TC128) [45] is also a widely used tracking benchmark. Compared to OTB2015, it has 128 challenging videos and also contains 11 tracking challenges. The UAV123 dataset consists of 123 videos captured from Unmanned Aerial Vehicle (UAV) at a low-altitude and all videos are fully annotated [46]. We used One Pass Evaluation (OPE) for evaluation. Performance evaluations are performed using precision and success to measure for the aforementioned datasets. The former metric is computed as the Euclidean distance between the center location of the ground truth and the predicted location and is defined as the percentage of the frames where the Euclidean distance lies within a 20 pixels threshold. The latter metric computes the Overlap Score (OS) using an intersection over union. A frame is considered a success if its OS exceeds a threshold of 0.50. The comparisons were conducted over the VOT2016 and VOT2017 datasets. During the evaluation of these datasets, the tracker was re-initialized if it failed. We used the official toolkit and three parameters including the Expected Average Overlap (EAO), Robustness (R), and accuracy (A) to compare the tracking performance. Details of the aforementioned datasets are presented in Table 2.

4.3. Experiments on OTB2015

On OTB2015 dataset, we compared our algorithms with eight existing state-of-the-art algorithms including TRACA [71], SRDCF [72], UDT [73], SiamtTri [31], SiameseFC [30], Staple [74], CFNet [34], and CNNSI [75]. Our proposed SCA-Siam, SCS-Siam and SSA-Siam trackers demonstrated better performance compared to TRACA and all other trackers in terms of both precision and success over OTB2015, as illustrated in Figure 7. SCA-Siam gained 62.2% success, which is 2% higher than TRACA. Likewise, SCA-Siam algorithm obtained 2.6% more precision than TRACA over OTB2015.
We also compared our SCA-Siam, SSA-Siam, and SCS-Siams version with other various state-of-the-art trackers over OTB2015. Table 3 presents the comparison of various trackers based on three parameters precision, success, and speed in FPS. The performance of our algorithm SCA-Siam surpassed the other methods in terms of precision and success. TRACA performed tracking at 101 FPS, but its precision and success score were less than our SCA-Siam by 2.6% and 2% respectively. Kuai et al. [81] secured 62.2% success but exhibited less precision and speed than our algorithms.
We also conducted the experiments to exploit the robustness for the SCA-Siam, SSA-Siam, and SCS-Siam algorithms for 11 tracking challenges including occlusion, deformation, fast motion, motion blur, light and scale variations, and others. We selected the OTB2015 benchmark to evaluate for different challenges, as presented in Figure 8 and Table 4. We observed that the proposed algorithm SCS-Siam secured first rank in terms of success when compared with other trackers for five challenges: fast motion, occlusion, deformation, motion blur, and in-plane rotation, as shown in Figure 8. Our SCA-Siam showed better performance for three challenges such as scale variations, background clutter, and low resolution in terms of success. However, our SCS-Siam revealed slightly less performance than TRACA for two groups of videos including illumination variations and out-of-plane rotation. Table 4 presents the attribute-based performance in terms of precision scores for eight state-of-the-art trackers. Our SCA-Siam demonstrated best performance for nine challenges including Fast Motion (FM), Background Clutter (BC), Deformation (Def), Illumination Variations (IV), Occlusion (OCC), In-Plane Rotation (IPR), Low Resolution (LR), Out of View (OV), and Scale Variations (SV). Our SCS-Siam showed better performance for the Motion Blur (MB) challenge in terms of precision. TRACA showed best performance for Out-of-Plane Rotation (OPR), while our SCA-Siam ranked second.
We further compared the tracking results qualitatively with CNNSI, UDT, SRDCF, SiamTri, and SiameseFC trackers, as illustrated in Figure 9. We preformed qualitative experiments for selected sequences from OTB2015 such as CarDark, Skating2-2, Basketball, Bird2, Jogging-1, and Box videos. Our SCS-Siam tracker did not lose the target object for any sequence, as illustrated in Figure 9.

4.4. Experiments on TC128 and UAV123

We evaluated the performance for TC128 and UAV123 datasets using precision and success. Table 5 presents the comparison of the proposed method over TC128 with the following trackers: SCT [94], KCF [7], CNNSI [75], ACT [89], UDT [73], CFNet [34] and baseline tracker SiameseFC [30]. Our algorithm SCS-Siam showed better performance compared to other trackers in terms of precision and achieved the highest scores 74.2 % . While our SSA-Siam secured best overlap 54.2 % . KCF is computationally effective, exhibiting its tracking 160 FPS; however, our SCA-Siam, SSA-Siam, and SCS-Siam surpassed the KCF for both success and precision.
We validate the performance of our trackers over the UAV123 dataset using precision and success. Figure 10 demonstrates the plots for precision and success for the UAV123 dataset. We note that SCS-Siam surpassed other methods in performance and achieved the best scores for precision (75.2 %) and overlap (52.4%).

4.5. Experiments on VOT2016 and VOT2017

Table 6 presents the Expected Average Overlap (EAO), accuracy (A), and Robustness (R) of the compared trackers over VOT2016 [47] for the baseline experiments. Our methods are compared with 11 trackers: MemTrack [58], MemDTC [90], ECO [21], SRDCF [72], DSiam [57], CMKCF [87], Staple [74], CCOT [20], UDT [73], and SiameseFC [30]. CCOT obtained the maximum EAO score, and its robustness value is larger than our proposed algorithm, and overlap scores is less than our algorithm, as presented in Table 6. Overall, our algorithms demonstrated the better accuracy and robustness. Moreover, our algorithms showed the highest accuracy score against all the compared trackers.
We compute Expected Average Overlap (EAO), accuracy (A), and Robustness (R) to perform comparison over VOT2017 [48] in Table 7. We compared our algorithms with 12 other trackers: CSRDCF [68], MemTrack [58], MemDTC [90], SRDCF [72], DSST [96], SATIN [88], SiamDCF [97], UCT [98], SiameseFC [30], GradNet [91], and SiameseRPN [92]. CSRDCF surpassed our method for both EAO (0.25). However, our trackers achieved more accuracy, low robustness, and demonstrated superior computational efficiency compared to CSRDCF. SATIN has a high EAO of 0.28, but our algorithms have comparatively high overlap scores and low robustness. Furthermore, our SCS-Siam obtained a favorably high overlapping score and low robustness compared to the other trackers.

4.6. Ablation Study

We performed an extensive ablation study to validate the effectiveness of different components of the proposed method. We performed experiments for each version and compared the results using precision and success over OTB2015 [44]. We also compared our methods with the baseline SiameseFC tracker, and evaluated the effectiveness of the three integrated modules.
First, we performed a series of experiments to select the best σ for soft mask generation. During training, we set different values of σ to weight background information to generate soft-mask images. During testing, we set the background weight to 1 for all experiments and present the performance in Table 8 in terms of precision and success. For example, in Table 8, SCS-Siam- 0.0 means that SCS-Siam was trained with σ = 0.0 such that it has no background information to generate soft mask images. Similarly, SCS-Siam- 0.3 σ = 0.3 was used during training. We observe that SCS-Siam- 0.9 exhibited the best performance where soft-mask images were generated when σ was set with σ = 0.9 . In an additional experiment, we also tested the proposed SCS-Siam with soft-masks generated with σ = 0.9 during the test time as well. We observed that this choice showed degradation in performance compared to σ = 1 . It is because that the target estimation at previous frame may not be accurate due to distractors. Therefore, this degradation in performance may be due to the construction of an inaccurate soft mask from the previous frame.
In Table 9, we present the comparison of different versions of the proposed framework. SiameseFC is a baseline tracker trained over the ImageNet-ILSVRC2015 dataset and the SiameseFC* is trained over GOT-10k dataset. The proposed framework is referred to as the ‘Extended-SiameseFC’, which is trained over GOT-10K without the soft-mask. We notice that SiameseFC* performed better than the SiameseFC, which is because the earlier was trained over a larger dataset compared to the former. We also observe that the proposed Extended-SiameseFC showed improved tracking performance than SiameseFC*. Because of the low level feature fusion, it achieved better localization. We compared our three versions of the proposed algorithm including SCA-Siam, SSA-Siam, and SCS-Siam to validate the integration of channel attention module, spatial attention module, and both modules, respectively. SCA-Siam refers to the channel attention module integrated in the proposed framework, while SSA-Siam represents the proposed spatial attention. SCS-Siam indicates that the proposed framework contains both channel and spatial attention modules. Table 9 shows that SCA-Siam achieved better performance compared to SCS-Siam and SSA-Siam trackers. We also investigate the order of the spatial and channel attention in the proposed framework. SCS-Siam means that it has a channel attention followed by the spatial attention module. While SSC-Siam represents the tracker that has spatial attention followed by channel attention. The experimental results demonstrate that the inclusion of the channel attention module before the spatial attention shows superior tracking performance. We also integrate the channel and spatial attention modules for exemplar and search branches within the proposed framework denoted as B-SCS-Siam in Table 9. It can be seen that no performance gain is obtained, whereas the speed decreases when channel and spatial attentions are applied for both exemplar and search branches.
Proposed channel attention and spatial attention modules consist of two sub-networks, as shown in Figure 5 and Figure 6, respectively. We performed different experiments to validate the importance of each sub-network for both channel and spatial attention modules, as shown in Table 10. For SCA-Siam, we performed experiments for each sub-network. For example, SCA-GMP indicates the SCA-Siam that utilizes global maximum pooling sub-network only while SCA-GAP represents the SCA-Siam that has a global average pooling sub-network only. Similar to SCA-Siam, we also performed two experiments for SSA-Siam. SSA-Context represents the SSA-Siam without transform sub-network while SSA-Transform indicates the SSA-Siam without context sub-network. Experimental results in Table 10 revealed to us that each sub-network of the channel and spatial attention modules is important to achieve better tracking performance.

5. Conclusions

In this work, a soft mask based spatial feature fusion method is proposed which is further strengthened with channel and spatial attention mechanisms. The proposed approach is embedded within a Siamese framework to obtain improved tracking accuracy and robustness. The soft mask based feature fusion mechanism produces efficient and discriminative features for effective tracking. During training, soft-masks were used to highlight the target information region. Features from original and soft-mask images were fused for both template and search branches, which enhanced the tracker’s localization ability. The channel attention mechanism was utilized to exploit the rich feature channels for discrimination while reducing the weights for less informative channels. A spatial attention module was also proposed to enhance the inter-channel localization ability of the tracker. The proposed framework was evaluated over five publicly available tracking benchmark datasets exhibiting significant improvement over 39 state-of-the-art trackers.

Author Contributions

Conceptualization, M.F.; methodology, M.F.; validation, M.F., A.M. and S.K.J.; formal analysis, A.M. and S.K.J.; investigation, S.K.J. and A.M.; resources, S.K.J.; data curation, M.F.; writing–original draft preparation, M.F. and A.M.; writing–review and editing, M.F., A.M. and S.K.J.; visualization, S.K.J.; supervision, A.M. and S.K.J.; project administration, S.K.J.; funding acquisition, S.K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This study was supported by the BK21 Plus project (SW Human Resource Development Program for Supporting Smart Life) funded by the Ministry of Education, School of Computer Science and Engineering, Kyungpook National University, Korea (21A20131600005). In part, this research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (NRF-2019R1A2C1010786).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Gupta, M.; Kumar, S.; Behera, L.; Subramanian, V.K. A novel vision-based tracking algorithm for a human-following mobile robot. IEEE Trans. Syst. Man, Cybern. Syst. 2016, 47, 1415–1427. [Google Scholar] [CrossRef]
  2. Menze, M.; Geiger, A. Object scene flow for autonomous vehicles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3061–3070. [Google Scholar]
  3. Renoust, B.; Le, D.D.; Satoh, S. Visual analytics of political networks from face-tracking of news video. IEEE Trans. Multimed. 2016, 18, 2184–2195. [Google Scholar] [CrossRef]
  4. Yao, H.; Cavallaro, A.; Bouwmans, T.; Zhang, Z. Guest editorial introduction to the special issue on group and crowd behavior analysis for intelligent multicamera video surveillance. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 405–408. [Google Scholar] [CrossRef]
  5. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
  6. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the circulant structure of tracking-by-detection with kernels. In Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
  7. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Kiani Galoogahi, H.; Fagg, A.; Lucey, S. Learning background-aware correlation filters for visual tracking. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1135–1143. [Google Scholar]
  9. Hu, S.; Ge, Y.; Han, J.; Zhang, X. Object Tracking Algorithm Based on Dual Color Feature Fusion with Dimension Reduction. Sensors 2019, 19, 73. [Google Scholar] [CrossRef] [Green Version]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  11. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  12. Feichtenhofer, C.; Pinz, A.; Zisserman, A. Convolutional two-stream network fusion for video action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1933–1941. [Google Scholar]
  13. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal segment networks: Towards good practices for deep action recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 20–36. [Google Scholar]
  14. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  15. Fiaz, M.; Mahmood, A.; Jung, S.K. Video Object Segmentation using Guided Feature and Directional Deep Appearance Learning. In Proceedings of the 2020 DAVIS Challenge on Video Object Segmentation—CVPR, Workshops, Seattle, WA, USA, 19 June 2020. [Google Scholar]
  16. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 483–499. [Google Scholar]
  17. Ranjan, R.; Patel, V.M.; Chellappa, R. Hyperface: A deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 121–135. [Google Scholar] [CrossRef] [Green Version]
  18. Fiaz, M.; Mahmood, A.; Javed, S.; Jung, S.K. Handcrafted and Deep Trackers: Recent Visual Object Tracking Approaches and Trends. Acm Comput. Surv. (CSUR) 2019, 52, 43. [Google Scholar] [CrossRef]
  19. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Convolutional features for correlation filter based visual tracking. In Proceedings of the 2015 IEEE International Conference on Computer Vision Workshop (ICCVW), Santiago, Chile, 7–13 December 2015; pp. 58–66. [Google Scholar]
  20. Danelljan, M.; Robinson, A.; Khan, F.S.; Felsberg, M. Beyond correlation filters: Learning continuous convolution operators for visual tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 472–488. [Google Scholar]
  21. Danelljan, M.; Bhat, G.; Shahbaz Khan, F.; Felsberg, M. Eco: Efficient convolution operators for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6638–6646. [Google Scholar]
  22. Ma, C.; Huang, J.B.; Yang, X.; Yang, M.H. Hierarchical convolutional features for visual tracking. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3074–3082. [Google Scholar]
  23. Mueller, M.; Smith, N.; Ghanem, B. Context-aware correlation filter tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1396–1404. [Google Scholar]
  24. Qi, Y.; Zhang, S.; Qin, L.; Yao, H.; Huang, Q.; Lim, J.; Yang, M.H. Hedged deep tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4303–4311. [Google Scholar]
  25. Song, Y.; Ma, C.; Gong, L.; Zhang, J.; Lau, R.W.; Yang, M.H. Crest: Convolutional residual learning for visual tracking. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2555–2564. [Google Scholar]
  26. Fan, H.; Ling, H. Sanet: Structure-aware network for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 42–49. [Google Scholar]
  27. Han, B.; Sim, J.; Adam, H. Branchout: Regularization for online ensemble tracking with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3356–3365. [Google Scholar]
  28. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar]
  29. Yun, S.; Choi, J.; Yoo, Y.; Yun, K.; Young Choi, J. Action-decision networks for visual tracking with deep reinforcement learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2711–2720. [Google Scholar]
  30. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 850–865. [Google Scholar]
  31. Dong, X.; Shen, J. Triplet loss in siamese network for object tracking. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 459–474. [Google Scholar]
  32. Fiaz, M.; Mahmood, A.; Jung, S.K. Deep Siamese Networks toward Robust Visual Tracking. In Visual Object Tracking with Deep Neural Networks; IntechOpen: London, UK, 2019. [Google Scholar]
  33. Held, D.; Thrun, S.; Savarese, S. Learning to track at 100 fps with deep regression networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 749–765. [Google Scholar]
  34. Valmadre, J.; Bertinetto, L.; Henriques, J.; Vedaldi, A.; Torr, P.H. End-to-end representation learning for correlation filter based tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2805–2813. [Google Scholar]
  35. Li, C.; Liang, X.; Lu, Y.; Zhao, N.; Tang, J. RGB-T object tracking: Benchmark and baseline. Pattern Recognit. 2019, 96, 106977. [Google Scholar] [CrossRef] [Green Version]
  36. Li, C.; Lin, L.; Zuo, W.; Tang, J.; Yang, M.H. Visual tracking via dynamic graph learning. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2770–2782. [Google Scholar] [CrossRef] [Green Version]
  37. Woo, S.; Park, J.; Lee, J.Y.; So Kweon, I. Cbam: Convolutional block attention module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27 October–3 November 2019. [Google Scholar]
  39. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 510–519. [Google Scholar]
  40. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  41. Fiaz, M.; Rahman, M.M.; Mahmood, A.; Farooq, S.S.; Baek, K.Y.; Jung, S.K. Adaptive Feature Selection Siamese Networks for Visual Tracking. In Proceedings of the International Workshop on Frontiers of Computer Vision, Ibusuki, Japan, 20–22 February 2020; pp. 167–179. [Google Scholar]
  42. Rahman, M.M.; Fiaz, M.; Jung, S.K. Efficient Visual Tracking with Stacked Channel-Spatial Attention Learning. IEEE Access 2020, 8. [Google Scholar] [CrossRef]
  43. Wang, Q.; Teng, Z.; Xing, J.; Gao, J.; Hu, W.; Maybank, S. Learning attentions: Residual attentional Siamese network for high performance online visual tracking. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4854–4863. [Google Scholar]
  44. Wu, Y.; Lim, J.; Yang, M.H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Liang, P.; Blasch, E.; Ling, H. Encoding color information for visual tracking: Algorithms and benchmark. IEEE Trans. Image Process. 2015, 24, 5630–5644. [Google Scholar] [CrossRef] [PubMed]
  46. Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for uav tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 445–461. [Google Scholar]
  47. Kristan, M.; Leonardis, A.; Matas, J.; Felsberg, M.; Pflugfelder, R.; Cehovin, L.; Vojir, T.; Hager, G.; Lukezic, A.; Fernandez, G.; et al. The Visual Object Tracking VOT2016 challenge results. In Proceedings of the European Conference on Computer Vision Workshop, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  48. Kristan, M.; Leonardis, A.; Matas, J.; Felsberg, M.; Pflugfelder, R.; Cehovin, L.; Vojir, T.; Hager, G.; Lukezic, A.; Eldesokey, A.; et al. The visual object tracking vot2017 challenge results. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1949–1972. [Google Scholar]
  49. Fiaz, M.; Mahmood, A.; Jung, S.K. Tracking Noisy Targets: A Review of Recent Object Tracking Approaches. arXiv 2018, arXiv:1802.03098. [Google Scholar]
  50. Li, P.; Wang, D.; Wang, L.; Lu, H. Deep visual tracking: Review and experimental comparison. Pattern Recognit. 2018, 76, 323–338. [Google Scholar] [CrossRef]
  51. Bhat, G.; Johnander, J.; Danelljan, M.; Shahbaz Khan, F.; Felsberg, M. Unveiling the power of deep tracking. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 483–498. [Google Scholar]
  52. Song, Y.; Ma, C.; Wu, X.; Gong, L.; Bao, L.; Zuo, W.; Shen, C.; Lau, R.W.; Yang, M.H. Vital: Visual tracking via adversarial learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8990–8999. [Google Scholar]
  53. Hong, S.; You, T.; Kwak, S.; Han, B. Online tracking by learning discriminative saliency map with convolutional neural network. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 597–606. [Google Scholar]
  54. Teng, Z.; Xing, J.; Wang, Q.; Lang, C.; Feng, S.; Jin, Y. Robust object tracking based on temporal and spatial deep networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1144–1153. [Google Scholar]
  55. Wang, L.; Ouyang, W.; Wang, X.; Lu, H. Visual tracking with fully convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3119–3127. [Google Scholar]
  56. Gordon, D.; Farhadi, A.; Fox, D. Re 3: Re al-Time Recurrent Regression Networks for Visual Tracking of Generic Objects. IEEE Robot. Autom. Lett. 2018, 3, 788–795. [Google Scholar] [CrossRef] [Green Version]
  57. Guo, Q.; Feng, W.; Zhou, C.; Huang, R.; Wan, L.; Wang, S. Learning dynamic siamese network for visual object tracking. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1763–1771. [Google Scholar]
  58. Yang, T.Y.; Antoni, B.C. Learning Dynamic Memory Networks for Object Tracking. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 152–167. [Google Scholar]
  59. Fiaz, M.; Mahmood, A.; Jung, S.K. Improving Object Tracking by Added Noise and Channel Attention. Sensors 2020, 20, 3780. [Google Scholar] [CrossRef]
  60. Zhao, Y.; Liu, Z.; Yang, L.; Cheng, H. Combing rgb and depth map features for human activity recognition. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–4. [Google Scholar]
  61. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 2017–2025. [Google Scholar]
  62. Du, W.; Wang, Y.; Qiao, Y. Rpan: An end-to-end recurrent pose-attention network for action recognition in videos. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 3725–3734. [Google Scholar]
  63. Chen, L.C.; Yang, Y.; Wang, J.; Xu, W.; Yuille, A.L. Attention to scale: Scale-aware semantic image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3640–3649. [Google Scholar]
  64. Cui, Z.; Xiao, S.; Feng, J.; Yan, S. Recurrently target-attending tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1449–1458. [Google Scholar]
  65. Fan, J.; Wu, Y.; Dai, S. Discriminative spatial attention for robust tracking. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 480–493. [Google Scholar]
  66. He, A.; Luo, C.; Tian, X.; Zeng, W. A twofold siamese network for real-time object tracking. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 4834–4843. [Google Scholar]
  67. Abdelpakey, M.H.; Shehata, M.S.; Mohamed, M.M. Denssiam: End-to-end densely-siamese network with self-attention model for object tracking. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 19–21 November 2018; pp. 463–473. [Google Scholar]
  68. Lukezic, A.; Vojir, T.; Cehovin Z, L.; Matas, J.; Kristan, M. Discriminative correlation filter with channel and spatial reliability. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6309–6318. [Google Scholar]
  69. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  70. Huang, L.; Zhao, X.; Huang, K. Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [Green Version]
  71. Choi, J.; Jin Chang, H.; Fischer, T.; Yun, S.; Lee, K.; Jeong, J.; Demiris, Y.; Young Choi, J. Context-aware deep feature compression for high-speed visual tracking. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 479–488. [Google Scholar]
  72. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Learning spatially regularized correlation filters for visual tracking. In Proceedings of the 2015 IEEE International Conference on Computer Vision Workshop (ICCVW), Santiago, Chile, 7–13 December 2015; pp. 4310–4318. [Google Scholar]
  73. Wang, N.; Song, Y.; Ma, C.; Zhou, W.; Liu, W.; Li, H. Unsupervised Deep Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  74. Bertinetto, L.; Valmadre, J.; Golodetz, S.; Miksik, O.; Torr, P.H. Staple: Complementary learners for real-time tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NY, USA, 26 June–1 July 2016; pp. 1401–1409. [Google Scholar]
  75. Fiaz, M.; Mahmood, A.; Jung, S.K. Convolutional neural network with structural input for visual object tracking. In Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing, Long Beach, CA, USA, 16–20 June 2019; pp. 1345–1352. [Google Scholar]
  76. Wang, L.; Liu, T.; Wang, B.; Lin, J.; Yang, X.; Wang, G. Learning Hierarchical Features for Visual Object Tracking With Recursive Neural Networks. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipai, Taiwan, 22–25 September 2019; pp. 3088–3092. [Google Scholar]
  77. Choi, J.; Kwon, J.; Lee, K.M. Deep meta learning for real-time target-aware visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 911–920. [Google Scholar]
  78. Baik, S.; Kwon, J.; Lee, K.M. Learning to Remember Past to Predict Future for Visual Tracking. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipai, Taiwan, 22–25 September 2019; pp. 3068–3072. [Google Scholar]
  79. Li, B.; Xie, W.; Zeng, W.; Liu, W. Learning to Update for Object Tracking With Recurrent Meta-Learner. IEEE Trans. Image Process. 2019, 28, 3624–3635. [Google Scholar] [CrossRef]
  80. Li, G.; Peng, M.; Nai, K.; Li, Z.; Li, K. Multi-view correlation tracking with adaptive memory-improved update model. Neural Comput. Appl. 2019. [Google Scholar] [CrossRef]
  81. Kuai, Y.; Wen, G.; Li, D. Masked and dynamic siamese network for robust visual tracking. Inf. Sci. 2019, 503, 169–182. [Google Scholar] [CrossRef]
  82. Tao, R.; Gavves, E.; Smeulders, A.W. Siamese instance search for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2016; pp. 1420–1429. [Google Scholar]
  83. Shen, J.; Tang, X.; Dong, X.; Shao, L. Visual object tracking by hierarchical attention siamese network. IEEE Trans. Cybern. 2019. [Google Scholar] [CrossRef] [PubMed]
  84. Zhang, J.; Ma, S.; Sclaroff, S. MEEM: Robust tracking via multiple experts using entropy minimization. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 188–203. [Google Scholar]
  85. Li, Y.; Zhu, J. A scale adaptive kernel correlation filter tracker with feature integration. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 254–265. [Google Scholar]
  86. Hong, Z.; Chen, Z.; Wang, C.; Mei, X.; Prokhorov, D.; Tao, D. Multi-store tracker (muster): A cognitive psychology inspired approach to object tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 749–758. [Google Scholar]
  87. Huang, B.; Xu, T.; Jiang, S.; Chen, Y.; Bai, Y. Robust Visual Tracking via Constrained Multi-Kernel Correlation Filters. IEEE Trans. Multimed. 2020. [Google Scholar] [CrossRef]
  88. Gao, P.; Yuan, R.; Wang, F.; Xiao, L.; Fujita, H.; Zhang, Y. Siamese attentional keypoint network for high performance visual tracking. Knowl. Based Syst. 2019. [Google Scholar] [CrossRef] [Green Version]
  89. Chen, B.; Wang, D.; Li, P.; Wang, S.; Lu, H. Real-time’Actor-Critic’Tracking. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 318–334. [Google Scholar]
  90. Yang, T.; Chan, A.B. Visual Tracking via Dynamic Memory Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  91. Li, P.; Chen, B.; Ouyang, W.; Wang, D.; Yang, X.; Lu, H. Gradnet: Gradient-guided network for visual object tracking. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6162–6171. [Google Scholar]
  92. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8971–8980. [Google Scholar]
  93. Choi, J.; Jin Chang, H.; Yun, S.; Fischer, T.; Demiris, Y.; Young Choi, J. Attentional correlation filter network for adaptive visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4807–4816. [Google Scholar]
  94. Choi, J.; Jin Chang, H.; Jeong, J.; Demiris, Y.; Young Choi, J. Visual tracking using attention-modulated disintegration and integration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NY, USA, 26 June–1 July 2016; pp. 4321–4330. [Google Scholar]
  95. Lukežič, A.; Zajc, L.Č.; Kristan, M. Deformable parts correlation filters for robust visual tracking. IEEE Trans. Cybern. 2017, 48, 1849–1861. [Google Scholar] [CrossRef] [Green Version]
  96. Danelljan, M.; Häger, G.; Khan, F.S.; Felsberg, M. Discriminative scale space tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1561–1575. [Google Scholar] [CrossRef] [Green Version]
  97. Chen, S.; Qiu, D.; Huo, Q. Siamese Networks with Discriminant Correlation Filters and Channel Attention. In Proceedings of the 2018 14th International Conference on Computational Intelligence and Security (CIS), Hangzhou, China, 16–19 November 2018; pp. 110–114. [Google Scholar]
  98. Zhu, Z.; Huang, G.; Zou, W.; Du, D.; Huang, C. Uct: Learning unified convolutional networks for real-time visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1973–1982. [Google Scholar]
  99. Song, K.; Zhang, W.; Lu, W.; Zha, Z.J.; Ji, X.; Li, Y. Visual Object Tracking via Guessing and Matching. IEEE Trans. Circuits Syst. Video Technol. 2019. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed Soft-mask with Channel and Spatial attentional Siamese (SCS-Siam) tracker. The input images and the soft mask features are fused after the pooling layer for both exemplar and search branches to enhance target information. The fused features are forwarded to an embedded Convolutional Neural Network (CNN) model ψ to compute higher level discriminative features. Attention modules are integrated into the exemplar branch using a skip connection before the correlation operation. The output of channel attention is forwarded to the spatial attention module. Finally, a response map is produced using a cross-correlation operation between the exemplar branch features and search branch features.
Figure 1. Framework of the proposed Soft-mask with Channel and Spatial attentional Siamese (SCS-Siam) tracker. The input images and the soft mask features are fused after the pooling layer for both exemplar and search branches to enhance target information. The fused features are forwarded to an embedded Convolutional Neural Network (CNN) model ψ to compute higher level discriminative features. Attention modules are integrated into the exemplar branch using a skip connection before the correlation operation. The output of channel attention is forwarded to the spatial attention module. Finally, a response map is produced using a cross-correlation operation between the exemplar branch features and search branch features.
Sensors 20 04021 g001
Figure 2. An example of a Siamese network that produces a feature space. Green and orange cubes in the cubic feature map indicate the features for green and orange regions in the image.
Figure 2. An example of a Siamese network that produces a feature space. Green and orange cubes in the cubic feature map indicate the features for green and orange regions in the image.
Sensors 20 04021 g002
Figure 3. During training. Template with soft mask generated from three videos. The first row is the original input image, the red box shows the target position. The second row is the computed soft mask with σ = 0.5. The third row is the template with the soft-mask where the target bounding box has weight 1 while the background has weight 0.5. The fourth row is the template resized and padded with the image mean to bring the target object at the center position. The fifth row is the resized soft-mask again with the target object on the center and image mean values are padded to preserve the input size.
Figure 3. During training. Template with soft mask generated from three videos. The first row is the original input image, the red box shows the target position. The second row is the computed soft mask with σ = 0.5. The third row is the template with the soft-mask where the target bounding box has weight 1 while the background has weight 0.5. The fourth row is the template resized and padded with the image mean to bring the target object at the center position. The fifth row is the resized soft-mask again with the target object on the center and image mean values are padded to preserve the input size.
Sensors 20 04021 g003
Figure 4. Applied soft-mask feature visualization. Column (a) shows the features without the soft-mask, (b) indicates features from the soft-mask images and (c) features after feature fusion from columns (a) and (b). Column (d) shows the feature fusion from the images without the applied soft-mask. The right side shows their corresponding templates with and without the soft-mask.
Figure 4. Applied soft-mask feature visualization. Column (a) shows the features without the soft-mask, (b) indicates features from the soft-mask images and (c) features after feature fusion from columns (a) and (b). Column (d) shows the feature fusion from the images without the applied soft-mask. The right side shows their corresponding templates with and without the soft-mask.
Sensors 20 04021 g004
Figure 5. Proposed channel attentional mechanism. GAP means Global Average Pooling and GMP stands for Global Maximum Pooling. ⨁ indicates pixel-wise addition operation while ⨂ indicates the pixel-wise multiplication operation.
Figure 5. Proposed channel attentional mechanism. GAP means Global Average Pooling and GMP stands for Global Maximum Pooling. ⨁ indicates pixel-wise addition operation while ⨂ indicates the pixel-wise multiplication operation.
Sensors 20 04021 g005
Figure 6. Proposed spatial attentional module. U represents the input feature map, V indicates the output feature map, and ⨂ indicates the element-wise multiplication operation. The softmax layer produces the same attention for all channels, while the later part after Sigmoid produces a different spatial attention across C channels.
Figure 6. Proposed spatial attentional module. U represents the input feature map, V indicates the output feature map, and ⨂ indicates the element-wise multiplication operation. The softmax layer produces the same attention for all channels, while the later part after Sigmoid produces a different spatial attention across C channels.
Sensors 20 04021 g006
Figure 7. Comparison over OTB2015 in terms of precision and success.
Figure 7. Comparison over OTB2015 in terms of precision and success.
Sensors 20 04021 g007
Figure 8. Comparison for eleven different challenges in terms of success over OTB2015.
Figure 8. Comparison for eleven different challenges in terms of success over OTB2015.
Sensors 20 04021 g008
Figure 9. Qualitative study of our method with state-of-the-art trackers over sequences including CarDark, Skating2-2, Basketball, Bird2, Jogging-1, and Box.
Figure 9. Qualitative study of our method with state-of-the-art trackers over sequences including CarDark, Skating2-2, Basketball, Bird2, Jogging-1, and Box.
Sensors 20 04021 g009
Figure 10. Comparison over UAV123 in terms of precision and success.
Figure 10. Comparison over UAV123 in terms of precision and success.
Sensors 20 04021 g010
Table 1. Network parameters θ of the proposed framework. Features are fused after the Max-Pool1 layer for both exemplar and search branches, as shown in Figure 1. The network Ψ ( · ) in Figure 1 contains Conv2, Conv3, Max-Pool2, Conv4, Conv5, and Conv6 layers.
Table 1. Network parameters θ of the proposed framework. Features are fused after the Max-Pool1 layer for both exemplar and search branches, as shown in Figure 1. The network Ψ ( · ) in Figure 1 contains Conv2, Conv3, Max-Pool2, Conv4, Conv5, and Conv6 layers.
LayerFilter SizeStrideIn and Out ChannelTemplate and Template with Soft-MaskSearch and Search with Soft-Mask
Input 3 3 × 135 × 135 3 × 263 × 263
Conv1 11 × 11 2 3 × 192 192 × 63 × 63 192 × 127 × 127
Max-Pool1 3 × 3 2- 192 × 31 × 31 192 × 63 × 63
Feature Fusion--- 192 × 31 × 31 192 × 63 × 63
Conv2 3 × 3 1 192 × 256 256 × 29 × 29 256 × 61 × 61
Conv3 3 × 3 1 256 × 256 256 × 27 × 27 256 × 59 × 59
Max-Pool2 3 × 3 2- 256 × 13 × 13 256 × 29 × 29
Conv4 3 × 3 1 256 × 512 512 × 11 × 11 512 × 27 × 27
Conv5 3 × 3 1 512 × 512 512 × 9 × 9 512 × 25 × 25
Conv6 3 × 3 1 512 × 384 384 × 7 × 7 384 × 23 × 23
Table 2. Details of different benchmarks.
Table 2. Details of different benchmarks.
BenchmarksOTB2015TC-128UAV123VOT2016VOT2017
Sequences1001281236060
Minimum frames71711094841
Mean frames590429915357356
Max frames38723872308515001500
Total frames59,04055,3461,12,57821,45521,356
Table 3. Comparison of the proposed algorithm with various state-of-the-art methods over OTB2015 using precision, success, and speed in Frames Per Second (FPS). Bold values represent the maximum value for precision, success and FPS.
Table 3. Comparison of the proposed algorithm with various state-of-the-art methods over OTB2015 using precision, success, and speed in Frames Per Second (FPS). Bold values represent the maximum value for precision, success and FPS.
TrackersPrecisionSuccessFPSReal-Time
RNN [76]79.860.6--
MLT [77]-61.148Yes
P2FNet [78]79.859.552Yes
SiamFc-lu [79]-62.082Yes
DSiamM [57]-60.518No
SiamTri [31]78.159.085Yes
Li’s [80]77.158.230Yes
CNNSI [75]72.152.21<No
Kuai et al. [81]82.262.225No
TRACA [71]81.460.2101Yes
SINT [82]-59.24No
ACFN [93]80.257.515No
CSRDCF [68]73.358.713No
Staple [74]78.457.880Yes
SRDCF [72]78.959.86No
HASiam [83]81.061.130Yes
UDT [73]76.059.470Yes
SiameseFC-G&M [99]81.061.368Yes
CMKCF [87]82.261.074.4Yes
CFNet [34]77.758.643Yes
SiameseFC [30]77.158.286Yes
SCA-Siam (Ours)84.062.276Yes
SSA-Siam (Ours)82.161.375Yes
SCS-Siam (Ours)82.761.873Yes
Table 4. Comparison for eleven different challenges in terms of precision over OTB2015. Bold values represent the maximum value for precision.
Table 4. Comparison for eleven different challenges in terms of precision over OTB2015. Bold values represent the maximum value for precision.
TrackersSiamTriCFNetSRDCFTRACASiameseFCUDTStapleCNNSISCA-SiamSCA-SiamSCS-Siam
FM77.677.477.376.275.875.372.967.580.278.379.2
BC71.573.177.579.969.074.974.968.780.274.677.5
MB74.476.178.277.172.471.471.969.678.277.979.2
Def68.066.973.476.969.067.075.168.780.077.679.3
IV75.176.378.783.374.070.078.260.083.681.283.3
IPR75.978.573.779.472.874.175.168.881.180.580.5
LR88.486.163.173.181.568.859.166.091.983.082.5
OCC72.671.373.577.572.270.672.864.478.577.177.3
OPR76.175.874.482.875.474.773.768.980.679.880.3
OV72.365.059.770.066.965.166.859.473.065.670.3
SV75.274.874.976.973.971.473.168.781.679.279.6
Table 5. Comparison of the proposed method with various state-of-the-art methods over TC128 using precision, success, and speed in FPS. Bold value represent the maximum value for precision, success and FPS.
Table 5. Comparison of the proposed method with various state-of-the-art methods over TC128 using precision, success, and speed in FPS. Bold value represent the maximum value for precision, success and FPS.
TrackersPrecisionSuccessFPS
SCT [94]62.746.640
KCF [7]54.938.7160
CNNSI [75]63.845.6<1
UDT [73]65.850.770
CFNet [34]60.745.643
ACT [89]73.853.230
SiameseFC [30]68.850.386
SCA-Siam (Ours)73.153.276
SSA-Siam (Ours)73.854.275
SCS-Siam (Ours)74.253.873
Table 6. Performance comparison for different trackers over VOT2016. Bold values represent the maximum value for overlap, robustness, and EAO.
Table 6. Performance comparison for different trackers over VOT2016. Bold values represent the maximum value for overlap, robustness, and EAO.
TrackersOverlap (↑)Robustness (↓)EAO (↑)
MemTrack [58]0.531.440.27
MemDTC [90]0.511.820.27
ECO [21]0.54-0.37
Staple [74]0.530.380.29
SRDCF [72]0.540.420.25
DSiam [57]0.492.930.18
CCOT [20]0.540.240.33
UDT [73]0.54-0.22
SiameseFC [30]0.530.460.23
CMKCF [87]0.530.180.30
SCA-Siam (Ours)0.550.230.28
SSA-Siam (Ours)0.550.230.27
SCS-Siam (Ours)0.550.210.28
Table 7. Performance comparison for different trackers over VOT2017. Bold values represent the maximum value for overlap, robustness, EAO, and FPS.
Table 7. Performance comparison for different trackers over VOT2017. Bold values represent the maximum value for overlap, robustness, EAO, and FPS.
TrackersOverlap (↑)Robustness (↓)EAO (↑)FPS
CSRDCF [68]0.490.490.2513
MemTrack [58]0.491.770.2450
MemDTC [90]0.491.770.2540
SRDCF [72]0.490.970.126
DSST [96]0.391.450.0824
SATIN [88]0.491.340.2824
SiamDCF [97]0.500.470.2560
UCT [98]0.490.480.2041
SiameseFC [30]0.500.590.1986
GradNet [91]0.500.370.2480
SiameseRPN [92]0.490.460.24200
SCA-Siam (Ours)0.510.360.2176
SSA-Siam (Ours)0.520.380.2075
SCS-Siam (Ours)0.520.290.2473
Table 8. Performance of the proposed SCS-Siam over the OTB2015 dataset using different values of σ to generate soft mask images during training as shown (see Section 3.2). ¶ denotes that σ is set at 0.9 during training as well as testing. However, σ is set at 1 for the test time for the rest of the experiments.
Table 8. Performance of the proposed SCS-Siam over the OTB2015 dataset using different values of σ to generate soft mask images during training as shown (see Section 3.2). ¶ denotes that σ is set at 0.9 during training as well as testing. However, σ is set at 1 for the test time for the rest of the experiments.
SCS-Siam- σ PrecisionSuccess
SCS-Siam-0.076.356.5
SCS-Siam-0.378.258.6
SCS-Siam-0.581.560.4
SCS-Siam-0.780.059.6
SCS-Siam-0.982.761.8
SCS-Siam-1.079.759.5
SCS-Siam- 0.9 81.760.5
Table 9. Comparison of different versions of the proposed framework on the OTB2015 dataset. SiameseFC is the baseline network, SiameseFC* shows that SiameseFC was retrained on GOT-10k dataset, Extended-SiamFC means the proposed framework without a soft mask and attentions, SSA-Siam is Extended-SiamFC with a soft mask and spatial attention, SCA-Siam is Extended-SiamFC with a soft mask and channel attention, SSC-Siam is Extended-SiamFC with a soft mask and spatial attention followed by channel attention, and SCS-Siam is the proposed framework with soft mask, channel attention followed by spatial attention. B-SCS-Siam means that the channel attention was followed by spatial attention at both exemplar and search branches.
Table 9. Comparison of different versions of the proposed framework on the OTB2015 dataset. SiameseFC is the baseline network, SiameseFC* shows that SiameseFC was retrained on GOT-10k dataset, Extended-SiamFC means the proposed framework without a soft mask and attentions, SSA-Siam is Extended-SiamFC with a soft mask and spatial attention, SCA-Siam is Extended-SiamFC with a soft mask and channel attention, SSC-Siam is Extended-SiamFC with a soft mask and spatial attention followed by channel attention, and SCS-Siam is the proposed framework with soft mask, channel attention followed by spatial attention. B-SCS-Siam means that the channel attention was followed by spatial attention at both exemplar and search branches.
TrackerPrecisionSuccessFPS
SiameseFC77.158.286
SiameseFC*79.259.786
Extended-SiamFC80.360.279
SSA-Siam82.161.375
SCA-Siam84.062.276
SSC-Siam82.060.973
SCS-Siam82.761.873
B-SCS-Siam81.560.459
Table 10. Ablation study of channel and spatial attention modules. SCA-GMP is the SCA-Siam using only global max pooling sub-network while SCA-GAP represents the SCA-Siam with only the global average pooling sub-network. Similarly, SSA-Context means the SSA-Siam with the context modeling sub-network, while SSA-Transform represents the SSA-Siam with the transform sub-network alone.
Table 10. Ablation study of channel and spatial attention modules. SCA-GMP is the SCA-Siam using only global max pooling sub-network while SCA-GAP represents the SCA-Siam with only the global average pooling sub-network. Similarly, SSA-Context means the SSA-Siam with the context modeling sub-network, while SSA-Transform represents the SSA-Siam with the transform sub-network alone.
TrackerPrecisionSuccess
SCA-GMP83.561.8
SCA-GAP83.962.0
SCA-Siam84.062.2
SSA-Context81.560.8
SSA-Transform81.861.0
SSA-Siam82.161.3

Share and Cite

MDPI and ACS Style

Fiaz, M.; Mahmood, A.; Jung, S.K. Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking. Sensors 2020, 20, 4021. https://doi.org/10.3390/s20144021

AMA Style

Fiaz M, Mahmood A, Jung SK. Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking. Sensors. 2020; 20(14):4021. https://doi.org/10.3390/s20144021

Chicago/Turabian Style

Fiaz, Mustansar, Arif Mahmood, and Soon Ki Jung. 2020. "Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking" Sensors 20, no. 14: 4021. https://doi.org/10.3390/s20144021

APA Style

Fiaz, M., Mahmood, A., & Jung, S. K. (2020). Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking. Sensors, 20(14), 4021. https://doi.org/10.3390/s20144021

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop