Next Article in Journal
A PI Control Method with HGSO Parameter Regulator for Trajectory Planning of 9-DOF Redundant Manipulator
Next Article in Special Issue
Joint Cross-Consistency Learning and Multi-Feature Fusion for Person Re-Identification
Previous Article in Journal
Trajectory Design for Multi-UAV-Aided Wireless Power Transfer toward Future Wireless Systems
Previous Article in Special Issue
Scale-Space Feature Recalibration Network for Single Image Deraining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single-Shot Object Detection via Feature Enhancement and Channel Attention

College of Mathematics and Computer Science, Zhejiang Normal University, Jinhua 321004, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 6857; https://doi.org/10.3390/s22186857
Submission received: 9 August 2022 / Revised: 30 August 2022 / Accepted: 4 September 2022 / Published: 10 September 2022

Abstract

:
Features play a critical role in computer vision tasks. Deep learning methods have resulted in significant breakthroughs in the field of object detection, but it is still an extremely challenging obstacle when an object is very small. In this work, we propose a feature-enhancement- and channel-attention-guided single-shot detector called the FCSSD with four modules to improve object detection performance. Specifically, inspired by the structure of atrous convolution, we built an efficient feature-extraction module (EFM) in order to explore contextual information along the spatial dimension, and then pyramidal aggregation module (PAM) is presented to explore the semantic features of deep layers, thus reducing the semantic gap between multi-scale features. Furthermore, we construct an effective feature pyramid refinement fusion (FPRF) to refine the multi-scale features and create benefits for richer object knowledge. Finally, an attention-guided module (AGM) is developed to balance the channel weights and optimize the final integrated features on each level; this alleviates the aliasing effects of the FPN with negligible computational costs. The FCSSD exploits richer information of shallow layers and higher layers by using our designed modules, thus accomplishing excellent detection performance for multi-scale object detection and reaching a better tradeoff between accuracy and inference time. Experiments on PASCAL VOC and MS COCO datasets were conducted to evaluate the performance, showing that our FCSSD achieves competitive detection performance compared with existing mainstream object detection methods.

1. Introduction

Object detection, as a basic, fundamental, and challenging task in computer vision, aims at the detection various visual instances in the real world. With the rapid development of deep neural networks, object detection has been significantly improved in comparison with traditional methods. As the cornerstone of scene understanding and image learning, object detection builds a solid foundation for dealing with other vision tasks, including object tracking [1], image captions [2], segmentation [3], and so on. The emergence of deep learning [4] has provided models with a strong ability to learn sophisticated and complicated representations, leading to remarkable progress in object detection. Generally speaking, there are two main categories of object detectors with deep learning; the first is that of two-stage detectors, which have a higher accuracy but slower inference speed, such as R-CNN [5], SPPNet [6], Fast R-CNN [7], Faster R-CNN [8], Mask R-CNN [9], and R-FCN [10]. The other category is that of one-stage detectors, such as YOLO v1 [11], v2 [12], and its improved versions [13,14], SSD [15], FSSD [16], DSSD [17], and RetinaNet [18], which run faster but with lower accuracy.
It is notable that the quality of features defines the upper limits of a model’s performance, which has been studied by researchers in recent years. The method of using low-level features combined with high-level features is effective for collecting rich information in order to get better detection performance. The feature pyramid network (FPN) [19] was the initial work of enhancing the feature representation by combining multi-level features by building pyramids of different scales. Specifically, shallow feature maps focus on the detection of larger instances, while deeper feature layers are used for smaller ones. Moreover, the design of lateral connections between the bottom-up and top-down layers was introduced in order to deliver the multi-scale details of various objects. Nevertheless, there remain some serious problems in FPNs: The first is information decay [20] during fusion; FPN-based methods adopt 1x1 convolution layers to cause the channel to have the same numbers between the input and the output, but they lose channel information within these operations. For the purpose of a lower computational cost and memory burden, a common practice is the reduction of the number of channels of the feature map, resulting in serious information loss.The second problem is that of aliasing effects [19] during cross-scale fusion. However, there are semantic and contextual differences among feature maps that are to be integrated together. Therefore, simply using element-add or c o n c a t fusion will lead to serious aliasing effects. Although some previous works [20,21] have ameliorated these issues to a certain degree, there is still much room for further improvement.
In this work, we propose FCSSD in order to enhance the feature representation and alleviate the aliasing affects that originate from FPN feature fusion in object detection. Firstly, inspired by atrous convolution [22], which is expected to capture a larger convolutional receptive size for more information in image segmentation tasks, we present an efficient feature-extraction module (EFM) for exploring the contextual information of a target in the process of a forward pass. The EFM has negligible parameters, has been proven to be effective in strengthening the representation ability of deep features, and contributes to fast and accurate detectors. As we all know, semantic information is vital for the detection of small objects, so a series of downsampling will cause useful details of instances to be lost, as the receptive field is not able to catch different dependencies in the scene. We provide a pyramidal aggregation module (PAM), which includes four levels of kernels with mixed sizes. The PAM improves the ability to process the input features of multi-scale feature maps for capturing rich semantic knowledge about smaller objects. Traditional FPN design has intrinsically restricted the model performance by using a single information flow because of the single lateral connection. In addition, the utilization of element summation is not an effective way to combine features, as it leads to a lack of refinement of the fused features. It is known that low-level features are beneficial for small object identification and detection; however, there is a long pathway for low-level features to reach deeper ones in order to get enough semantic information, thereby increasing the difficulty of accessing accurate localization details. We designed a feature pyramid refinement fusion (FPRF) for generating the refined features for a multi-scale feature map in order to make the most of the shallower and deeper features. Fusing multi-scale feature maps would degrade the power of the model’s representation due to the inconsistent information flow. Furthermore, it would also bring in loss of detail in the information in the highest pyramid due to its the limited channel numbers. Motivated by these innovations, we designed an attention-guided module (AGM), which is a channel attention aggregation module. The AGM aims to optimize the final integrated layers on each level and alleviates the aliasing effects. To validate the performance of our proposed model, we tested our FCSSD on public datasets—PASCAL VOC 2007 and MS COCO 2017. The main contributions of this work can be summarized as follows:
  • For feature enhancement, we propose a lightweight efficient feature-extraction module (EFM) and pyramidal aggregation module (PAM). The EFM applies various dilation rates, batch normalization (BN), and ReLU to more richly explore the contextual information of the CNN. The PAM uses different adaptive average pooling sizes to exploit richer semantic information from the deep layers, and upsampling is embraced to keep the same feature size as that of the original input feature map.
  • Aiming to make full multi-scale features, we built an effective feature pyramid refinement fusion (FPRF) to calibrate the multi-scale features during the fusion process. The FPRF broadens the ways of the single lateral connections of the traditional FPN and enriches the approaches of multi-scale feature fusion, thus greatly improving the detection performance.
  • To alleviate the aliasing effects of the FPN, we introduce an attention-guided module (AGM); an improved channel attention mechanism was developed to ameliorate the problem of fused features, and it is efficient and speeds up the training process.
  • By applying the above four improvements, we designed a feature-enhancement and channel-attention-guided single-shot detector (FCSSD). Experiments on the PASCAL VOC2007 and MS COCO2017 datasets showed the effectiveness of our proposed FCSSD and that it can outperform mainstream object detectors.

2. Related Work

2.1. Deep-Learning-Based Object Detectors

Deep-learning-based object detection can be divided into the categories two-stage detectors and one-stage detectors. R-CNN [5] was first to show that the use of a convolutional neural network (CNN) could enable one to reach unprecedented object detection performance on the PASCAL VOC [23] dataset. R-CNN firstly produces fixe-length features from generated proposals and then uses linear SVMs and a bounding box regressor to classify objects for a precise bounding-box prediction. Although R-CNN represents a breakthrough in object detection, it is a multi-stage pipeline and has a higher computational cost, and it is slow in its running time because each part of the whole network must be trained separately. SPPNet [6] introduced a spatial pyramid pool (SPP) module with the aim of obtaining a fixed length for fully connected layers. In an evaluation, the SPP was shown to be faster than the R-CNN to some extent, but still could not achieve real-time object detection. Fast R-CNN [7] makes the combination of classification and regression, achieving a fast end-to-end training paradigm. Faster R-CNN [8] involved the development of a region proposal network, requiring nearly cost-free region proposals and boosting the performance significantly. Cascade R-CNN [24] was used to study the effect of regression proposals for under different IoU thresholds based on faster R-CNN and achieved better performance in object detection. In addition to these, there are still many other excellent two-stage detectors, such as Mask R-CNN [9], R-FCN [10], and CBNet [25].
One-stage detectors usually apply a unified network to directly accomplish localization and classification with more efficiency but lower accuracy. YOLO [11] frames the detection task as a regression problem and divides the input image into some grids; thus, extracting features from the input image results in the direct prediction of the bounding box and classification from each unit of the grid within a united architecture. For the sake of achieving real-time detection without giving away much accuracy, SSD [15] combines the idea of the RPN in Faster R-CNN, YOLO, and multi-scale convolutional features to achieve fast detection. RetinaNet [18] introduces a new balanced loss function called focal loss to downweight the contributions of an easy sample v . s . a hard example, thus improving the detection performance of one-stage detectors. There are also other one-stage detectors, including the YOLO series [12,13,14], FSSD [16], and DSSD [17].

2.2. Enhancement of Feature Representation

Deep learning techniques have a powerful ability to learn rich feature representations with multi-level features directly from raw images. The extraction of effective features is a critical issue for more accurate classification and localization. Earlier object detectors usually directly performed predictions based on the pyramid feature hierarchy extracted from a backbone network [15]. In a pioneering work, the effective extraction of the low-level features of a network was determined to be a key problem in object detection. The feature pyramid network (FPN) [19] was the first to enhance a CNN’s representation by fusing features from different levels and constructing feature pyramids. The FPN proposes a top-down pathway and uses lateral connections to combine multi-level features. The design of the FPN shows that fully utilizing the multi-scale features can promote accuracy in object detection. PFPNet [26] was used to investigate the width of the FPN by means of building feature pyramid blocks to widen the network and further improve the performance. PANet [27] was used to explore an extra bottom-up pathway to improve the low-level information in deep layers. The proposal of TridentNet [28] was the first to use the effects of the receptive field in object detection with different scale sizes, and it was constructed with a parallel multi-branch architecture in which there were three branches for training, but only one of them was used for testing, which ensured that no additional parameters or computations were added during forward inference. EfficientDet [21] used a weighted bi-directional FPN to perform easy and fast feature fusion. RefineDet [29] used an anchor-refined module and object detection module to get better features after refinement and fusion, achieving excellent accuracy and high efficiency.

2.3. Attention Mechanisms

Attention mechanisms imitate human cognitive awareness about specific information, amplifying critical details to focus more on the essential aspects of data. In addition, attention mechanisms are able to build long-range dependencies within a model and become the workhorse of many challenging tasks, including image classification [30], semantic and instance segmentation [31], and natural language processing [32]. Attention mechanisms have a wide range of applications in object detection because they help the model to better locate and recognize objects in images, thus further improving detection performance [33,34]. SENet [35] involved a s q u e e z e a n d e x c i t a t i o n (SE) block with the aim of collecting global information along with channel-wise relationships and strengthening the representation ability of a CNN with efficient operations. ECANet [36] involved an e f f i c i e n t c h a n n e l a t t e n t i o n (ECA) block, a local cross-channel interaction strategy without dimensionality reduction implemented through one-dimensional convolution, which improved the training speed with a lower model complexity. GENet [37] mainly included two operations, Gather and Excite. The former efficiently aggregated feature responses over a large spatial area, and the latter redistributed the combined information to local features in a spatial domain. SKNet [38] involved s e l e c t i v e k e r n e l (SK) convolution, which enabled the network to adjust the kernel size of the convolution operation according to the input, achieving significant performance gains at a small computational cost. CBAM [39] was used to study the effects of model performance in the spatial dimension and channel dimension. For CBAM, two parallel branches with max pooling and average pooling were designed to explore the interrelations of features between channels, and a concatenation operation was employed to gather the final outputs. CABM told the network where to focus and where to pay attention by modeling the spatial and channel dimensions of the features.
Based on the methods listed above, we focused on building an effective method for exploring multi-level feature fusion in the FPN; thus, we propose the FCSSD. The FCSSD consists of four modules—the EFM, PAM, FPRF, and AGM—based on SSD. The details of FCSSD are introduced in the following sections.

3. Methodology

In this section, each part of the proposed FCSSD, which is shown in Figure 1, is demonstrated in detail. First, we introduce the specific design of our network in Section 3.1. Then, the efficient feature-extraction module (EFM) is presented in Section 3.2. Next, the pyramidal aggregation module (PAM) is shown in Section 3.3. Section 3.4 describes the feature pyramid refinement fusion (FPRF). Finally, in Section 3.5, the attention-guided module (AGM) is explained.

3.1. FCSSD Architecture

In this work, we adopt a one-stage SSD for object detection as our baseline due to its good tradeoff between high speed and detection accuracy. The SSD predicts layers to match the output, including both the bounding box regression and classification. Its object detection results are from six final feature maps; the adjusted VGG-16 is used as the backbone, and each feature map predicts different scales of an individual object. NMS (non-maximum suppression) is usually adopted to filter redundant and overlapping predictions to generate the final detection result.
Figure 1 shows the overall architecture of our proposed method, consisting of four parts: the standard SSD network; feature enhancement fusion component: the efficient feature-extraction module (EFM), pyramidal aggregation module (PAM), feature pyramid refinement fusion (FPRF) process X, and attention-guided module (AGM). As we mentioned before, the standard SSD employs VGG-16 as the network backbone. We then regard the six layers of the SSD as the prediction feature maps. The strides are { 8 , 16 , 32 , 64 , 100 , 300 } pixels of multiple feature maps with respect to the input image size of 300 × 300 . For brevity, we refer to these as C 1 , C 2 , C 3 , C 4 , C 5 , and C 6 . The EFM exploits rich contextual information from receptive fields with different sizes, and the PAM collects prior global-scene-level semantic knowledge from the CNN. An effective FPRF is applied to the multi-scale features to generate a refined map X. The AGM is a channel-attention-guided module that reduces the aliasing effects for fused feature maps and accelerates the training speed. We will discuss the above modules and fusion process in the following subsections.

3.2. Efficient Feature-Extraction Module

As we know, object detection requires contextual information, especially for small objects [40]. The efficient feature-extraction module (EFM) was inspired by the mode in which humans distinguish objects by relying on different sizes, colors, backgrounds, and shapes. For example, it is difficult for a human to distinguish a bird very high in the sky, but it is easy to recognize when the sky is taken into consideration as the contextual information. Therefore, we believe that contextual knowledge is helpful for object detection. The EFM takes advantage of dilated convolution [22] to exploit larger receptive fields on the feature map, thus acquiring a considerable contextual information about the detected objects. To improve the accuracy of the multi-scale object detection, we apply dilated convolution with three kinds of dilation rates to obtain the contextual information from the receptive fields. As Figure 2 shows, the EFM has a multi-branch convolutional design, and it has two main components: a multi-dilated convolution layer and a feature aggregation layer.
For the input features R C × H × W , the multi-dilated convolution layer has parallel branches consisting of dilated convolution, a BN layer, and the ReLU activation function. Three different dilation rates are used, but with the same kernel size. Specifically, the kernel of the dilated convolution is c o n v 3 × 3 , and the dilation rates are 1, 3, and 5 for the different branches. We also use the same padding to keep the input and output feature maps at the same size. We adopt a scale ratio r to reduce the channels of the feature map for efficiency. This process can be expressed as:
F o u t = C o n v d ( 1 , 3 , 5 ) R C r × H × W
where R C r × H × W is the input feature map with a reduced channel ratio; C, H, and W refer to the channel, height, and width of the feature map; r equals 8 as the hyper-parameter; C o n v d ( 1 , 3 , 5 ) indicates c o n v 3 × 3 with three different dilation rates of 1, 3, and 5; F o u t refers to the output feature map.
The feature aggregation layer is used to fuse contextual information from different parallel branches and make the most of the diverse feature details. Element concatenation and summation are adopted to produce the multi-branch feature representation. It is notable that stacks of c o n v 1 × 1 and c o n v 3 × 3 are used to reduce the channels of the concatenation feature map. As residual learning, the input feature map is added to that. The whole process can be expressed as:
F o u t = C o n v ( C o n c a t [ C o n v d ( 1 , 3 , 5 ) R C r × H × W ] + R C × H × W )

3.3. Pyramidal Aggregation Module

Deep convolutional neural network (DCNN) methods have strengthened object detection performance by a great margin, but still face tough challenges when considering different scenes, object sizes, and backgrounds. Semantic relationships are universal and important for object detection, especially when the target is very small [21,27]. Prior works proved that not collecting the necessary semantic information increases the possibility of misclassification and mislocalization. Overlooking the global scene information may result in failure to consider an object’s details, such as its pattern, texture, and shape. Therefore, one should focus on separate subregions that contain inconspicuous information of interest. In order to learn more suitable prior global-scene-level knowledge for the CNN, we built a pyramidal aggregation module (PAM). This was aimed at the aggregation of global semantic information along with subregions, as well as at the reduction of the training loss between different subregions.
Global average pooling is commonly used in image classification tasks by enforcing correspondences between feature maps and categories. The PAM fuses features with four different pyramid scales. As shown in Figure 3, a stack of pyramid levels with four different AVG sizes is used to separate the input features and form pooled representations for locations; it is followed by CONV, BN, and ReLU. The input features first go through the average pooling pyramid to generate the aggregation information. Next, all of the aggregated features are upsampled through bilinear interpolation, followed by the concatenation of the original input feature to produce rich semantic information. Finally, a series of conv operations are used to balance the channel number. The whole process can be formulated as:
F o u t = C o n v ( C o n c a t [ U p s a m p l e ( A v g k = 1 , 2 , 3 , 5 ( F i n ) ) , F i n ] )
where F i n and F o u t refer to the input and output feature maps, respectively. C o n v is C o n v 3 × 3 operations, and C o n c a t is the concatenation for aggregating all of the feature maps. U p s a m l e indicates an upsampling operation for the generated output. A v g is adaptive average pooling with different sizes of 1, 2, 3, and 5.

3.4. Feature Pyramid Refinement Fusion

Traditional FPN models are inefficient in exploring stacks of feature maps that contain a wide range of scales, giving rise to inferior detection performance. The SSD [15] introduced the design of a pyramidal feature hierarchy for detecting multi-scale objects, as shown in Figure 4b. However, it failed to exploit the relationships of the information between shallower layers and deeper layers and just used a single-stage feature map while neglecting the complementary effects of all layers, which made it hard to achieve good performance for small object detection. To alleviate this problem, an intuitive notion is that of acquiring multi-scale features, such as with an FPN [19], as shown in Figure 4c. However, the information flow in the FPN becomes another critical problem, as the lateral connections and top-down pathways greatly limit the upper bound of feature utilization. To improve the model’s efficiency and enrich the single lateral connections of the traditional FPN, we propose feature pyramid refinement fusion. Different input resolutions contribute unequally to the final detection results because they carry different characteristic features of the object. If we gather them together into a refined feature, then they will contain different information at different scales. Finally, we just apply the refined feature map to previous features through multiplication to get the final output for detection. This process can be formulated as:
P 4 o u t = C o n v ( C 4 i n ) P 3 o u t = C o n v ( C 3 i n + U p s a m p l e ( P 4 o u t ) ) P 2 o u t = C o n v ( C 2 i n + U p s a m p l e ( P 3 o u t ) ) P 1 o u t = C o n v ( C 1 i n + U p s a m p l e ( P 2 o u t ) )
R 4 = R e s i z e ( P 4 o u t , S ) R 3 = R e s i z e ( P 3 o u t , S ) R 2 = A M P ( P 2 o u t , S ) R 1 = A M P ( P 1 o u t , S ) X = ( R 1 + R 2 + R 3 + R 4 ) 4
M i = R i × X , i { 1 , 2 , 3 , 4 } M j = P j , j { 5 , 6 }
where AMP refers to adaptive max pooling, S is the scale size of feature map C 2 , X represents the refined features, and M refers to the final feature maps for prediction.

3.5. Attention-Guided Module

The multi-scale features are used to improve small object detection to some extent. Nevertheless, serious semantic differences exist among features of multiple shapes, especially in the fusion process. Miscellaneous integrated information may lead to aliasing effects [19,42], causing the mistakes in localization and classification. In the original FPN, a series of 1 × 1 and 3 × 3 convolutions is usually followed by each fused feature map to reduce these effects. Features generated from the VGG-16 backbone contain more serious aliasing effects after the EFM and PAM. In order to mitigate the negative influences of the aliasing effects, it is appropriate to employ an attention mechanism on the fused feature maps. However, just applying attention modules on integrated features would bring in a vast computational burden because the SSD takes on six feature maps. We hope that the attention mechanism will not only alleviate this aliasing effect, but also that it will require fewer parameters. Inspired by the channel attention design of CBAM [39], we developed an attention-guided module (AGM), as illustrated in Figure 5, to solve the inconsistency between layers and mitigate the aliasing effects between multi-scale feature maps. Two different types of spatial contextual information are generated by employing the two main pooling paradigms (AMP and AAP). These two kinds of information descriptors independently head to the FC layers. Element-wise summation and a sigmoid activation function are used to get the final output. The whole process can be expressed as:
A F G = σ ( F C ( G A P ( F i n ) ) + F C ( G M P ( F i n ) ) ) M i = A F G ( X ) × P i
where i refers to the index of the pyramid levels in the backbone, P i refers to fused features, M i refers to the final outputs, F i n and A F G are the input and output, respectively, σ is a sigmoid activation function, F C is the fully connected layer, and G A P and G M P refer to global average pooling and global max pooling, respectively.

4. Results

4.1. Dataset and Experimental Details

We used the PASCAL VOC 2007 [23] and MS COCO 2017 [40] datasets to validate the model proposed in this work. Training was performed on the VOC 2007 trainval data with 20 classes, which were in combination with 5k images and the VOC 2012 trainval dataset. The evaluation was performed on the VOC 2007 test set with 5k images. Here, we used the mAP as the evaluation criterion for this dataset. The input was 300 × 300 , the batch size was set to 16 with a total of 250 epochs during training on a single 2080Ti GPU, and the learning rate was 4 × 10 3 at the beginning. The warming-up training strategy was used to adjust the learning rate, which gradually increased the learning rate from 6 × 10 1 to 4 × 10 3 in the first five epochs. It dropped the original point and was divided by 10 at 150, 200, and 230 epochs. The weight decay was set to 1 × 10 5 , and the momentum was set to 9 × 10 1 . For the input size of 512 × 512 , the total of the training epochs was up to 200, the batch size was decreased to 8, and the other settings were kept unchanged.
Our FCSSD was trained based on the MS COCO 2017 dataset, which contains 80 classes, 115k images for training ( t r a i n 2017 ) , 5k images for validation ( v a l 2017 ) , and 20k images for testing ( t e s t d e v ) . We set the total number of training epochs to 150, and also adopted the warming-up technique to increasingly raise the learning rate from 6 × 10 1 to 4 × 10 3 in the first five epochs and then decreasing it after 60 and 100 epochs by a factor of 10, finishing up at 140 epochs. With these experimental settings, we used a batch size of 8, and the weight decay, momentum, and other settings were the same as those with the PASCAL VOC training strategy. Table 1 shows detailed information on the two datasets. Figure 6 provides a detailed workflow of the training and detection process for the improved model used in this paper.

4.2. Evaluation Metric

To evaluate the model’s performance in object detection, the precision, recall, and mean average precision (mAP) were used. They can be expressed as:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where T P indicates true positives, F N indicates false negatives, and F P indicates false positives.
The intersection over union (IoU) was used to set the thresholds of the ground-truth box and the prediction box to determine the truth.
I o U = a r e a ( B p r e d B g t ) a r e a ( B p r e d B g t )
where B p r e d indicates the predicted bounding box, and B g t is the ground-truth bounding box. The IoU threshold in PASCAL VOC was set to 0.5, and the mAP of the MS COCO dataset ranged from 0.5 to 0.9 with a step size of 0.05.
m A P = m A P 0.50 + m A P 0.55 + m A P 0.60 + . . . + m A P 0.90 + m A P 0.95 10
In addition to the basic evaluation of the mAP, the average precision (AP) was also included. A P , A P 50 , A P 75 , A P s m a l l , A P m e d i u m , and A P l a r g e depended on different thresholds from those of the criteria for MS COCO.

4.3. Experimental Analysis

4.3.1. PASCAL VOC 2007

We compared our method with the mainstream one-stage and two-stage object detectors, as shown in Table 2. Obviously, in order to achieve better performance, most existing two-stage detectors usually have a larger input size (typically 1000 × 800 ). CoupleNet [44] achieved an mAP of 82.7, which was 1.2 higher than the mAP of our method with a 300 × 300 input. For the one-stage methods, we used two input variants for a fair comparison: 300 × 300 and 512 × 512 scales. The baseline SSD [15] achieved detection with an mAP of 77.2 with a 300 × 300 input. Our approach provided a considerable increase of 4.3% in terms of mAP in comparison with the baseline SSD. Our FCSSD surpassed the DSSD [17], which used a strong feature extraction backbone, ResNet-101, by a large margin of 2.9%. In comparison with the complicated FPN-style methods, such as PFPNet [26] and RefineDet [29], our method achieved a slightly higher mAP than they did, and this highlighted the effectiveness of our proposed FCSSD. With the input size of 512 × 512 , RefineDet [29], RFBNet [45], and PFPNet [26] achieved accuracies of 81.8, 82.2, and 82.3, respectively. With the same input size and backbone, our method outperformed them with an accuracy of 83.2 mAP. Table 2 shows the results of our FCSSD versus those of mainstream object detectors on this dataset. The FPS evaluation metric was used to examine the model’s inference speed, and our FCSSD also achieved a better tradeoff between accuracy and speed.

4.3.2. MS COCO

Table 3 shows the results on the MS COCO dataset. For a 300 × 300 input, the baseline SSD [15] achieved a detection performance score of 25.1. With the same feature extraction backbone, our method achieved a notable refinement of 13.1% for the overall detection score in comparison with the baseline SSD. For large objects ( A P l ) , the baseline SSD attained a pure performance of 41.4 AP. However, it seriously deteriorated to 6.6 AP for small objects ( A P s ) , as it showed inferior quality in small object detection. Significantly, our FCSSD achieved a detection performance of 15.3 AP for this criterion, which was over double compared with the SSD baseline. Likewise, we also accomplished a great margin of improvement in detection performance with medium objects ( A P m ) . Among the existing single-stage detectors, RFBNet [45] and RefineDet320+ [29] provided scores for overall detection of 30.3 and 35.2, respectively. With the same VGG-16 backbone, our detector achieved results superior to those obtained with both methods.
For the 512 × 512 input size, the baseline SSD achieved an overall detection score of 28.8. Our method provided a significant gain of 12.7% in terms of AP with the same VGG backbone. Among the existing methods, EfficientDet-D1 [21] and YOLOv4 [14] provided detection AP scores of 39.6 and 41.2, respectively, due to their efficient framework design. Compared with the anchor-free-style method CornerNet [54], our method showed a slightly better performance in terms of AP and outperformed it in terms of inference time (65 versus 227 ms), which can be expensive in terms of costs when processing an image. The low quality of feature representations of small objects was due to their limited size and the general feature extraction, which is still a tough and challenging problem in both types of detectors. The experiments demonstrated that the feature representation power could be strengthened through our FCSSD network design. In the MS COCO dataset, more than 70% of the images are composed of objects, the size of which is usually less than 32 × 32 pixels in one image. In addition, MS COCO contains more detailed information on the objects than the images in PASCAL VOC do, and this can further enhance the learning ability of our method.
The two-stage detectors were able to achieve superior accuracy, but with a high computational cost. They usually require a considerably larger input resolution and need more than 100 ms to process an image during the inference time. For example, Mask-RCNN [9] achieved an AP of 39.8, but needed 210 ms to infer an image. Our method provided competitive accuracy with a high efficiency and achieved excellent performance, with a detection AP score of 41.5 and an inference time of 65 ms. This shows that our FCSSD not only achieves better performance, but also operates at high efficiency.

4.3.3. Ablation Study

To verify the effectiveness of the proposed modules in our FCSSD, we conducted various ablation experiments on the PASCAL VOC 2007 test set and MS COCO m i n i v a l . To be specific, we used VGG-16 as the backbone and an input size of 300 × 300 in all experiments. As shown in Table 4, when we used only the EFM, the mAP was increased by 1.4% in comparison with that of the baseline SSD (77.2%) [15]. Based on the EFM, we found that the mAP was further increased from 78.6% to 79.5% with the embedding of the PAM. The FPRF refined multi-scale features from the shallow layers and deep layers and showed a more effective feature fusion method, which boosted the performance by 2% in comparison with that of the original SSD. Since the AGM plays an important part in balancing the channel weights of features and reduces the aliasing effects caused by the FPN design, it boosted the performance of the model by 1% based on the FPRF, from 79.2% to 80.2%. Finally, the FCSSD reached an mAP of 81.5% on the PASCAL VOC2007 test set with the image size of 300 × 300 . As shown in Table 5, the AP of the baseline SSD gradually increased as we added the EFM and PAM to the model. When the FPRF was applied to the model, the AP underwent a large increase, showing that multi-scale feature pyramids are beneficial for object detection and provide richer semantic and spatial information. The AGM was intended to reduce the redundancy and balance the channel weights, and it further increased the AP to 38.2. These experimental results validate the effectiveness of our modules; our FCSSD performs accurately and efficiently.
We show some qualitative results of the SSD [15] in comparison with those of our FCSSD on the PASCAL VOC 2007 test set and MS COCO m i n i v a l in Figure 7. The figure compares images containing objects of various sizes. When the FCSSD was applied to the test images, even small objects were successfully detected due to the effectiveness our model design.

5. Conclusions

In this work, we proposed a feature-enhancement- and channel-attention-guided single-shot detector for object detection called the FCSSD. Our method brings in four effective improvements: an efficient feature-extraction module (EFM), pyramidal aggregation module (PAM), feature pyramid refinement fusion (FPRF), and attention-guided module (AGM). The EFM is designed to exploit multi-scale contextual information in the shallow layers, while the PAM is used to explore richer semantic knowledge in the deep layers. The FPFR introduces an effective means of feature fusion for a PFN by compressing multi-scale features into a refined feature. The AGM balances the channel weights and deals with the aliasing effects of the fused features in the FPN, thus accelerating the training process and further improving the performance. Experiments on public datasets proved that our approach achieved competitive results in comparison with those of mainstream one-stage and two-stage methods, achieving better a tradeoff between accuracy and speed. Since one-stage object detectors achieve a faster speed and better performance, further research is needed to move on to lightweight models for real applications that could easily be embedded into mobile devices.

Author Contributions

Conceptualization, Y.L. and L.W.; methodology, Y.L. and L.W.; software, L.W. and Y.L.; validation, L.W. and Y.L.; formal analysis, L.W. and Y.L.; investigation, L.W. and Y.L.; resources, L.W. and Y.L.; data curation, Y.L. and L.W.; writing—original draft preparation, L.W. and Y.L.; writing—review and editing, L.W., Y.L., and Z.W.; visualization, L.W. and Y.L.; supervision, L.W.; project administration, L.W.; funding acquisition, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This researchwas funded by the National Natural Science Foundation of China (NSFC), 11871438, and the Zhejiang Provincial Natural Science Foundation of China, LZ22F020010.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

PASCAL VOC: http://host.robots.ox.ac.uk/pascal/VOC/; MS COCO: https://cocodataset.org/ accessed on 9 July 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kang, K.; Li, H.; Yan, J.; Zeng, X.; Yang, B.; Xiao, T.; Zhang, C.; Wang, Z.; Wang, R.; Wang, X.; et al. T-cnn: Tubelets with convolutional neural networks for object detection from videos. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2896–2907. [Google Scholar] [CrossRef]
  2. Wu, Q.; Shen, C.; Wang, P.; Dick, A.; Van Den Hengel, A. Image captioning and visual question answering based on attributes and external knowledge. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1367–1381. [Google Scholar] [CrossRef] [PubMed]
  3. Dai, J.; He, K.; Sun, J. Instance-aware semantic segmentation via multi-task network cascades. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3150–3158. [Google Scholar]
  4. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  5. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  7. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision; 2015; pp. 1440–1448. Available online: https://openaccess.thecvf.com/content_iccv_2015/html/Girshick_Fast_R-CNN_ICCV_2015_paper.html (accessed on 9 July 2022).
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. Available online: https://proceedings.neurips.cc/paper/2015/hash/14bfa6bb14875e45bba028a21ed38046-Abstract.html (accessed on 9 July 2022). [CrossRef] [PubMed]
  9. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  10. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar]
  11. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  12. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  13. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  14. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  16. Li, Z.; Zhou, F. FSSD: Feature fusion single shot multibox detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  17. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. Dssd: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  18. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  19. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  20. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 821–830. [Google Scholar]
  21. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020; pp. 10781–10790. [Google Scholar]
  22. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  23. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  24. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  25. Liu, Y.; Wang, Y.; Wang, S.; Liang, T.; Zhao, Q.; Tang, Z.; Ling, H. Cbnet: A novel composite backbone network architecture for object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11653–11660. [Google Scholar]
  26. Kim, S.W.; Kook, H.K.; Sun, J.Y.; Kang, M.C.; Ko, S.J. Parallel feature pyramid network for object detection. In Proceedings of the European Conference on Computer Vision (ECCV), online, 8–14 September 2018; pp. 234–250. [Google Scholar]
  27. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  28. Li, Y.; Chen, Y.; Wang, N.; Zhang, Z. Scale-aware trident networks for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6054–6063. [Google Scholar]
  29. Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-shot refinement neural network for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4203–4212. [Google Scholar]
  30. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  31. Ren, M.; Zemel, R.S. End-to-end instance segmentation with recurrent attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6656–6664. [Google Scholar]
  32. Lu, J.; Xiong, C.; Parikh, D.; Socher, R. Knowing when to look: Adaptive attention via a visual sentinel for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 375–383. [Google Scholar]
  33. Li, H.; Liu, Y.; Ouyang, W.; Wang, X. Zoom out-and-in network with map attention decision for region proposal and object detection. Int. J. Comput. Vis. 2019, 127, 225–238. [Google Scholar] [CrossRef]
  34. Choi, H.T.; Lee, H.J.; Kang, H.; Yu, S.; Park, H.H. SSD-EMB: An improved SSD using enhanced feature map block for object detection. Sensors 2021, 21, 2842. [Google Scholar] [CrossRef] [PubMed]
  35. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  36. Qilong, W.; Banggu, W.; Pengfei, Z.; Peihua, L.; Wangmeng, Z.; Qinghua, H. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In CVF Conference on Computer Vision and Pattern Recognition (CVPR); Available online: https://arxiv.org/abs/1910.03151 (accessed on 9 July 2022).
  37. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Vedaldi, A. Gather-excite: Exploiting feature context in convolutional neural networks. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  38. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 510–519. [Google Scholar]
  39. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), online, 8–14 September 2018; pp. 3–19. [Google Scholar]
  40. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  41. Jiang, D.; Sun, B.; Su, S.; Zuo, Z.; Wu, P.; Tan, X. FASSD: A feature fusion and spatial attention-based single shot detector for small object detection. Electronics 2020, 9, 1536. [Google Scholar] [CrossRef]
  42. Luo, Y.; Cao, X.; Zhang, J.; Guo, J.; Shen, H.; Wang, T.; Feng, Q. CE-FPN: Enhancing channel information for object detection. Multimed. Tools Appl. 2022, 81, 30685–30704. [Google Scholar] [CrossRef]
  43. Li, H.; Liu, L.; Du, J.; Jiang, F.; Guo, F.; Hu, Q.; Fan, L. An Improved YOLOv3 for Foreign Objects Detection of Transmission Lines. IEEE Access 2022, 10, 45620–45628. [Google Scholar] [CrossRef]
  44. Zhu, Y.; Zhao, C.; Wang, J.; Zhao, X.; Wu, Y.; Lu, H. Couplenet: Coupling global structure with local parts for object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4126–4134. [Google Scholar]
  45. Liu, S.; Huang, D.; Wang, Y. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Salt Lake City, UT, USA, 18–22 June 2018; pp. 385–400. [Google Scholar]
  46. Bell, S.; Zitnick, C.L.; Bala, K.; Girshick, R. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2874–2883. [Google Scholar]
  47. Kong, T.; Sun, F.; Yao, A.; Liu, H.; Lu, M.; Chen, Y. Ron: Reverse connection with objectness prior networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5936–5944. [Google Scholar]
  48. Jeong, J.; Park, H.; Kwak, N. Enhancement of SSD by concatenating feature maps for object detection. arXiv 2017, arXiv:1705.09587. [Google Scholar]
  49. Woo, S.; Hwang, S.; Kweon, I.S. Stairnet: Top-down semantic aggregation for accurate one shot detection. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1093–1102. [Google Scholar]
  50. Zhang, Z.; Qiao, S.; Xie, C.; Shen, W.; Wang, B.; Yuille, A.L. Single-shot object detection with enriched semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5813–5821. [Google Scholar]
  51. Pang, Y.; Wang, T.; Anwer, R.M.; Khan, F.S.; Shao, L. Efficient featurized image pyramid network for single shot detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7336–7344. [Google Scholar]
  52. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  53. Cao, J.; Pang, Y.; Li, X. Triply supervised decoder networks for joint detection and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7392–7401. [Google Scholar]
  54. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), online, 8–14 September 2018; pp. 734–750. [Google Scholar]
  55. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. Scaled-yolov4: Scaling cross stage partial network. In Proceedings of the IEEE/cvf Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13029–13038. [Google Scholar]
  56. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
Figure 1. Overall architecture of the proposed FCSSD. Our approach consists of four components: a standard SSD network; feature-enhancement components: the efficient feature-extraction module (EFM) and pyramidal aggregation module (PAM); the feature pyramid refinement fusion (FPRF) process X; the attention-guided module (AGM). Here, C i denotes the feature map extracted from the CNN backbone, P i denotes the corresponding pyramid levels in the FCSSD, and M i is the final output of every level. An input image is first downsampled and passed through the EFM and PAM to produce features with more semantic and rich global contextual information. The FPRF uses an FPN-style design to refine the previous features for fusion. The AGM adopts a channel attention scheme to extract channel-balanced weights, and it reduces the aliasing effects of the fused features.
Figure 1. Overall architecture of the proposed FCSSD. Our approach consists of four components: a standard SSD network; feature-enhancement components: the efficient feature-extraction module (EFM) and pyramidal aggregation module (PAM); the feature pyramid refinement fusion (FPRF) process X; the attention-guided module (AGM). Here, C i denotes the feature map extracted from the CNN backbone, P i denotes the corresponding pyramid levels in the FCSSD, and M i is the final output of every level. An input image is first downsampled and passed through the EFM and PAM to produce features with more semantic and rich global contextual information. The FPRF uses an FPN-style design to refine the previous features for fusion. The AGM adopts a channel attention scheme to extract channel-balanced weights, and it reduces the aliasing effects of the fused features.
Sensors 22 06857 g001
Figure 2. The architecture of the efficient feature-extraction module (EFM). The multi-dilated convolution layer has three different dilation rates: 1, 3, and 5. The feature aggregation layer contains two operations: element-wise concatenation and element-wise summation.
Figure 2. The architecture of the efficient feature-extraction module (EFM). The multi-dilated convolution layer has three different dilation rates: 1, 3, and 5. The feature aggregation layer contains two operations: element-wise concatenation and element-wise summation.
Sensors 22 06857 g002
Figure 3. Pyramidal aggregation module (PAM).
Figure 3. Pyramidal aggregation module (PAM).
Sensors 22 06857 g003
Figure 4. (a) Using an image pyramid to build a feature pyramid, which is an inefficient method [41]. (b) Reusing the pyramidal feature hierarchy computed by a ConvNet as if it were a featurized image pyramid, as with a single-shot multi-box detector (SSD) [15]. (c) Bottom-up and top-down pathway feature fusion, as in an FPN [19]. (d) Our proposed feature pyramid refinement fusion.
Figure 4. (a) Using an image pyramid to build a feature pyramid, which is an inefficient method [41]. (b) Reusing the pyramidal feature hierarchy computed by a ConvNet as if it were a featurized image pyramid, as with a single-shot multi-box detector (SSD) [15]. (c) Bottom-up and top-down pathway feature fusion, as in an FPN [19]. (d) Our proposed feature pyramid refinement fusion.
Sensors 22 06857 g004
Figure 5. Structure of the attention-guided module (AGM).
Figure 5. Structure of the attention-guided module (AGM).
Sensors 22 06857 g005
Figure 6. Flow diagram of the training and detection process [43].
Figure 6. Flow diagram of the training and detection process [43].
Sensors 22 06857 g006
Figure 7. Qualitative results of the SSD in comparison with those of the FCSSD on the PASCAL VOC and MS COCO datasets.
Figure 7. Qualitative results of the SSD in comparison with those of the FCSSD on the PASCAL VOC and MS COCO datasets.
Sensors 22 06857 g007
Table 1. Dataset statistics of PASCAL VOC and MS COCO.
Table 1. Dataset statistics of PASCAL VOC and MS COCO.
NameTrain ImagesValidation ImagesTest ImagesCategory
VOC 2007 [23]25012510495220
MS COCO 2017 [40]1128750004067080
Table 2. Comparison of our method with existing detectors on the PASCAL VOC 2007 test set (with IoU = 0.5); 07 + 12: 07 trainval + 12 trainval.
Table 2. Comparison of our method with existing detectors on the PASCAL VOC 2007 test set (with IoU = 0.5); 07 + 12: 07 trainval + 12 trainval.
MethodTraining DataBackboneInput sizemAPFPS
Two-stage detectors:
Faster R-CNN [8]07 + 12VGG-16 1000 × 600 73.27
Faster R-CNN [8]07 + 12ResNet-101 1000 × 600 76.45
R-FCN [10]07 + 12ResNet-101 1000 × 600 80.59
CoupleNet [44]07 + 12VGG-16 1000 × 600 82.78
ION [46]07 + 12VGG-16 1000 × 600 79.21.3
One-stage detectors:
SSD [15]07 + 12VGG-16 300 × 300 77.246
RON320++ [47]07 + 12VGG-16 320 × 320 76.620
DSSD [17]07 + 12ResNet-101 321 × 321 78.610
R-SSD [48]07 + 12VGG-16 300 × 300 78.535
YOLOv2 [12]07 + 12DarkNet-19 544 × 544 78.640
StrairNet [49]07 + 12VGG-16 300 × 300 78.830
DES [50]07 + 12VGG-16 300 × 300 79.776
PFPNet [26]07 + 12VGG-16 300 × 300 80.040
RFBNet [45]07 + 12VGG-16 300 × 300 80.583
RefineDet [29]07 + 12VGG-16 320 × 320 79.776
FCSSD (Ours)07 + 12VGG-16 300 × 300 81.555
SSD [15]07 + 12VGG-16 512 × 512 77.246
FSSD [16]07 + 12VGG-16 512 × 512 80.935.7
DES [50]07 + 12VGG-16 300 × 300 81.731
RefineDet [29]07 + 12VGG-16 512 × 512 81.821
RFBNet [45]07 + 12VGG-16 512 × 512 82.238
PFPNet [26]07 + 12VGG-16 512 × 512 82.326
FCSSD (Ours)07 + 12VGG-16 512 × 512 83.235
Table 3. The results of the mainstream one-stage and two-stage detectors on the MS COCO testdev.
Table 3. The results of the mainstream one-stage and two-stage detectors on the MS COCO testdev.
MethodBackboneInput SizeTime (ms) AP AP 50 AP 75 AP s AP m AP l
Two-stage detectors:
Faster R-CNN [8]VGG-16 1000 × 600 14724.245.323.57.726.437.1
Faster FPN [19]ResNet-101-FPN 1000 × 600 24036.259.139.018.239.048.2
CoupleNet [44]ResNet-101 1000 × 600 12134.454.837.213.438.150.8
Mask R-CNN [9]ResNetXt-101-FPN 1280 × 800 21039.862.343.422.143.251.2
Cascade R-CNN [24]ResNet-101-FPN 1280 × 800 14142.862.146.323.745.555.2
One-stage detectors:
SSD [15]VGG-16 300 × 300 2025.143.125.86.625.941.4
DSSD [17]ResNet-101 321 × 321 -28.046.129.27.428.147.6
RetinaNet [18]ResNet-101 500 × 832 9034.453.136.814.738.549.1
DES [50]VGG-16 300 × 300 -28.347.329.48.529.945.2
RFBNet [45]VGG-16 300 × 300 1530.349.331.811.831.945.9
EFIPNet [51]VGG-16 300 × 300 1430.048.831.710.932.846.3
RefineDet320+ [29]VGG-16 320 × 320 -35.256.137.719.537.247.0
M2det [52]VGG-16 320 × 320 -38.959.142.424.441.547.6
FCSSD(ours)VGG-16 300 × 300 3538.259.441.515.340.358.6
YOLOv2 [12]DarkNet 544 × 544 2521.644.019.25.022.435.5
YOLOv3 [13]DarkNet-53 416 × 416 3531.055.332.315.333.242.8
SSD [15]VGG-16 512 × 512 2828.848.530.310.931.843.5
DSSD [17]ResNet-101 513 × 513 15633.253.335.213.035.451.1
RefineDet [29]VGG-16 512 × 512 4533.054.535.516.336.344.3
RefineDet  [29]ResNet-101 512 × 512 -36.457.539.516.639.951.4
RFBNet-E [45]VGG-16 512 × 512 3034.455.736.417.637.047.6
TripleNet [53]ResNet-101 512 × 512 -37.459.339.618.539.052.7
EfficientDet-D1 [21]EfficientNet-B1 640 × 640 5039.658.642.317.944.356.0
CornerNet [54]Hourglass-104 511 × 511 22740.556.543.119.442.753.9
YOLOv4 [14]CSPDarkNet53 416 × 416 3841.262.844.320.444.456.0
M2det [52]VGG-16 512 × 512 -42.962.547.228.047.452.8
Scaled-YOLOv4 [55]CD53 608 × 608 -45.564.149.527.049.056.7
YOLOX-M [56]Modified CSP v5 640 × 640 -46.465.450.626.351.059.9
FCSSD(ours)VGG-16 512 × 512 6541.563.146.921.846.256.5
Table 4. Ablation studies of the FCSSD on the PASCAL VOC 2007 test set with an image size of 300 × 300 .
Table 4. Ablation studies of the FCSSD on the PASCAL VOC 2007 test set with an image size of 300 × 300 .
MethodEFMPAMFPRFAGMmAP
Baseline SSD 77.2
(a) 78.6
(b) 78.4
(c) 79.2
(d) 79.5
(e) 80.2
FCSSD (ours)81.5
Table 5. Ablation studies of the FCSSD on the MS COCO m i n i v a l with an image size of 300 × 300 .
Table 5. Ablation studies of the FCSSD on the MS COCO m i n i v a l with an image size of 300 × 300 .
SchemeMethods AP AP 50 AP 75 AP s AP m AP l
ABaseline SSD25.143.125.86.625.941.4
BA + EFM27.245.227.27.927.344.6
CB + PAM29.347.930.69.130.145.2
DC + FPRF37.556.938.813.538.354.2
ED + AGM38.259.441.515.340.358.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Y.; Wang, L.; Wang, Z. Single-Shot Object Detection via Feature Enhancement and Channel Attention. Sensors 2022, 22, 6857. https://doi.org/10.3390/s22186857

AMA Style

Li Y, Wang L, Wang Z. Single-Shot Object Detection via Feature Enhancement and Channel Attention. Sensors. 2022; 22(18):6857. https://doi.org/10.3390/s22186857

Chicago/Turabian Style

Li, Yi, Lingna Wang, and Zeji Wang. 2022. "Single-Shot Object Detection via Feature Enhancement and Channel Attention" Sensors 22, no. 18: 6857. https://doi.org/10.3390/s22186857

APA Style

Li, Y., Wang, L., & Wang, Z. (2022). Single-Shot Object Detection via Feature Enhancement and Channel Attention. Sensors, 22(18), 6857. https://doi.org/10.3390/s22186857

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop