Next Article in Journal
Existence Conditions and General Solutions of Closed-form Inverse Kinematics for Revolute Serial Robots
Previous Article in Journal
3D Convolutional Neural Networks for Remote Pulse Rate Measurement and Mapping from Facial Video
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced Feature Pyramid Object Detection Network for Autonomous Driving

1
Graduate School of Information, Production and Systems, Waseda University, Fukuoka 8080135, Japan
2
Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(20), 4363; https://doi.org/10.3390/app9204363
Submission received: 29 August 2019 / Revised: 12 October 2019 / Accepted: 14 October 2019 / Published: 16 October 2019
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Feature Pyramid Network (FPN) builds a high-level semantic feature pyramid and detects objects of different scales in corresponding pyramid levels. Usually, features within the same pyramid levels have the same weight for subsequent object detection, which ignores the feature requirements of different scale objects. As we know, for most detection networks, it is hard to detect small objects and occluded objects because there is little information to exploit. To solve the above problems, we propose an Enhanced Feature Pyramid Object Detection Network (EFPN), which innovatively constructs an enhanced feature extraction subnet and adaptive parallel detection subnet. Enhanced feature extraction subnet introduces Feature Weight Module (FWM) to enhance pyramid features by weighting the fusion feature map. Adaptive parallel detection subnet introduces Adaptive Context Expansion (ACE) and Parallel Detection Branch (PDB). ACE aims to generate the features of adaptively enlarged object context region and original region. PDB predicts classification and regression results separately with the two features. Experiments showed that EFPN outperforms FPN in detection accuracy on Pascal VOC and KITTI datasets. Furthermore, the performance of EFPN meets the real-time requirements of autonomous driving systems.

1. Introduction

Autonomous driving systems cognize the surrounding environment by modeling the street scenery information acquired by sensors, and then makes driving decisions. Among them, object detection plays an important role in the construction of realistic scenes. In recent years, the improvement of deep learning theory and computing power has significantly accelerated the development of object detection. Nowadays, object detection methods are divided into two branches, one-stage and two-stage. Two-stage algorithms are widely applied in tasks with a high degree of accuracy. A typical two-stage deep learning-based detection framework includes a feature extraction subnet, a proposal extraction subnet, and a detection subnet. The feature extraction subnet [1,2,3] uses a convolutional neural network to extract features with high robustness and rich information through end-to-end training process. The proposal extraction subnet [4,5] generates regions of interest (RoIs), including foreground positive samples and background negative samples from the feature map. The detection subnet utilizes the pooling feature of RoIs to predict the classification and regression results of detected objects [5,6,7,8,9].
To utilize the features obtained from the feature extraction subnet, initially, researchers feed the single scale features from the last convolution layer into the subsequent subnet [5,7,8]. Although these features are rich in higher-level semantic information, they lack detailed information. This leads to a poor detection performance for small objects and occluded objects. Two kinds of network structures are proposed to solve this problem. One is to do prediction separately on feature layers of different resolutions [10,11]. The other is to merge multi-resolution features at first and then to predict on the merged feature map [12,13]. Further studies show that taking advantages of both structures can get more accurate results [14,15,16,17]. FPN is a state-of-the-art network of such structure. FPN constructs a feature pyramid with high-level semantics throughout and independently predicts at each pyramid level. The key design for building feature maps for each pyramid level is the lateral connection. It merges the semantically stronger feature maps from top-down with feature maps which are rich in detail localization information from the same bottom-up level. However, feature maps within the same pyramid level have the same weight for subsequent object detection, which ignores the feature requirements of different scale objects. Inspired by the features strengthen block in image classification [18] and segmentation [19] field, we propose a generic FWM to recalibrate feature maps in each pyramid level. For each pyramid layer, learnable weights are applied between features along space and channel to boost useful features and suppress useless features specific for different scale detection requirements. After that, a new enhanced feature pyramid is reconstructed with more powerful feature representation.
After calculating the shared features, the proposal extraction subnet generates RoIs. Then, the detection subnet adopts RoI pooling to map and extract the corresponding fixed-size features of RoIs for classification and regression. However, for small and occluded objects whose feature information is limited by their size or physical occlusion, their RoI features are less informative for subsequent prediction. Researchers argue that using image evidence beyond RoI content benefits object detection [10,13,20,21,22,23] and segmentation [21]. Such context information can be integrated by a spatial recurrent neural network or acquired by expanding or deforming original RoI regions. We mainly focus on local context embedding methods by expanding RoI regions. Gidaris et al. [20] defined ten kinds of context regions around RoI to enrich the RoI representation, and then integrated each region’s information in separate fully connected layers and concatenated them together as final features of each RoI. It does improve accuracy at the cost of calculating a lot of redundant and overlapping context blocks, which is incompatible with real-time autonomous driving systems. Cai et al. [10] stacked features of a RoI and its larger context RoI together, and then compressed them by a convolution operation as a merged RoI representation. However, all the RoIs share a fixed context expansion ratio, which is set by human and requires careful tuning. This causes insufficient or redundant context information supplementation of various scale objects. Wang et al. [22] introduced local competition mechanism to select the most useful context region among three different expansion ratios of RoIs, but all of the expansion values are still set manually. Since each object has different requirements for context information, the amount of context complement should be distinct between each RoI. For instance, it is difficult to detect small objects without referring to the surrounding context, while the pure object characteristics is what a detection model should focus on in the case of large objects. To meet this requirement, we construct an adaptive parallel detection subnet, which innovatively introduces Adaptive Context Expansion(ACE) and Parallel Detection Branch (PDB). We naturally leverage the pyramid shape of FPN to adaptively supply the context information of RoI of different sizes. For example, we provide more context information for low-resolution small object while less for large object. Then, features of context expansion RoI (ceRoI) and original RoI, respectively, serve as the input of PDB to separately predict classification and regression results. With the parallel design, we can make the best of two features, which not only maintains the location accuracy but also improves classification performance.
The rest of this paper is organized as follows. In Section 2, we introduce our proposed EFPN with an innovative enhanced feature extraction subnet and an adaptive parallel detection subnet. In Section 3, we evaluate and discuss the object detection results of EFPN on several open datasets. In Section 4, we introduce the application of EFPN as vision-based object detection module in an autonomous driving cart system. Finally, conclusions are drawn in Section 5.

2. Proposed Method

EFPN is our proposed object detection network. Its architecture is shown in Figure 1. Firstly, in enhanced feature extraction subnet, we generate pyramid features in the same way as FPN. Features in each pyramid level are weighted by our proposed FWM, and a new enhanced feature pyramid is reconstructed as the input for the following procedure. Secondly, in the proposal extraction subnet, Region Proposal Network (RPN) [5] is used to generate anchors of various shapes on the enhanced pyramidal feature map. Thirdly, in adaptive parallel detection subnet, ACE is applied to extract the feature of ceRoI and RoI for each foreground RoI. Two kinds of RoI features are, respectively, fed into PDB to predict classification and regression as the final detection results.

2.1. Enhanced Feature Extraction Subnet

Generally, FPN first builds the bottom-up layers { C 2 , C 3 , C 4 , C 5 } by the feedforward computation of backbone ConvNet. Then, FPN constructs each top-down feature maps by element-wised adding the top-down feature maps of the last pyramid level with the bottom-up feature maps of the same pyramid level, which is shown in t Figure 2 (left). The set of pyramidal feature maps built by FPN is { P 2 , P 3 , P 4 , P 5 } .
Despite such a careful design for generating refined merged feature maps for different levels, it is not strong enough for the information of spatial and channel features to different scaled objects. We hypothesize that both spatial-wise and channel-wise recalibrating merged feature maps can encourage current pyramid layer detection. Hence, we propose FWM to enhance the pyramid feature. The structure of FWM is shown in Figure 2 (right).
FWM starts by modeling the feature dependency of the feature maps in each pyramid level, and further learns the feature importance vector to recalibrate the feature maps to emphasize the useful features. Specially, FWM in each pyramid level is in the same structure but has different learnable weights, which results in different calculated feature weights. Each FWM consists of three sub-modules: Feature Channel Weight Module (FCWM), Feature Spatial Weight Module (FSWM) and Feature Channel Spatial Weight Module (FCSWM). FCWM and FSWM calculate the feature importance vector along channel and spatial location. FCSWM combines the recalibrated weighted feature maps after FCWM and FSWM as the new pyramidal feature maps. The detailed design of the three submodules are described in the following subsections.

2.1.1. Feature Channel Weight Module (FCWM)

FCWM focuses on enhancing features along channel of each pyramid level. FCWM first explicitly models the dependency of features along channel and learns a channel specific descriptor through the squeeze-and-excitation method [18]. Then, it emphasizes the useful channels for more efficient global information expression of feature maps in each pyramid level.
Suppose the feature maps in nth pyramid level is P n , which is generated by FPN. H n and W n are the spatial height and width of P n , respectively. The ith channel feature is P n i .
At the beginning, we do global average pooling on P n i to get the global distribution response Z n i :
Z n i = 1 H n × W n p H n q W n P n i ( p , q ) .
We use two fully connected layers to map the non-linear correlation between all global distribution responses Z n and obtain the feature importance vectors Z n ^ :
Z n ^ = W n 1 ( δ ( W n 2 Z n ) ) ,
where W n 1 is the weight of the first fully connected layer. W n 2 is the weight of the second fully connected layer. δ represents the ReLU function.
Then, we normalize Z n ^ to [ 0 , 1 ] as a weight vector:
r n = σ ( Z n ^ ) ,
where σ represents Sigmoid function.
Finally, we assign the weight r n to the original feature P n and get the new pyramid feature P c r n after channel-wised recalibration:
P c r n = P n r n = [ P n 1 r n 1 , P n 2 r n 2 , . . . , P n n r n n ] .

2.1.2. Feature Spatial Weight Module (FSWM)

Similar to the design of FCWM, FSWM enhances the features along spatial location of each pyramid level, which emphasizes the effective pixels and depresses the ineffective or low-effect pixels.
We define P n ( p , q ) as the clipping of all channel features at each feature point ( p , q ) of P n . First, we integrate all the features of each point through a convolution operation to get the spatial importance vector O n ( p , q ) :
O n ( p , q ) = W n 3 P n ( p , q ) ,
where W n 3 is the convolution kernel weight.
Then, we normalize O n ^ to [ 0 , 1 ] as a weight vector t n
t n = σ ( O n ^ ) ,
where σ represents Sigmoid function.
Finally, the normalized weights are spatially weighted to each pixel to get the new feature P s r n :
P s r n = P n t n = [ P n ( 1 , 1 ) t n ( 1 , 1 ) , P n ( 1 , 2 ) t n ( 1 , 2 ) , . . . , P n ( H n , W n ) t n ( H n , W n ) ] .

2.1.3. Feature Channel Spatial Weight Module (FCSWM)

FCSWM combines the channel-wised weighted P c r n obtained by FCWM and the spatially weighted P s r n obtained by FSWM to generate a new recalibrated feature P r n . The combination operation is implemented by addition:
P r n = P c r n + P s r n .
P r n encourages original feature maps to be both spatial-wise and channel-wise more informative. In EFPN, we replace the initial feature pyramid features { P 2 , P 3 , P 4 , P 5 } by the recalibrated enhanced pyramid features { P r 2 , P r 3 , P r 4 , P r 5 } as the input feature of proposal extraction subnet and detection subnet.

2.2. Adaptive Parallel Detection Subnet

To inject object context information, we design adaptive parallel detection subnet, as shown in Figure 1. Adaptive parallel detection subnet includes ACE and PDB. ACE calculates the context region extension ratio of each RoI, and then generates and extracts the feature of ceRoI and original RoI after RoI pooling. PDB inputs the feature of ceRoI and RoI separately to object classification branch and regression branch to predict the classification and regression as the final detection results.

2.2.1. Adaptive Context Expansion (ACE)

As we know, small objects need extra information to help detection. FPN assigns small RoIs to finer-resolution level to add extra detailed RoI feature. Suppose a RoI has width w and height h (on the input image of the network); the level k of RoI is determined by its area size S:
S = w h ,
k = k 0 + log 2 ( S / S 0 ) .
Similarly, in ACE, an extra context feature is embedded for adding surrounding information for small RoIs. We define two context expanding criteria, which guide the calculation of context region expansion ratio for each RoI.
Vertical expansion criterion leverages the hierarchical structure of pyramid network to calculate the vertical context expansion ratio R v :
R v = α k ,
where α denotes the vertical context enlarge coefficient. Since context expending of RoI is positively correlated with the level of RoI, α is always positive. In Figure 3, the small R o I b and the large R o I c belong to different pyramid levels, hence their R v are clearly distinct from each other, that is a larger one for R o I b and a smaller one for R o I c .
Horizontal expansion criterion focuses on the calculation of refined expansion ratios between different objects on the same pyramid level. Although RoIs within the same scale range have the same R v , their varying sizes lead to different requirements. We take this factor into account to provide more precise horizontal context expansion ratio R h :
R h = β ( log 2 ( S / S 0 ) + k 0 - k ) ,
where β denotes the horizontal context enlarge coefficient. log 2 ( S / S 0 ) + k 0 - k is the residual part of level vector, ranging [ 0 , 1 ] , which indicates the size difference among the RoIs of the same pyramid level. In Figure 3, for R o I a and R o I d of the same pyramid level, the smaller its area is, the more context information it needs. Thus, they have slightly different R h , with a larger one for R o I a and a smaller one for R o I d .
Above all, the final context expansion ratio R, which considers the two independent criteria of each RoI, is:
R = R v + R h + γ ,
where γ is used to adjust magnitude of extension ratio among different dataset resulting from factors, such as background complexity, objects density, etc.
Next, the scale-up operation Θ based on the center point of each RoI is used to get the context region ceRoI:
c e R o I = Θ ( R o I ) ,
where the ceRoI has the width of w × R and height of h × R with the same center as the original RoI.

2.2.2. Parallel Detection Branch (PDB)

ceRoI enriches the feature of RoI by aggregating its context environment information around, which will be helpful for object recognition. However, ceRoI changes the geometric position representation of original RoI. The misalignment of two features may somehow cause negative effects to sensitive bounding box regression.
To make better use of the context information, unlike most popular detection subnets, we design a divide-and-conquer prediction strategy, as shown in Figure 1. Features of ceRoI and RoI after RoI pooling are extracted and separately fed into two parallel branches. Each branch includes two hidden 1024D fully connected layers, which are attached by ReLU for object classification and regression.

2.2.3. Loss Function

When training the adaptive parallel detection subnet, the parameters W are randomly initialized. Then, W are learned from a set of training samples S = { ( X i , Y i ) } i = 1 N , where X i is a training RoI, and Y i = ( y i , b i ) is the combination of its class label y i 0 , 1 , 2 , . . . K and bounding box coordinates b i = ( b i x , b i y , b i w , b i h ) . K is the number of classes.
We use a multi-task loss L on each labeled X i to jointly train for classification and regression:
L ( X , Y | W ) = L c l s ( p ( Θ ( X ) ) , y ) + λ [ y 1 ] L l o c ( b ^ , b ) ,
where p ( X ) = ( p 0 ( X ) , . . . , p K ( X ) ) is the probability distribution over classes, λ is a trade-off coefficient, L c l s ( p ( Θ ( X ) ) , y ) = - log p y ( Θ ( X ) ) is the cross-entropy loss, b ^ = ( b ^ x , b ^ y , b ^ w , b ^ h ) is the regressed bounding box, and
L l o c ( b ^ , b ) = 1 4 j { x , y , w , h } s m o o t h L 1 ( b ^ j , b j )
is the smoothed bounding box regression loss [7]. The bounding box loss is only used for positive samples and the optimal parameters W * = a r g m i n W L ( W ) are learned by stochastic gradient descent optimizer.

3. Experiments on Open Datasets

We evaluated our method using three challenging open datasets.
Pascal VOC [24] contains 20 categories of indoor and outdoor objects class. We mainly focused on the average precision (AP) of six classes appearing in road scenes and the mean average precision (mAP) of the whole dataset. We used VOC07+12 dataset, which contains 16551 images for training and 4952 images for testing to evaluate our proposed method. To further investigate the effectiveness of each module structure, we used VOC07 dataset, which contains 5011 training images and 4952 testing images for ablation study.
KITTI [25] is a large automatic driving dataset. Here, we used its 2D object detection dataset for evaluation. In our experiment, we redefined the object classes into two classes: c a r and p e d e s t r i a n . c a r includes [ V a n , T r u c k , C a r , T r a m ] and p e d e s t r i a n includes [ P e r s o n s i t t i n g , P e d e s t r i a n , C y c l i s t ] . KITTI provides 7481 images for training and 7518 for testing. Since no ground truth is available for the test set, we split the training set into training set and validation set by 8:1.
Cityscapes [26] is a road scene image segmentation dataset. Pixel-level annotation for segmentation task contains more small and occluded labeled objects than detection datasets, which is full of challenges. Therefore, we converted Cityscapes into detection dataset to further test our model. The definition of object class is the same as that of KITTI. Cityscapes detection dataset consists of 2842 images.

3.1. Implementation Details

We implemented EFPN in Python within Pytorch deep-learning framework. Following Lin et al. [14], we resized each image so that its shorter side has 600 pixels. The network was trained using one NVIDIA RTX 2080 GPU with 1 image per mini-batch.
For proposal extraction subnet, we adopted the same design and training parameters as FPN.
For adaptive parallel detection subnet, we adopted RoI align [27] as RoI Pooling mechanism. We used a weight decay of 0.0001 and a momentum of 0.9. The learning rate started from 0.001 and was divided by 10 at every five epochs. The model had 10 total epochs.
For the parameter setting in ACE, we set k 0 to 4 and S 0 to 224 in Equation (10), consistent with the canonical ImageNet pre-training size. We set the [ α , β , γ ] in Equation (13) for different datasets: [0.1, 0.01, 0.6] for VOC07+12, [0.2, 0.01, 1.1] for VOC07, [0.15, 0.1, 0.9] for KITTI, and [0.15, 0.1, 0.9] for Cityscapes.

3.2. Object Detection Results on Pascal VOC

To evaluate the performance of EFPN and the two new subnets proposed in this paper, we compared them with other related object detection algorithms on VOC07+12 dataset. Results are shown in Table 1. Note that Approaches (b)–(d) and (i) were implemented and tested on the same platform as the proposed EFPN, while the results of Approaches (e)–(h) and (j) were from their corresponding publications.
Compared with baseline (d), the controlled experiments (b) and (c), which separately replace one part of the original FPN model, prove the validity of the proposed enhanced feature extraction subnet and adaptive parallel detection subnet. Since the two subnets enhance the feature by applying weight distribution of the entire feature map and feature supplementary of each RoI, respectively, they do not overlap or inhibit each other. This is evidenced by the increased accuracy of the merged method (a) compared with either single module (b) or (c). Overall, among all related methods, the proposed EFPN (a) holds the highest mAP and highest AP in car, person and train.
Adaptive parallel detection subnet assembles two new module ACE and PDB. We designed ablation studies to quantify the effect of the two modules on the VOC07 dataset. The results are reported in Table 2.
The increased detection accuracy in (b)–(e) compared with baseline (a) proves that adding context to a tight-fitting RoI is beneficial. However, different AP improvement of different context expansion ratio indicates that the expansion range has different effects on various objects. If we simply introduce a fixed context expansion ratio for all objects, it may result in mismatching for different object requirements. In the case that the detection branches are all in parallel design, the performance of our proposed ACE (i) is better than all fixed context expansion methods, which proves the effectiveness of our adaptive strategy.
We compared four detection architectures in context integration networks.
S h a r e in (a) denotes the normal object detection network, which shares the fully connection layer of both classification and regression without context embedding.
M e r g e is the context embedding method proposed in [10], where the features extracted from RoI and ceRoI after RoI pooling are paired and merged by a 3 × 3 convolution, and then same as S h a r e design. We tested a fixed context expansion ratio of 0.5, which is the same as [10] in (f). The result shows it slightly improves by 0.4 mAP. Nevertheless, we found that the accuracy is reduced when applying the adaptive context ratio to (j). One cause may be that the inconsistent expansion ratio leads to variable feature representation of ceRoI, thus it is hard to train a convolution kernel to merge the various feature pair.
C o n c a t in (h) reproduces the context embedding design proposed in [20]. It sends each ceRoI into different fully-connected layers for feature generation, and then concatenates the features. The rest of the design is the same as S h a r e . Since we mainly focused on the design of detection branch, for lighting parameters and equally contrasting with other methods, we concatenated one ceRoI with context expansion ratio of 0.5 with the original RoI in our experiment. Among all the architectures, C o n c a t with maximal parameters only improves by 0.3 mAP. We suppose it is hard to take full advantage of this design for the integration of a few features.
Unlike the popular design of sharing parameters for classification and regression, PDB (i) offers precision improvements over traditional context embedded detection branch paradigms. Parallel branches can be trained to attentively strengthen their information integration ability for specific task. Moreover, cooperating with ACE, PDB can maximize the usage of context features, boosting detection performance by 0.8 points within the affordable calculations increase.

3.3. Object Detection Results on KITTI

We evaluated model performance in autonomous driving scene dataset KITTI in Table 3.
(a) and (b) show the results of FPN and EFPN, which were both tested and trained on KITTI with a 0.5 IoU threshold. The mAP of EFPN is 1.4 points higher than baseline FPN. Moreover, the AP of pedestrian, which contains more small size objects, is increased lot by 2.7 points.
The comparison of (c) and (d) shows that the performance of EFPN with an IoU threshold of 0.75 also increases by 1.4 points. We suppose that the improvement of localization accuracy is probably caused by proposed PDB. The parallel design forces each fully connected layer to attentively focus on its own task. The AP improvement of high IoU threshold denotes the increase of high-quality detections, which demonstrates the proposed EFPN is well-qualified for high-security applications such as autonomous driving. Figure 4 shows several detection examples on KITTI. We can see that more accurate boxes are generated by EFPN and more small and occluded objects have been detected.

3.4. Object Detection Results on Cityscapes

To verify the model generalization ability of EFPN, we trained the model on KITTI and tested on Cityscapes. The results are shown in Table 4.
EFPN is better than the baseline by 0.4, 1.5, and 0.9 points in the AP of car, pedestrian and mAP, respectively. Figure 5 shows some detection examples on Cityscapes, where more small objects and occluded objects are detected by proposed EFPN. To a certain extent, EFPN can be extended to other environments and maintain its accuracy improvement.
Compared with the test results in KITTI, the overall APs in Cityscapes are much lower for some reasons. One cause is that the proportion of small objects and occluded objects increases greatly in Cityscapes and some of the pixel-level small object vanished after several down samplings. It may also be due to the great differences in image quality, lighting situation and traffic environment complexity between Cityscapes and KITTI.

4. Application in Autonomous Driving System

To further test the model practicality in real autonomous driving scenes, we installed the proposed EFPN in an autonomous driving cart for freight transportation inside a factory. The testing shows that EFPN can effectively provide surrounding object information for autonomous vehicles.
EFPN works in the perception module of autonomous driving cart. The perception module first feeds the data obtained from binocular vision camera into EFPN for object detection. Then, the results are validated by laser radar information. Next, all object information in current frame is wrapped into a ROS message and sent to decision module for further path planning and vehicle control. More broadly, EFPN supports plenty of functions such as obstacle avoidance, trajectory following and automatic parking in autonomous driving cart.
Three of the most crucial measures for object detection algorithms in autonomous driving cart are accuracy, real-time performance and vision range.
For accuracy, to satisfy production demand, the autonomous driving cart needs full-time work, including daytime, nighttime and extreme weather such as rain and snow. Besides, there are many special engineering vehicles in the factory whose appearances are significantly different from normal cars. Moreover, closely parked custom vehicles along roadside are always viewed as overlapped objects. The aforementioned challenges increase the detection difficulty and put forward higher requirements on the accuracy of object detection algorithm. The proposed deep learning model EFPN ensures the capacity of robustly detecting complex objects. For better adapting to the factory scene, we constructed a factory object detection dataset based on Pascal VOC, KITTI, and special vehicles that appear in the factory. Meanwhile, we defined a new object class c o n e for further traffic control and parking management. Overall, the factory dataset includes 21392 images, and four classes, which are c a r , p e r s o n , c o n e and b i c y c l e . We trained EFPN on the factory dataset using the same setting as KITTI.
For real-time performance, EFPN operates up to 16 fps cooperating with the 10 Hz vehicle controller, adequately ensuring the safety of autonomous driving cart.
For vision range, due to the restricted limit in the factory, the speed of the autonomous cart is lower than 10 km/h. The max braking distance is less than 0.5 m, which results in 15 m visualization demand of object detection system. Therefore, the proposed EFPN with the small object detection ability within 80 m in daytime and 50 m at nighttime well satisfies the vehicle reaction requirement within safe distances.
Figure 6 shows the real-time detection results of EFPN under some challenging scenarios. Until now, the autonomous driving cart has been operated in the factory for several months. The practical applicability of EFPN has been effectively verified.

5. Conclusions

This paper proposes EFPN, a feature reinforced update based on FPN. By weighting the features in each pyramid level and adaptively utilizing the context information of object, EFPN enhances the feature expression and further improves object detection accuracy. Experiments showed that the accuracy of EFPN surpasses FPN in open datasets. The application of EFPN in autonomous driving cart proves that it can meet the demands of accuracy, efficiency and visibility in multiple scenarios.

Author Contributions

Conceptualization, Y.W. and S.T.; data curation, S.Z. and Y.W.; formal analysis, S.Z. and Y.W.; investigation, Y.W.; methodology, Y.W. and S.T.; software, Y.W. and S.Z.; project administration, Y.W.; resources, Y.W.; supervision, S.T. and H.O.; Validation, Y.W. and S.Z.; visualization, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, Y.W., S.T. and H.O.; and Funding acquisition, H.O.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  2. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  3. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  4. Uijlings, J.R.R.; Sande, K.E.A.V.D.; Gevers, T.; Smeulders, A.W.M. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
  5. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  7. Girshick, R.B. Fast R-CNN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1440–1448. [Google Scholar]
  8. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object detection via region-based fully convolutional networks. In Proceedings of the Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  9. Li, Z.; Peng, C.; Yu, G.; Zhang, X.; Deng, Y.; Sun, J. Light-head r-cnn: In defense of two-stage object detector. arXiv 2017, arXiv:1711.07264. [Google Scholar]
  10. Cai, Z.; Fan, Q.; Feris, R.S.; Vasconcelos, N. A unified multi-scale deep convolutional neural network for fast object detection. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 354–370. [Google Scholar]
  11. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  12. Kong, T.; Yao, A.; Chen, Y.; Sun, F. Hypernet: Towards accurate region proposal generation and joint object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 845–853. [Google Scholar]
  13. Bell, S.; Zitnick, C.L.; Bala, K.; Girshick, R.B. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2874–2883. [Google Scholar]
  14. Lin, T.; Dollár, P.; Girshick, R.B.; He, K.; Hariharan, B.; Belongie, S.J. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  15. Fu, C.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. DSSD: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  16. Kong, T.; Sun, F.; Yao, A.; Liu, H.; Lu, M.; Chen, Y. RON: Reverse connection with objectness prior networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5244–5252. [Google Scholar]
  17. Shrivastava, A.; Sukthankar, R.; Malik, J.; Gupta, A. Beyond skip connections: Top-down modulation for object detection. arXiv 2016, arXiv:1612.06851. [Google Scholar]
  18. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  19. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent spatial and channel ‘squeeze & excitation’ in fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 421–429. [Google Scholar]
  20. Gidaris, S.; Komodakis, N. Object detection via a multi-region and semantic segmentation-aware CNN model. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1134–1142. [Google Scholar]
  21. Zhu, Y.; Urtasun, R.; Salakhutdinov, R.; Fidler, S. segdeepm: Exploiting segmentation and context in deep neural networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4703–4711. [Google Scholar]
  22. Wang, S. PCN: Part and context information for pedestrian detection with cnns. In Proceedings of the British Machine Vision Conference, Oxford, UK, 4–7 September 2017. [Google Scholar]
  23. Chen, X.; Zhu, Y. 3d object proposals for accurate object class detection. In Proceedings of the Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  24. Everingham, M.; Gool, L.V.; Williams, C.K.I.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  25. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  26. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  27. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2961–2969. [Google Scholar]
Figure 1. The structure of EFPN.
Figure 1. The structure of EFPN.
Applsci 09 04363 g001
Figure 2. The structure of enhanced feature extraction subnet.
Figure 2. The structure of enhanced feature extraction subnet.
Applsci 09 04363 g002
Figure 3. The structure of Adaptive Context Expansion (ACE).
Figure 3. The structure of Adaptive Context Expansion (ACE).
Applsci 09 04363 g003
Figure 4. Detection examples on KITTI dataset with FPN and EFPN model. For each pair, on the left is the result of FPN (a) and on the right is the result of EFPN (b). We show detections with scores higher than 0.3.
Figure 4. Detection examples on KITTI dataset with FPN and EFPN model. For each pair, on the left is the result of FPN (a) and on the right is the result of EFPN (b). We show detections with scores higher than 0.3.
Applsci 09 04363 g004
Figure 5. Detection examples on Cityscapes dataset with FPN and EFPN model. For each pair, one the left is the result of FPN (a) and on the right is the result of EFPN (b). We show detections with scores higher than 0.3.
Figure 5. Detection examples on Cityscapes dataset with FPN and EFPN model. For each pair, one the left is the result of FPN (a) and on the right is the result of EFPN (b). We show detections with scores higher than 0.3.
Applsci 09 04363 g005
Figure 6. Real-time detection results in different lighting conditions and weathers in factory scenarios. (a) Detection results on sunny days; (b) Detection results on rainy days; (c) Detection results in the nighttime.
Figure 6. Real-time detection results in different lighting conditions and weathers in factory scenarios. (a) Detection results on sunny days; (b) Detection results on rainy days; (c) Detection results in the nighttime.
Applsci 09 04363 g006
Table 1. Object detection results evaluated on VOC07+12.
Table 1. Object detection results evaluated on VOC07+12.
ApproachBackboneCarPersonBusBikeMotorbikeTrainmAP
(a) EFPNRes10188.785.488.486.888.288.081.6
(b) FPN with enhanced
feature extraction subnet
Res10188.685.486.786.689.086.481.3
(c) FPN with adaptive
parallel detection subnet
Res10188.584.488.088.586.486.981.4
(d) FPNRes10188.284.786.985.585.587.281.1
(e) SSD 513Res10188.183.088.287.687.587.280.6
(f) DSSD 513Res10188.783.789.086.287.585.781.5
(g) R-FCNRes10188.581.286.887.279.985.980.5
(h) MR-CNNVGG85.976.488.084.185.085.078.2
(i) Faster R-CNNRes10185.375.485.180.780.985.376.4
(j) IONVGG85.174.485.483.182.284.275.6
Table 2. Ablation study of the proposed ACE and PDB in adaptive parallel detection subnet on VOC07.
Table 2. Ablation study of the proposed ACE and PDB in adaptive parallel detection subnet on VOC07.
ApproachContext ExpansionDetection BranchmAP
(a) FPN0share75.8
(b) FPN with0.1PDB76.3
(c) FPN with0.3PDB76.3
(d) FPN with0.5PDB75.8
(e) FPN with0.7PDB76.2
(f) FPN with0.5merge76.2
(j) FPN withACEmerge75.5
(h) FPN withACEconcat76.1
(i) FPN withACEPDB76.6
Table 3. Object detection results evaluated on KITTI.
Table 3. Object detection results evaluated on KITTI.
ApproachTest onCarPedestrianmAP
(a) [email protected]KITTI90.378.384.3
(b) [email protected]KITTI90.481.085.7
(c) [email protected]KITTI73.633.453.5
(d) [email protected]KITTI74.035.854.9
Table 4. Object detection results evaluated on Cityscapes.
Table 4. Object detection results evaluated on Cityscapes.
ApproachTest onCarPedestrianmAP
(a) [email protected]Cityscapes38.519.929.2
(b) [email protected]Cityscapes38.921.430.1

Share and Cite

MDPI and ACS Style

Wu, Y.; Tang, S.; Zhang, S.; Ogai, H. An Enhanced Feature Pyramid Object Detection Network for Autonomous Driving. Appl. Sci. 2019, 9, 4363. https://doi.org/10.3390/app9204363

AMA Style

Wu Y, Tang S, Zhang S, Ogai H. An Enhanced Feature Pyramid Object Detection Network for Autonomous Driving. Applied Sciences. 2019; 9(20):4363. https://doi.org/10.3390/app9204363

Chicago/Turabian Style

Wu, Yutian, Shuming Tang, Shuwei Zhang, and Harutoshi Ogai. 2019. "An Enhanced Feature Pyramid Object Detection Network for Autonomous Driving" Applied Sciences 9, no. 20: 4363. https://doi.org/10.3390/app9204363

APA Style

Wu, Y., Tang, S., Zhang, S., & Ogai, H. (2019). An Enhanced Feature Pyramid Object Detection Network for Autonomous Driving. Applied Sciences, 9(20), 4363. https://doi.org/10.3390/app9204363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop