Next Article in Journal
Numerical Analysis of Laser Pattern Effects to Residual Stress on Metal 3D Printing
Previous Article in Journal
Effective Multi-Frame Optical Detection Algorithm for GEO Space Objects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ODEM-GAN: An Object Deformation Enhancement Model Based on Generative Adversarial Networks

1
School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
2
School of Computer Science and Technology, Soochow University, Suzhou 215006, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(9), 4609; https://doi.org/10.3390/app12094609
Submission received: 26 March 2022 / Revised: 27 April 2022 / Accepted: 29 April 2022 / Published: 3 May 2022
(This article belongs to the Topic Applied Computer Vision and Pattern Recognition)

Abstract

:
Object detection has attracted great attention in recent years. Many experts and scholars have proposed efficient solutions to address object detection problems and achieve perfect performance. For example, coordinate-based anchor-free (CBAF) module was proposed recently to predict the category and the adjustments to the box of the object by its feature part and its contextual part features, which are based on feature maps divided by spatial coordinates. However, these methods do not work very well for some particular situations (e.g., small object detection, scale variation, deformations, etc.), and the accuracy of object detection still needs to be improved. In this paper, to address these problems, we proposed ODEM-GAN based on CBAF, which utilizes generative adversarial networks to implement the detection of a deformed object. Specifically, ODEM-GAN first generates the object deformation features and then uses these features to enhance the learning ability of CBFA for improving the robustness of the detection. We also conducted extensive experiments to validate the effectiveness of ODEM-GAN in the simulation of a parachute opening process. The experimental results demonstrate that, with the assistance of ODEM-GAN, the AP score of CBAF for parachute detection is 88.4%, thereby the accuracy of detecting the deformed object by CBAF significantly increases.

1. Introduction

With large-scale datasets [1,2], object detection based on convolutional neural networks (CNNs) has achieved impressive performance. These CNN-based methods are not only improving bounding box object detection [3,4,5], but also making progress on local correspondence [6,7]. However, for some particular situations, such as small object detection, scale variation, and deformations, the traditional CNN-based methods do not perform very well. To address this problem, some scholars have proposed to build multi-level feature pyramids and expand datasets for improving the detection accuracy. The former is from the perspective of knowledge modeling, and the latter is from a data-driven perspective. Most state-of-the-art approaches are based on preset anchor-boxes [8,9,10,11,12,13] or use the feature pyramid network (FPN) [5,14,15,16,17] module to solve small object detection and scale variation.
Although these methods achieve good performance, there are still some inherent limitations. First, these methods tile a large number of artificially preset anchor boxes, and then predict the category of these anchor boxes. To obtain the high intersection over union (IOU) rate between anchor boxes and ground-truth boxes, they refine these anchor boxes one or more times, which will produce too many negative samples and considerable computational redundancy during the prediction process. Second, the anchor boxes have predefined scales and aspect ratios, which belong to discrete space. However, the ground-truth boxes belong to continuous space, so in object detection, there will be anchor boxes and ground-truth boxes that are difficult to accurately align, which will lead to greater loss regression. For small object detection and scale variation, using feature pyramids is the principal method. Small objects are mapped to high-resolution feature maps (the low-level feature maps of the feature pyramid) [17] as they have fine-grained appearance information [18,19]. The feature pyramid is the basic module of object detection at different scales. In addition, the object deformation should also be taken into consideration during detection. A common method to solve deformation is to use a data-driven learning strategy, that is, to collect as many images in different states as possible, and use it to train the detector network to improve the performance of the detector [20,21]. However, it is impossible to collect the state of the object all the time. Moreover, the cost required to collect and label a large amount of data is also inconceivable.
To tackle these limitations, Tang et al. proposed a coordinate-based anchor-free (CBAF) module, which can effectively detect rigid objects of small and scale variation [22]. However, it does not work well with deformation objects such as parachutes. The reason is that the basic principle of the object detection algorithm is to calculate the similarity of the feature on the search image. When the similarity is greater than the pre-set threshold, it is considered the same object. If the parachute is just thrown as the object, its violent shape change will cause the confidence score during the detection process to drop sharply, resulting in detection failure (the detection object is lost multiple times during the experiment). If the fully opened parachute is used as the target, although the parachute will no longer deform drastically afterwards, the algorithm can stably detect the target, but when the parachute is just thrown from the aircraft, the parachute is not fully opened, and the detection algorithm cannot detect the parachute. As shown in Figure 1a–d, the movement process of the parachute is divided into four stages: (1) The free-fall stage of the airdropped item leaving the aircraft (Figure 1a); (2) The stage of pulling out the main parachute from the umbrella bag (Figure 1b); (3) The stage of inflating the main parachute (Figure 1c); and (4) The stage of stable descent of the main parachute (Figure 1d). If we choose the parachute in the stable descent phase as the training sample, it is difficult for the detector to detect the parachute in the first three phases. Because the parachute is gradually opened, its violent shape change will cause the confidence score during the detection process to drop sharply, resulting in detection failure (the detection object is lost multiple times during the experiment). If a severely deformed object is detected in a continuous image sequence, the boundary box will be inaccurately positioned and severely jitter. Figure 2 shows that the jitter of the bounding box is extremely unfavorable for the subsequent analysis of parachute shape and aerodynamic layout.
In order to address these issues, we propose ODEM-GAN, an object deformation enhancement model based on a generative adversarial network (GAN) for object detection. ODEM-GAN has two branches: one is the CBAF branch and the other is the GAN branch. The CBAF branch is used for small- and scale variation object detection, which divides the feature map into several patches by spatial coordinates, and each of these patches has a size of m × m. These patches are used as the input of the contextual merging layer, and the output of the contextual merging layer, regarded as contextual feature-patches, is used for classification and regression, as shown in Figure 3. The GAN branch uses contextual feature patches to generate deformation features for enhancing the performance of classification and regression networks.
We highlight our main contributions as follows:
  • We propose ODEM-GAN, an effective object deformation enhancement model based on GAN, which focuses on the irrelevance of object deformation and scale variability to improve the prediction of hard sample probabilities and coordinates.
  • In ODEM-GAN, an online deformable sample creation module and an anchor-free box prediction layer are developed. The online deformable sample creation module is to deform and augment the dataset. The anchor-free box prediction layer not only avoids accurately aligning discrete spatial anchor boxes and ground-truth boxes but also reduces the amount of calculation.
  • We propose the positioning accuracy loss which reduces the bounding box’s inaccurate positioning and jitter caused by the detection of severely deformed objects in a continuous image sequence. The hard deformed samples generated by GAN can improve the regression accuracy of the object detector.
  • The experimental results on the airdrop dataset show that our ODEM module has obvious advantages in object detection with severe deformation. Moreover, the experimental results on MS-COCO show that the ODEM module also has advantages for small target detection, while improving the baseline of small object detection. In addition, the method also provides a new idea for object detection to replace object tracking algorithms.
The rest of this paper is organized as follows. Section 2 presents a review of related works. The main procedures and ODEM-GAN are studied in Section 3. Comparison performance experiments are given in Section 4. Finally, the conclusion is drawn in Section 5.

2. Related Works

Most object detection approaches are classified as one-stage methods and two-stage methods. The two-stage approach for object detection includes two steps. First, the extraction of multiple regions of interest (RoI) using the region proposal network. Second, these ROIs are classified and refined using the classification and regression module, respectively. Unlike the two-stage method, the one-stage method predicts the item immediately using a single neural network structure. The one-stage approach has a faster prediction speed than the two-stage method, but its accuracy is not as excellent. Among them, YOLO, RetinaNet, CornerNet, etc. are the most representative methods, and the contributions of these methods are described in detail in the next paragraph. Feature pyramids, also known as multi-level feature towers, are progressively becoming a frequent structure for object detection, since they handle object scale variances and produce considerable results. SSD [11] first predicts class scores and bounding boxes at different resolutions of feature maps so that objects of different sizes can be processed. FPN [14] and DSSD [5] proposed a top–down architecture with lateral connections which enhance low-level features with high-level semantic feature maps at all scales. In order to better adapt to the scale changes of anchors and to solve the limitations of anchor boxes, the anchor-free approach has attracted much attention in recent years. Among them, YOLO and CornerNet left a deep impression. CornerNet solved object detection as the key point detection problem, wherein the prediction box was obtained by detecting two key points in the upper left corner and the lower right corner.

2.1. Object Detection

Recent object detection approaches are either CNN approaches or transformer approaches. Using a transformer to complete visual tasks has become a new research direction to reduce the complexity of structures and explore scalability. The transformer performs better than CNNs in various computer vision tasks. Among them, the vision transformer (VIT) [23] is the pioneering work of the transformer in the field of computer vision. VIT divides the image into fixed-size patches, adds position embedding to the patch, preserves space/location information in the global scope through different strategies, and then inputs the linear projections of these patches together with their position codes into the transformer. Although these transformer-based approaches achieve good performance, they need to consume many computing resources and training data during the training process.
In the case of a small dataset, the performance of the CNN approach is equally impressive. The CNN approaches are divided into a two-stage approach and a one-stage approach. The one-stage method is currently a popular research direction. RetinaNet [13] mainly makes the one-stage network accuracy better than the two-stage network, and the focal loss is proposed to solve the problem of positive and negative sample imbalance. FPN [5] proposes a top–down architecture with lateral connections, which enhances low-level features with high-level semantic feature maps at all scales. DetNet [14] designs a new backbone network to maintain a higher spatial resolution in the upper layers to improve the accuracy of bounding box regression. However, they all use pre-set anchor boxes to regress the bounding boxes. With the recent development of the idea of anchor-free detection, YOLO [12,17,24] is impressive in balance speed and accuracy. DenseBox [25] first proposes a fully convolutional framework that directly predicted bounding boxes. CornerNet [26] solves object detection as the key point detection problem, and the prediction box is obtained by detecting two key points in the upper left corner and the lower right corner. Tang et al. proposed the CBAF model [22], which uses contextual information to effectively detect the parachute. Inspired by their work, we propose ODEM-GAN in this paper to improve the performance of CBAF in terms of the accuracy.

2.2. Generative Adversarial Networks

GAN is a powerful generative model based on deep learning. Since it was proposed by I. Goodfellow et al. in 2014 [27], GAN has been successfully applied in many fields, such as computer vision and natural language processing. In recent years, GANs are also unitized to tackle object detection problems. The main approaches are divided into two categories. The first is to create a spatial deformation by occluding certain regions on a feature map or manipulating feature maps to generate an adversary network of hard samples [28,29], as shown in Figure 4. The key idea is to create adversarial samples in the convolutional feature space instead of directly generating pixel-level data. The other is to generate super-resolution feature maps to reduce the difference in representation between the small and large objects for improving small object detection performance. Specifically, the generator learns to transform small object representations into super the resolution representations that are sufficiently similar to real large objects to deceive against the discriminator. Perceptual GAN [30], the model discriminator, and the generator are confronted to identify the generated representation and impose conditional requirements on the generator; the generated small object representation must be conducive to the detection of the target. The discriminator of the multitask generative adversarial network (MTGAN) [31] is a multitask network that uses true/false scores, object category scores, and bounding box regressors to describe each super-resolution image patch. Compared with these methods of directly generating pixel-level images and directly expanding the dataset, it is easier to generate samples in the convolutional feature space.
Therefore, inspired by the methods proposed by Wang [29] and Li [30], we used GAN networks to generate deformation samples. Unlike their methods, our method does not use masks or generate similar training samples. We generate deformation samples to achieve the detection of objects with dramatic deformation. For the problem of the accuracy of the one-stage method not being as good as that of the two-stage method, we used focal loss [13] to solve it. The Vit [23] and YOLO [12] algorithms achieve the establishment of object contextual relations by dividing the feature map, thus improving the accuracy of small target detection. Vit is a transformer-based detection network as it requires a large amount of training data, while YOLO is not as accurate as the two-stage network, although its prediction is fast. In this paper, we used object contextual relations to improve the accuracy of small object detection, then propose the ODEM-GAN module.

3. Approach

In this section, we describe the design process of the ODEM-GAN module, and show how to apply it to the backbone. We instantiate our ODEM-GAN module by using the state-of-the-art RetinaNet as the baseline. The design of ODEM-GAN is described from the following aspects: (1) how to choose hard samples for GAN; (2) how to design the GAN generator; (3) how to design the GAN discriminator; and (4) how to train and test the ODEM module and the branch of GAN.

3.1. Overview

To facilitate the demonstration, we use the 512 × 512 resolution of the image as input, and only P F 3 (the third of feature pyramid) of CBAF [22] is introduced; the other levels are the same. The resolution of P F 3 is 64 × 64 and the channel is c = 256 . Therefore, the shape of P F 3 is [ b , 64 , 64 , 256 ] (b is the batch size). As shown in the black-dashed region of Figure 3, the feature map division layer takes the feature map of P F 3 as input and divides the feature map of P F 3 into N f p feature patches of size n n according to the space coordinates. We set n = 8 . Therefore, the shape of a feature patch is [ b , n , n , c ] = [ b , 8 , 8 , 256 ] , the number of feature patches for the P F 3 level is N f p = h 3 n × w 3 n = 64 8 × 64 8 = 64 , and the output of the feature map division layer is [ b , N f p , n , n , c ] = [ b , 64 , 8 , 8 , 256 ] . After obtaining the output of the feature map division layer, the feature patches are used as input for the contextual merger layer to concatenate with the surrounding feature patches, as shown in the red-dashed region of Figure 3. We used a m m = 9 (we set m = 3 ) sliding window to obtain the contextual information of each feature patch; second, we concatenated the m m feature patches obtained in the sliding window along their channel dimensions. The shape of the merged feature patch is [ b , n , n , c × m m ] ; finally, the merged feature pieces are passed to a 3 × 3 convolutional layer with c f = 512 filters. The output shape of the contextual merger layer is [ b × N f p , n , n , c f ] = [ b × 64 , 8 , 8 , 512 ] .
A GAN is composed of a generator network and a discriminator network, and its structural diagram is shown in Figure 5 [27]. The generator network (abbreviated as G) uses a given noise vector z (generally subject to uniform distribution or normal distribution) to generate data. The discriminator network (abbreviated as D) is used to discriminate whether it is generated data or real data. The former tries to generate more real data, while the latter tries to perfectly distinguish between real data and generated data. The two learn from each other in the adversarial process and make progress together until the data generated by G become perfect so that D cannot distinguish. We can consider D as a binary classifier, so the basic model of GAN can be represented by cross entropy. The formula is shown in Equation (1):
m i n G   m a x   D V D , G = E x p d a t a ( x ) l o g D ( x ) + E z p z ( z ) l o g ( 1 D G z ) ,
where x is the real data, l o g D ( x ) represents the probability of the discriminator network on the real data, G ( z ) represents the data generated by G, and l o g ( 1 D ( G ( z ) ) ) represents the probability of D on the generated data. Through the max-min game, G and D are trained alternately until D cannot distinguish between the real and fake data.
Similar to traditional GAN, we propose that ODEM-GAN also consists of G and D, which compete by adversarial loss and alternating training. However, the difference is that this paper adopts conditional GAN (conditional GAN, referred to as C-GAN); that is, the input of G not only has a noise vector z but also a given condition. Adding conditions to G cannot only generate samples under the conditions (specify the desired object that hard deformation samples are generated) but also reduce the training difficulty of GAN. For the overall structure of ODEM-GAN, this paper adopts a structure similar to that of BE-GAN [32]. The structural feature of BE-GAN is that D adopts the structure of the autoencoder, which is different from the data distribution of the learning samples of other GAN models. BE-GAN learns the distribution of the loss of the autoencoder through the Wasserstein adversarial loss.
As shown in the blue-dashed region of Figure 6, the ODEM-GAN module includes a CBAF module, a GAN branch, and a classification and regression module for each level of the feature pyramid.

3.2. Network Architecture of the ODEM-GAN Module

3.2.1. ODEM-GAN Module

Figure 7 shows the GAN branch in the red-dashed box includes an online deformable sample creation module, a generator G, and a discriminator D. Among them, the online deformable sample creation module is used to choose and create the deformation samples that are difficult to predict by the classification and regression module (which refers to the hard sample feature of the deformation) and uses the deformation samples as the input of D. G is used to generate diverse deformation feature patches and add them to the merged feature patches with context information, which is used as the input of the classification and regression module so that the detector can learn more features of the deformed samples and improve the purpose of the detector to stably recognize the deformed samples. D is used to compete with G so that G can generate a variety of deformed samples. Note that the GAN branch only participates in training. Note that GAN can only generate one certain category of object.
Online deformable sample creation module. To obtain more hard samples, the most direct method is the horizontal and vertical flipping of existing hard samples. Therefore, this module proposes an online creation module for deformed samples, as shown in Figure 8. In Section 3.1, we obtain the current feature patches and contextual feature patches and the flip operation on the current feature patches and their contextual feature patches to obtain the new deformation samples. Specifically, if the sliding window contains the feature patches of hard samples, the online deformable sample creation module flips contextual feature patches horizontally, vertically, on the left diagonal and right diagonal in the sliding window (see the red-dashed box on the feature map in Figure 8). However, the current feature patches in the sliding window do not contain the feature patches of hard samples, the contextual feature patches are filled with 0, and the flip operation is not performed.
The flip operation obtains four new hard sample deformation features (we call the current feature patch and its contextual feature patches features) which are used as the input of the contextual merger layer to obtain new merged feature patches. Then, the new merged feature patches are used as the input of the classification and regression modules, and the sum of their classification loss and regression loss is calculated. What the detector needs is to make it difficult to classify and regress samples, and the purpose of improving robustness and detection accuracy by learning these hard samples. The greater the loss of the sample, the more difficult it is for the detector to predict it. Therefore, the merged feature patches corresponding to the maximum loss are selected as the hard samples. Through the above steps, we can complete the online creation of deformation samples.
The design of generatorG. ODEM-GAN is built on each level of the pyramid feature, and the parameters of each GAN branch are not shared. For ease of presentation, only the ODEM-GAN module of P F 3 is introduced, and the others are the same. As shown in Figure 9, the input of G is the noise vector z and the current feature patch, where z can increase the diversity of the generated samples, and the current feature patch is used as the condition of the generator, which can ensure that the expected sample is generated or can reduce the difficulty of training. Specifically, z is a 64-dimensional vector that is a uniform distribution of [ 1 , 1 ] . We take the average of the feature patches along the channel dimension; then, the shape of the feature patches on P F 3 will change from n , n , c = [ 8 , 8 , 256 ] to [ 8 , 8 , 1 ] and flatten it into a 64-dimensional vector. Finally, the 64-dimensional vector is concatenated with z to obtain a new 128-dimensional vector as the input of the generator. The 128-dimensional vector achieves a shape of feature patches [ n , n , c f ] = [ 8 , 8 , 512 ] by upsampling (the upsampling strategy uses nearest neighbor interpolation) and three convolution modules.
As shown in Figure 9, the number of channels of the convolutional layer is the same as the multiple of the resolution increase. There are two upsamplings, and n c = c f 2 2 = 128 in Figure 9. The parameters of generator G are shown in Table 1.
The design of discriminatorD. As shown in Figure 10, the discriminator adopts an autoencoder structure. It consists of two parts: an encoder and a decoder. The encoder encodes the input (the samples created by the generator and online deformable sample creation module) into a set of multidimensional feature vectors, while the function of the decoder is the opposite of the encoder. This restores the multidimensional feature vectors generated by the encoder to the shape of the original input. Therefore, the output dimension of the autoencoder is the same as its input dimension. The decoding structure used in this article is the same as the generator G. The encoder is the inverse process of the decoder. The parameters of the discriminator D are shown in Table 2.

3.2.2. Loss of the GAN Branch

Since L 2 loss will lead to blurred sample generation, L 1 loss is used for the autoencoder, as shown in Equation (2):
L ( x ) = x D ( x ) 1 ,
where L ( x ) is the loss of the autoencoder, D ( x ) is the output of the autoencoder, and x is the input of the autoencoder. The loss of generator G is shown in Equation (3):
L G = L ( x s ) ,
where x s is the sample generated by generator G. The loss of the discriminator is shown in Equation (4):
L D = L x r L ( x s ) ,
where x r are samples created by the online deformable sample creation module. To enhance the diversity of the samples generated, a coefficient of balance k t balances the generator and discriminator in each alternate training. The update of k t is shown in Equation (5):
k t + 1 = k t + ζ ( η L x r L ( x s ) ) ,
where ζ is the learning rate of k t , and the parameter η can control the proportion of sample diversity. In this article, we set ζ = 0.01 and η = 0.5 . After adding k t , the loss function of the discriminator L D becomes the following Equation (6):
L D = L x r k t L ( x s ) .
Although generator G can generate sufficiently realistic hard samples by minimizing Equations (2) and (5), to make the generated hard samples help improve the regression accuracy of the detector, we define that positioning accuracy loss is introduced here. The positioning accuracy loss L l p is as shown in (7):
L l p = | t i , j t i , j g | 2 0.5 0.9 , | t i , j t i , j g | < 1 / 9 | t i , j t i , j g | 0.5 0.9 , e l s e .
The hard sample generated by generator G is taken as the input of the detector regression module, and after the forward propagation of the regression module, the 4-dimensional offset t i , j g = [ t t i , j g , t l i , j g , t b i , j g , t r i , j g ] of the object bounding box will be output. Then, the smooth L 1 loss of t i , j g = [ t t i , j g , t l i , j g , t b i , j g , t r i , j g ] and the 4-dimensional offset t i , j of the ground-truth box are calculated.
The purpose of the positioning accuracy loss is to make the hard samples generated by the generator improve the regression accuracy of the object detector. Therefore, L l p is added to the loss of generator G. The loss function of generator L G becomes the following Equation (8):
L G = L x s + λ l p L l p ,
where λ l p can control the weight of L x s and L l p . In this article, we set λ l p = 0.1 . By minimizing Equations (6) and (8) and alternately training the discriminator D and generator G, the generator can generate sufficiently realistic hard samples.

3.3. Ground Truth and Loss of the Detector

Since the GAN branch does not participate in the backpropagation of the detector, nor does it participate in inference, the loss of the GAN branch is not included in the loss of the detector.

3.3.1. Classification Modules

The ODEM-GAN module predicts objects based on the anchor-free module. This approach is similar to labeling for image segmentation. Pixels inside the bounding box are regarded as positive samples, pixels outside the bounding box are regarded as negative samples, and these pixels are filled with 0. The classification output is a feature map of K channels, and each channel represents a class rather than an N × K dimensional probability (N is the number of anchor boxes). The classification module consists of four 3 × 3 convolutional layers. Each convolutional layer has 256 filters, and ReLU is added after each convolutional layer. The output layer consists of a 3 × 3 convolutional layer with K filters and sigmoid activation functions. We used focal loss [13] with hyperparameters α t = 0.25 and γ = 2:0 as the total classification loss for the ODEM-GAN module, as shown in Equation (9):
L c l s = α t ( 1 p t ) γ l o g ( p t ) ,
where p t = p , y = 1 1 p , e l s e , and p is the output of the classification module.

3.3.2. Regression Modules

The structure of the regression module is the same as that of the classification module. The difference is that the output layer is a 3 × 3 convolutional layer with four filters. The regression module outputs four offset maps of the ground truth and the bounding box. The distance from each pixel to the top, left, bottom, and right of the bounding box is represented by a 4-dimensional vector t i , j = [ t t i , j , t l i , j , t b i , j , t r i , j ] . The regression loss is the same as RetinaNet, using smooth L 1 loss, as shown in Equation (10).
L r e g = t i , j t i , j 2 0.5 0.9 , t i , j t i , j < 1 / 9 t i , j t i , j 0.5 0.9 , e l s e .
where t i , j is the 4-dimensional offset of the output of the bounding box.
Note that the parameters for classification are shared across each pyramid level, and the parameters for the regression module are also shared across each pyramid level. However, the parameters for classification and regression are independent. The architectures of the classification and regression modules are shown in Table 3.
The total loss of the ODEM-GAN is shown in Equation (11).
L O D E M G A N = 1 N c l s l = 3 7 i , j L c l s + 1 N r e g l = 3 7 i , j L r e g ,
where N c l s is the total of the positive and negative samples and N r e g is the total of positive samples.

3.4. Joint Inference and Training

Our ODEM-GAN module can work with the original RetinaNet, and all hyperparameters in training and inference are the same as those in RetinaNet. Note that the GAN branch of the ODEM-GAN module only participates in training and does not participate in inference.
Model pretraining: Experiments show that it is necessary to pretrain ODEM-GAN before using ODEM-GAN to improve the performance of the detector. The specific pretraining process is as follows:
  • The training dataset: a dataset that only contains specific hard deformation objects that need to be detected;
  • Use the trained CBAF object detection model to extract the features of specific hard deformation objects and use the online deformable sample creation module to create various hard deformation object samples. Note: During the pretraining process of ODEM-GAN, all parameters of the CBAF object detection model are frozen and not updated;
  • Minimize Formulas (6) and (8) to alternately train the discriminator and the generator (note: update the discriminator parameters first, then update the generator parameters) until the deformable samples generated by the generator are enough to be real. Throughout the pretraining process, ODEM-GAN is trained for 30 epochs using a dataset containing only specific hard deformation objects.
    The purpose of pretraining is to allow the ODEM-GAN module to have a preliminary understanding of specific hard deformation object samples and reduce the difficulty of joint training of the ODEM-GAN module and CBAF object detection model.
Joint pretraining:
  • The training set of joint training includes two parts: a dataset that only contains specific hard deformation object samples and a dataset that does not contain specific hard deformation object samples;
  • Unfreeze all parameters of the CBAF object detection model and update the parameters;
  • In each iteration, two images were taken from the datasets containing only specific hard deformation object samples and those without as the input of the network, so batch size = 4;
  • After using the detector to extract the feature patches of the image containing specific hard deformation object samples and fuse the context information, use the training method of the pretraining process to update the parameters of the ODEM-GAN;
  • Use the generator of ODEM-GAN to generate deformable samples of specific objects;
  • Use the detector to extract feature patches that do not contain specific hard deformation object images and fuse context information;
  • Place the fused feature patches of the specific hard deformation object images extracted by the detector, the fused contextual feature patches that do not contain the specific hard deformation object images, and the specific hard deformation object samples generated by ODEM-GAN together as a batch. Then, they are taken together as input to the classification and regression modules, and the optimizer is directly used to update the network parameters of the CBAF object detection model. Note that if the other specific category is changed, the GAN branch must be retrained.
Inference: The GAN branch of the ODEM-GAN module only participates in training and does not participate in prediction. Its prediction process is the same as that of the CBAF object detection model. Therefore, there is no additional calculation during the prediction, and the running speed of the detection algorithm will not be affected.
We only decoded box predictions whose confidence score threshold is higher than 0.05 and only selected up to at most 1k top-scoring in each pyramid level. The box predictions from each pyramid level based on the anchor boxes branch and the anchor-free box branch are merged together; then, non-maximum suppression is performed with a threshold of 0.5 to produce the final detection result.
Initialization: We use ImageNet1k [33] to pre-train the backbone network. For convolutional layers in our CBAF module, the classification and regression layers are initialized as in [13]. For the initialization of the convolutional layer and the fully connected layer in the generator and discriminator, this paper uses “Xavier” initialization [34].
Optimization: The loss of the whole network consists of the original RetinaNet loss and the ODEM-GAN loss, as shown in Equation (12):
L = L r e t + λ L O D E M G A N ,
where L r e t is the loss of RetinaNet and λ controls the weight of the ODEM-GAN module. In this article, we set λ = 0.5 . A stochastic gradient descent (SGD) optimizer is used to train the whole network. The entire network is trained on two GPUs, and the batch size is four. For the CBAF module, the initial learning rate of training is 0.01, which is divided by 10 at 10 K, 60 K and 90 K iterations. For the GAN branch, we set the weight decay to 0.0005 and the momentum to 0.9. For both the generator and the discriminator, the learning rate is set to 0.00008, and the learning rate decays dynamically, decaying to 0.00002 during 50 K iterations. The batch size is set to four.

4. Experiments

We experiment with the ODEM-GAN module on the MS-COCO dataset [2], our airdrop experiment dataset and VOC 2012. The training date of MS-COCO is ‘trainval35k’, which includes 80 K images from the train and 35 K images from the 40 k val. The airdrop dataset is made up of two parts: (1) the matching image that was scraped from the Internet using a web crawler approach; (2) relevant images that were taken during the actual airdrop experiments. Human faces, quadrotor UAVs, fixed-wing aircraft, and parachutes are among the categories in the airdrop dataset. Faces and quadrotor UAVs are among the objects introduced to help with system debugging because commissioning a system with an air-drop experiment is inefficient and expensive. Faces and quadrotor drones, which are ubiquitous targets in life, may be used to decrease costs and make debugging the system more convenient and adaptable. Take 9/10 of our dataset as the training set (20,373 images), and the remaining 1/10 as the test set (2264 images). Part of the training dataset (we cannot release the images collected during the airdrop experiments without the permission of the project funding provider, but we can release the images from the Internet) is available from https://pan.baidu.com/s/1nsZAWnVy-3rrVMZypNdnGg (accessed on 25 March 2022). We evaluated our module by ablation studies on our dataset. The experimental comparison with state-of-the-art methods was carried out on “test-dev” and our dataset, and the relevant AP value was calculated. To easily show the performance of each category, we calculated the AP value of each category on VOC 2012.

4.1. Ablation Studies

We used our airdrop experiment dataset in the ablation studies. In the airdrop experiment, the parachute had severe deformation, so we set the parachute as the specific hard deformation object. The resolution of the image is scaled to 800 × 800 for training and testing in ablation experiments, and we use ResNet-50 as the backbone network and all networks based on the CBAF module [22]. We mainly evaluated the impact of adding a conditional generator and the positioning accuracy loss on the detector.
When using the current feature patch as the condition of the generator and the generator does not add the condition, after using ODEM-GAN to enhance the training of the CBAF detection model, the CBAF detection model only has an AP score of 76.4% for parachute detection on the airdrop dataset (third row in Table 4), which is lower than the result of augmentation training without ODEM-GAN (second row in Table 4). When conditions are added to the generator, the AP score after enhanced training for the CBAF object detection model is 85.1% (the fourth row in Table 4), an increase of 0.3%. The experimental results show that the current feature patch must be used as the input of the generator to improve the parachute detection performance of the CBAF object detection model. The reason for this situation is analyzed in this paper: if conditions are not added to the generator, the hard samples generated by ODEM-GAN have no directionality, which will interfere with the learning of the classification and regression modules, resulting in a decrease in accuracy. In contrast, after the generator adds conditions, the generated hard deformation parachute samples are under specific conditions, have clear directionality, and can enhance the detection ability of the classification and regression modules.
When the loss of positioning accuracy is added, after using ODEM-GAN to enhance the training of the CBAF model, the AP score of the CBAF model for parachute detection on the airdrop dataset is 88.4% (the fifth row in Table 4). The AP score increased by 3.6%. The experimental results show that when the generator adds the positioning accuracy loss, the generator will generate parachute deformation samples with clear position information and can effectively reduce the impact of bounding box jitter on the regression accuracy in consecutive frames. We use the intersection-over-union (IOU) of two adjacent frames as the evaluation criterion for bounding box jitter. If the size and position of the bounding boxes in the adjacent two frames of images change little, the overlap ratio of the bounding boxes in the two frames is high; otherwise, it is small. The experimental results of the bounding box IOU in adjacent frames are shown in Figure 11a,b, respectively. The experimental results show that the IOU in adjacent frames of the model without ODEM-GAN is only approximately 0.2; after the ODEM-GAN module is added, the IOU in adjacent frames is increased to approximately 0.78, and the jitter is effectively suppressed. After adding the deformation enhancement model, the detection effect of the detector on the parachute in the open parachute state is visualized, as shown in Figure 12a,b. Therefore, using these samples to enhance the training of the regression module can further improve the accuracy of the detector on the object. The comprehensive experimental results show that using ODEM-GAN to enhance the training of the CBAF model can effectively improve the robustness of the CBAF model to deformation parachute detection.
The experimental results are shown in Figure 13. We found that there are both guided parachutes (the smaller parachute in Figure 13) and the main parachute which is not fully opened. The original RetinaNet can only detect the larger parachute (the first row); our proposed CBAF model is for the small and deformable guided parachute detection performance is not robust enough (the second row, the second column, the third column); and the ODEM-GAN model can detect the guide parachute and the main parachute well.
The data shown in Table 4 are the AP score of the parachute detection by the CBAF target detection model on the airdrop dataset after using ODEM-GAN to enhance the training of the CBAF target detection model. Note: The data shown in the second row in Table 4 are the AP score for parachute detection by the CBAF target detection model on the airdrop data.

4.2. Determination of Hyperparameters

For the determination of the weight λ l p in Formula (8), we used different λ l p values to jointly train the CBAF module and ODEM-GAN module for training on our airdrop dataset, and the backbone is ResNet-50. The AP scores for parachute detection are shown in Table 5 and Table 6. The experimental results show that when λ l p = 0.1 , the ODEM-GAN module achieves the best results for parachute detection.

4.3. Results on PASCAL VOC 2012

A GAN can only generate one certain category of object. To easily show the performance of the ODEM-GAN module for each category, the PASCAL VOC 2012 dataset has 20 categories, and we optimized each category and calculated the AP value of each category on VOC 2012. We show the results of the original RetinaNet-50, the CBAF module, and our ODEM-GAN module on VOC 2012 in Table 7. We found that the AP score increased by 3% compared with ResNet-50, and there are varying degrees of improvement in AP scores for each category, which proves that our model can improve the performance of the detector very well.

4.4. Comparison to State of the Art

We compare the ODEM-GAN module with other state-of-the-art detectors [32,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,50] on the test-dev split. Through the experimental results, we found that the detector based on ODEM-GAN achieved 27.9% in APS, which is 2% better than our previously proposed CBAF model, as shown in Table 8, indicating that our model not only enhances the deformation target detection, but also improves the performance of small object detection compared with state-of-the-art detectors [29,39,41,42,51]. We visualized some experimental results, as shown in Figure 14, and we performed both deformation and small object detection with good performance; in Table 8, we previously proposed CBAF with no improvement in detecting larger objects, which is because the global features of large objects are destroyed due to the feature maps divide to feature patches in the feature map division layer, but on the ODEM-GAN model. There is a 1.5% improvement, which is due to the invariance of the samples generated by the GAN branch learned by the detector.
In addition, DETR [41], DINO [43], GLIP [44], and DyHead [45] in Table 8 are all transformer-based object detection models, in which DINO and DyHead are based on the Swin-Transformer, although these algorithms implement higher general detection accuracy, because the transformer encoding method has a larger receptive field, which has a similar idea to our feature map division layer but needs to consume more training samples and consume more computing resources, resulting in lower operating rates. Although the QueryInst [46] algorithm has excellent performance in detection tasks, it is mainly used for instance segmentation, which is a query-based instance segmentation method. YOLOR-D6 [47] proposed a method to use semantic information in upper layers of CNNs to represent implicit knowledge and combine multiple discriminators for object detection, which has excellent performance and real-time performance. Comparing YOLOR-D6 and DINO, YOLOR-D6 makes full use of implicit knowledge, and DINO uses the transformer as a backbone, and its AP scores have been greatly improved, which is an improvement for ODEM-GAN as it provides a new direction for the future.

5. Conclusions

In the continuous image sequence, the CBAF detection model will have the problem of bounding box jitter in the continuous detection of the object during the deformation process. This paper proposes ODEM-GAN to enhance the training of the CBAF detection model and improve the invariance of deformation for object detection. ODEM-GAN consists of three parts: an online deformable sample creation module, a generator, and a discriminator. The online deformable sample creation module is mainly used to create object deformation samples, the generator is used to generate diverse object deformation samples, and the discriminator is used to compete with the generator to ensure that the samples generated by the generator are sufficiently fake. In addition, based on the adversarial loss function of GAN, a positioning accuracy loss function is proposed, which effectively improves the regression accuracy of the detector. Through comprehensive comparative experiments, it is verified that the proposed ODEM-GAN can enhance the pose and deformation irrelevance of the CBAF detection model for parachute detection and effectively suppress the phenomenon of bounding box jitter.
Although our algorithm has obvious advantages for specific tasks, it has certain limitations in the training process, the training complexity is relatively high, and if the specific optimized objects are changed in the training of the process, it needs to be retrained, which is a limitation caused by the fact that GAN networks can only generate one category of specific objects. At the same time, because GAN branches are added to each level of the feature pyramid resulting in a large number of parameters, the training time will be relatively long. To address the above limitations, our future improvement can be based on two aspects: (1) improving the GAN network so that it can optimize multiple categories’ objects to improve the robustness of the classifier; and (2) optimizing the network structure so that the output of each layer of the feature pyramid is connected to a unified GAN branch to reduce the complexity of training and shorten the training time.

Author Contributions

Conceptualization, Z.Z. and Z.P.; data collection, Z.T.; analysis and interpretation of results, Z.Z. and F.G.; validation, Z.Z. and F.G.; writing—original draft preparation, Z.Z. and F.G.; writing—review and editing, F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the China Postdoctoral Science Foundation (2020M671597), Jiangsu Postdoctoral Research Foundation (2020Z100), the National Science Foundation of the Jiangsu Higher Education Institutions of China (20KJB520002), Suzhou Planning Project of Science and Technology (No. SYG202024), and the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Everingham, M.; Zisserman, A.; Williams, C.K.; Van Gool, L.; Allan, M.; Bishop, C.M.; Chapelle, O.; Dalal, N.; Deselaers, T.; Dorkó, G.; et al. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results. 2008. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.230.2543 (accessed on 25 March 2022).
  2. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  3. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv 2013, arXiv:1312.6229. [Google Scholar]
  4. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  5. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. Dssd: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  6. Hariharan, B.; Arbeláez, P.; Girshick, R.; Malik, J. Hypercolumns for object segmentation and fine-grained localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 447–456. [Google Scholar]
  7. Fischer, P.; Dosovitskiy, A.; Brox, T. Descriptor matching with convolutional neural networks: A comparison to sift. arXiv 2014, arXiv:1405.5769. [Google Scholar]
  8. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  11. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  13. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  14. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  15. Li, Z.; Peng, C.; Yu, G.; Zhang, X.; Deng, Y.; Sun, J. Detnet: A backbone network for object detection. arXiv 2018, arXiv:1804.06215. [Google Scholar]
  16. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  17. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  18. Zhu, C.; He, Y.; Savvides, M. Feature selective anchor-free module for single-shot object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 840–849. [Google Scholar]
  19. Long, J.L.; Zhang, N.; Darrell, T. Do convnets learn correspondence? Adv. Neural Inf. Process. Syst. 2014, 27, 1–9. [Google Scholar]
  20. Shrivastava, A.; Gupta, A.; Girshick, R. Training region-based object detectors with online hard example mining. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 761–769. [Google Scholar]
  21. Takác, M.; Bijral, A.; Richtárik, P.; Srebro, N. Mini-batch primal and dual methods for SVMs. In Proceedings of the International Conference on Machine Learning (PMLR), Atlanta, GA, USA, 17–19 June 2013; pp. 1022–1030. [Google Scholar]
  22. Tang, Z.; Yang, J.; Pei, Z.; Song, X. Coordinate-based anchor-free module for object detection. Appl. Intell. 2021, 51, 9066–9080. [Google Scholar] [CrossRef]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  24. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  25. Huang, L.; Yang, Y.; Deng, Y.; Yu, Y. Densebox: Unifying landmark localization with end to end object detection. arXiv 2015, arXiv:1509.04874. [Google Scholar]
  26. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  27. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 1–9. [Google Scholar]
  28. Chen, Y.; Song, L.; Hu, Y.; He, R. Adversarial occlusion-aware face detection. In Proceedings of the 2018 IEEE 9th International Conference on Biometrics Theory, Applications and Systems (BTAS), Redondo Beach, CA, USA, 22–25 October 2018; pp. 1–9. [Google Scholar]
  29. Wang, X.; Shrivastava, A.; Gupta, A. A-fast-rcnn: Hard positive generation via adversary for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2606–2615. [Google Scholar]
  30. Li, J.; Liang, X.; Wei, Y.; Xu, T.; Feng, J.; Yan, S. Perceptual generative adversarial networks for small object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1222–1230. [Google Scholar]
  31. Bai, Y.; Zhang, Y.; Ding, M.; Ghanem, B. Sod-mtgan: Small object detection via multi-task generative adversarial network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 206–221. [Google Scholar]
  32. Berthelot, D.; Schumm, T.; Metz, L. Began: Boundary equilibrium generative adversarial networks. arXiv 2017, arXiv:1703.10717. [Google Scholar]
  33. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  34. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  35. Bell, S.; Zitnick, C.L.; Bala, K.; Girshick, R. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2874–2883. [Google Scholar]
  36. Vinukonda, P. A Study of the Scale-Invariant Feature Transform on a Parallel Pipeline. Master’s Thesis, Louisiana State University, Baton Rouge, LA, USA, 2011. [Google Scholar]
  37. Yang, Z.; Xu, Y.; Xue, H.; Zhang, Z.; Urtasun, R.; Wang, L.; Lin, S.; Hu, H. Dense reppoints: Representing visual objects with dense point sets. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 227–244. [Google Scholar]
  38. Duan, K.; Xie, L.; Qi, H.; Bai, S.; Huang, Q.; Tian, Q. Corner proposal network for anchor-free, two-stage object detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 399–416. [Google Scholar]
  39. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  40. Yao, L.; Xu, H.; Zhang, W.; Liang, X.; Li, Z. SM-NAS: Structural-to-modular neural architecture search for object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12661–12668. [Google Scholar]
  41. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  42. Wang, S.; Gong, Y.; Xing, J.; Huang, L.; Huang, C.; Hu, W. RDSNet: A new deep architecture forreciprocal object detection and instance segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12208–12215. [Google Scholar]
  43. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  44. Li, L.H.; Zhang, P.; Zhang, H.; Yang, J.; Li, C.; Zhong, Y.; Wang, L.; Yuan, L.; Zhang, L.; Hwang, J.N.; et al. Grounded Language-Image Pre-training. arXiv 2021, arXiv:2112.03857. [Google Scholar]
  45. Dai, X.; Chen, Y.; Xiao, B.; Chen, D.; Liu, M.; Yuan, L.; Zhang, L. Dynamic head: Unifying object detection heads with attentions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7373–7382. [Google Scholar]
  46. Fang, Y.; Yang, S.; Wang, X.; Li, Y.; Fang, C.; Shan, Y.; Feng, B.; Liu, W. Instances as queries. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 6910–6919. [Google Scholar]
  47. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. You only learn one representation: Unified network for multiple tasks. arXiv 2021, arXiv:2105.04206. [Google Scholar]
  48. Lu, X.; Li, B.; Yue, Y.; Li, Q.; Yan, J. Grid r-cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7363–7372. [Google Scholar]
  49. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 6569–6578. [Google Scholar]
  50. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. Foveabox: Beyound anchor-based object detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  51. Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P. Brief: Binary robust independent elementary features. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 778–792. [Google Scholar]
Figure 1. The process of parachute movement.
Figure 1. The process of parachute movement.
Applsci 12 04609 g001
Figure 2. The detection of the continuous image sequence (timing: from left to right, top to bottom).
Figure 2. The detection of the continuous image sequence (timing: from left to right, top to bottom).
Applsci 12 04609 g002
Figure 3. General concept of the proposed CBAF module.
Figure 3. General concept of the proposed CBAF module.
Applsci 12 04609 g003
Figure 4. Overview of GANs-based object detection.
Figure 4. Overview of GANs-based object detection.
Applsci 12 04609 g004
Figure 5. The process of generating an adversary network.
Figure 5. The process of generating an adversary network.
Applsci 12 04609 g005
Figure 6. Overview of network architecture.
Figure 6. Overview of network architecture.
Applsci 12 04609 g006
Figure 7. ODEM-GAN module includes a CBAF module (in the green-dashed box), a GAN branch (in the red-dashed box).
Figure 7. ODEM-GAN module includes a CBAF module (in the green-dashed box), a GAN branch (in the red-dashed box).
Applsci 12 04609 g007
Figure 8. Online deformable sample creation module. Each sample passed through the CBAF module to compute the focal loss and smooth L 1 loss. Then, the level with the largest sum of the two losses is selected as the supervision signal for hard samples.
Figure 8. Online deformable sample creation module. Each sample passed through the CBAF module to compute the focal loss and smooth L 1 loss. Then, the level with the largest sum of the two losses is selected as the supervision signal for hard samples.
Applsci 12 04609 g008
Figure 9. The structure of generator G.
Figure 9. The structure of generator G.
Applsci 12 04609 g009
Figure 10. The structure of discriminator D.
Figure 10. The structure of discriminator D.
Applsci 12 04609 g010
Figure 11. The IOU in adjacent frames: (a) the result of the CBAF module is only approximately 0.2; (b) the result of the ODEM-GAN module is increased to approximately 0.78, and the jitter is effectively suppressed.
Figure 11. The IOU in adjacent frames: (a) the result of the CBAF module is only approximately 0.2; (b) the result of the ODEM-GAN module is increased to approximately 0.78, and the jitter is effectively suppressed.
Applsci 12 04609 g011
Figure 12. Visualization of the IOU in adjacent frames result: (a) the RetinaNet with CBAF module; and (b) the RetinaNet with ODEM-GAN module.
Figure 12. Visualization of the IOU in adjacent frames result: (a) the RetinaNet with CBAF module; and (b) the RetinaNet with ODEM-GAN module.
Applsci 12 04609 g012
Figure 13. Comparison between our detector and the original RetinaNet. From the first row to the last row, the detection results of the original RetinaNet, the RetinaNet with the CBAF module, and the RetinaNet with the ODEM-GAN module are represented, respectively.
Figure 13. Comparison between our detector and the original RetinaNet. From the first row to the last row, the detection results of the original RetinaNet, the RetinaNet with the CBAF module, and the RetinaNet with the ODEM-GAN module are represented, respectively.
Applsci 12 04609 g013
Figure 14. Comparison between our improved detector and recent state-of-the-art detectors. From the first row to the last row, the detection results of YOLOv4, EfficientDet-2, DETR, RDSNet, RetinaNet and ODEM-GAN, respectively.
Figure 14. Comparison between our improved detector and recent state-of-the-art detectors. From the first row to the last row, the detection results of YOLOv4, EfficientDet-2, DETR, RDSNet, RetinaNet and ODEM-GAN, respectively.
Applsci 12 04609 g014
Table 1. The parameters of generator G.
Table 1. The parameters of generator G.
NameOutput Size
(Height, Width, Channel)
Act.Layer
(Filter_h, Filter_w, Num, Stride)
feature-patch + z(128, -, -)--
full connect layer(2 × 2 × n c , -, -)--
module1conv1(2, 2, n c )Relu(3, 3, n c , 1)
conv2(2, 2, n c )Relu(3, 3, n c , 1)
module2conv1(4, 4, 2 n c )Relu(3, 3, 2 n c , 1)
conv2(4, 4, 2 n c )Relu(3, 3, 2 n c , 1)
module3conv1(8, 8, 2 2 n c )Relu(3, 3, 2 2 n c , 1)
conv2(8, 8, 2 2 n c )-(3, 3, 2 2 n c , 1)
Table 2. The parameters of the discriminator D.
Table 2. The parameters of the discriminator D.
NameOutput Size
(Height, Width, Channel)
Act.Layer
(Filter_h, Filter_w, Num, Stride)
x s / x r (8, 8, 22 n c )Provided by generator and online deformable sample creation module
en-module1conv1(8, 8, 2 n c )Relu(3, 3, 2 n c , 1)
conv2(4, 4, 2 n c )Relu(3, 3, 2 n c , 2)
en-module2conv1(4, 4, n c )Relu(3, 3, n c , 1)
conv2(2, 2, n c )Relu(3, 3, n c , 2)
full connect layer1( 2 × 2 × n c , -, -)--
full connect layer1( 2 × 2 × n c , -, -)--
conv1(2, 2, n c )Relu(3, 3, n c , 1)
conv2(2, 2, n c )Relu(3, 3, n c , 1)
conv1(4, 4, 2 n c )Relu(3, 3, 2 n c , 1)
conv2(4, 4, 2 n c )Relu(3, 3, 2 n c , 1)
conv1(8, 8, 2 2 n c )Relu(3, 3, 2 2 n c , 1)
conv2(8, 8, 2 2 n c )-(3, 3, 2 2 n c , 1)
Table 3. Only PF3 is introduced, similarly to in the other levels.
Table 3. Only PF3 is introduced, similarly to in the other levels.
NameOutput Size
(Height, Width, Channel)
Act.Layer
(Filter_h, Filter_w, Num, Stride, Padding)
classification moduleclass_conv1(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
class_conv2(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
class_conv3(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
class_conv4(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
class_output(8, 8, K) × 64Sigmoid(3, 3, K, 1, 1)
regression moduleregress_conv1(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
regress_conv2(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
regress_conv3(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
regress_conv4(8, 8, 256) × 64ReLU(3, 3, 256, 1, 1)
regress_output(8, 8, K) × 64-(3, 3, 4, 1, 1)
Table 4. ODEM-GAN ablation experiment results.
Table 4. ODEM-GAN ablation experiment results.
1stODEM-GANGenerator Addition ConditionsLoss of Positioning AccuracyAP (%)
2ndno--84.8
3rdyesnono76.4
4thyesyesno85.1
5thyesyesyes88.4
Table 5. Experimental results using different λ values.
Table 5. Experimental results using different λ values.
λ 0.10.20.30.40.50.60.70.80.9
AP35.636.336.536.736.936.436.336.135.7
Table 6. Experimental results using different λ l p values.
Table 6. Experimental results using different λ l p values.
λ l p 0.020.040.060.080.10.120.140.160.18
AP (parachute)85.185.586.387.388.488.086.384.283.4
Table 7. The performance based on the AP score of the ODEM-GAN module for each category on the PASCAL VOC 2012 dataset.
Table 7. The performance based on the AP score of the ODEM-GAN module for each category on the PASCAL VOC 2012 dataset.
ObjectRetinaNetCBAFODEM-GAN
mAP73.875.176.8
aero85.386.887.5
bike79.280.682.8
bird76.978.279.5
boat59.962.166.2
bottle50.851.658.1
bus80.98283.2
car73.37577.1
cat93.99494
chair56.559.160
cow80.18183
table60.262.162.5
dog8989.989.9
horse85.88788.5
mbike84.885.485.9
person78.68082.3
plant45.647.250.1
sheep71.372.374.3
sofa70.67175.6
train82.683.283
tv70.972.774
Table 8. Results of the comparison of our improved detector (RetinaNet with our ODEM-GAN module) and state-of-the-art detectors.
Table 8. Results of the comparison of our improved detector (RetinaNet with our ODEM-GAN module) and state-of-the-art detectors.
MethodsBackbone A P AP 50 AP 75 AP S AP M AP L
DINO [43]ResNet-5050.26854.732.85364.8
YOLOR-D6 [47]YOLOv4-P6-light52.570.557.637.157.265.4
YOLOv4 [35]CSPDarknet-5341.262.844.320.444.456
CPN [38]DLA-3441.758.944.920.244.156.4
EfficientDet-D2 [39]Efficient-B24362.346.222.54758.4
DETR [41]Transformer43.563.846.421.94861.8
SM-NAS: E2 [40]-4058.243.421.142.451.7
CenterNet [49]Hourglass-10444.962.448.125.647.457.4
DINO [43]Swin-T63.380.869.946.76676.5
DyHead [45]Swin-T49.76854.333.354.264.2
GLIP [44]Swin-T61.579.567.745.364.975
QueryInst [46]Swin-T49.174.253.831.551.863.2
RDSNet [42]ResNet-10138.158.540.821.241.548.2
CBAF [22]ResNet-10140.960.243.222.543.350.4
DyHead [45]ResNet-10146.564.550.728.350.357.5
QueryInst [46]ResNet-10142.865.646.724.64555.5
our (RetinaNet + ODEM-GAN)ResNet-10141.360.543.224.843.351.4
Grid R-CNN w/FPN [48]ResNeXt-10143.26346.625.146.555.2
Dense RepPoints [37]ResNeXt-10140.263.843.123.143.652
FoveaBox [50]ResNeXt-10142.161.945.224.946.855.6
RetinaNet [7]ResNeXt-10140.861.144.124.144.251.2
CBAF [22]ResNeXt-1014363.246.325.945.651.4
DyHead [45]ResNeXt-10147.765.751.931.551.760.7
QueryInst [46]ResNeXt-10144.668.148.726.646.957.7
our (RetinaNet + ODEM-GAN)ResNeXt-10143.263.446.827.945.652.9
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Pei, Z.; Tang, Z.; Gu, F. ODEM-GAN: An Object Deformation Enhancement Model Based on Generative Adversarial Networks. Appl. Sci. 2022, 12, 4609. https://doi.org/10.3390/app12094609

AMA Style

Zhang Z, Pei Z, Tang Z, Gu F. ODEM-GAN: An Object Deformation Enhancement Model Based on Generative Adversarial Networks. Applied Sciences. 2022; 12(9):4609. https://doi.org/10.3390/app12094609

Chicago/Turabian Style

Zhang, Zeyang, Zhongcai Pei, Zhiyong Tang, and Fei Gu. 2022. "ODEM-GAN: An Object Deformation Enhancement Model Based on Generative Adversarial Networks" Applied Sciences 12, no. 9: 4609. https://doi.org/10.3390/app12094609

APA Style

Zhang, Z., Pei, Z., Tang, Z., & Gu, F. (2022). ODEM-GAN: An Object Deformation Enhancement Model Based on Generative Adversarial Networks. Applied Sciences, 12(9), 4609. https://doi.org/10.3390/app12094609

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop