Next Article in Journal
Automated Georectification, Mosaicking and 3D Point Cloud Generation Using UAV-Based Hyperspectral Imagery Observed by Line Scanner Imaging Sensors
Next Article in Special Issue
Unified Transformer with Cross-Modal Mixture Experts for Remote-Sensing Visual Question Answering
Previous Article in Journal
Impact of Spectral Resolution and Signal-to-Noise Ratio in Vis–NIR Spectrometry on Soil Organic Matter Estimation
Previous Article in Special Issue
ARE-Net: An Improved Interactive Model for Accurate Building Extraction in High-Resolution Remote Sensing Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MHLDet: A Multi-Scale and High-Precision Lightweight Object Detector Based on Large Receptive Field and Attention Mechanism for Remote Sensing Images

1
Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China
2
School of Computer and Information Engineering, Henan University, Kaifeng 475004, China
3
Henan Province Engineering Research Center of Spatial Information Processing and Shenzhen Research Institute, Henan University, Kaifeng 475004, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(18), 4625; https://doi.org/10.3390/rs15184625
Submission received: 22 June 2023 / Revised: 5 September 2023 / Accepted: 18 September 2023 / Published: 20 September 2023

Abstract

:
Object detection in remote sensing images (RSIs) has become crucial in recent years. However, researchers often prioritize detecting small objects, neglecting medium- to large-sized ones. Moreover, detecting objects hidden in shadows is challenging. Additionally, most detectors have extensive parameters, leading to higher hardware costs. To address these issues, this paper proposes a multi-scale and high-precision lightweight object detector named MHLDet. Firstly, we integrated the SimAM attention mechanism into the backbone and constructed a new feature-extraction module called validity-neat feature extract (VNFE). This module captures more feature information while simultaneously reducing the number of parameters. Secondly, we propose an improved spatial pyramid pooling model, named SPPE, to integrate multi-scale feature information better, enhancing the model to detect multi-scale objects. Finally, this paper introduces the convolution aggregation crosslayer (CACL) into the network. This module can reduce the size of the feature map and enhance the ability to fuse context information, thereby obtaining a feature map with more semantic information. We performed evaluation experiments on both the SIMD dataset and the UCAS-AOD dataset. Compared to other methods, our approach achieved the highest detection accuracy. Furthermore, it reduced the number of parameters by 12.7% compared to YOLOv7-Tiny. The experimental results illustrated that our proposed method is more lightweight and exhibits superior detection accuracy compared to other lightweight models.

Graphical Abstract

1. Introduction

Object detection plays a crucial role in remote sensing image processing, enabling the automated extraction of objects of interest from RSIs, including buildings, roads, vehicles, and more. In the past, manual visual interpretation was the predominant method for acquiring geographic information. However, with the ongoing advancements in remote sensing technology and the continuous innovation of image-processing algorithms, object detection for RSIs has become one of the essential means to obtain large-scale geographic information efficiently. Currently, object detection technology in RSIs has found widespread applications across diverse domains, including urban planning [1], environmental monitoring [2], agricultural production [3], and other fields. This technology holds promising application prospects.
Traditional object-detection methods, such as scale-invariant feature transform (SIFT) [4], support vector machine (SVM) [5], and histogram of oriented gradients (HOG) [6], have certain limitations due to manual features extraction and classifier construction. With the development of deep learning, object-detection methods can automatically learn features and construct efficient classifiers, resulting in higher accuracy and faster processing speeds. For example, Edge YOLO [7] can run in real-time on edge devices while maintaining high accuracy. At present, object detectors can be categorized into two main types: two-stage object detectors and one-stage object detectors. Two-stage detectors are represented by Faster R-CNN [8], Mask R-CNN [9], Cascade R-CNN [10], etc. Faster R-CNN is among the early pioneers of two-stage object detectors, distinguished in achieving both high detection accuracy and precise localization. The region proposal network (RPN) has been introduced to generate object candidate regions, followed by performing object classification and localization on these candidate regions. Mask R-CNN is a further advancement built upon the foundation of Faster R-CNN, which is able to not only detect object location and category, but also to generate accurate segmentation masks of objects. This makes Mask R-CNN perform well in image segmentation tasks. Cascade R-CNN is a two-stage detector proposed for small object detection, it significantly enhances the detection accuracy of small objects by means of cascade classifiers. Compared with other two-stage object detectors, Cascade R-CNN has higher detection accuracy for small objects, but slightly decreased detection accuracy for large objects.
Although two-stage detectors excel in high detection accuracy and positioning accuracy, their calculation speed is relatively slow. Moreover, training a two-stage object-detection model requires substantial data, and its efficiency diminishes when faced with a high volume of objects to detect. This situation is a challenge for both data acquisition cost and time cost. One-stage object detection models represented by SSD [11], RetinaNet [12], and you only look once (YOLO) [13] exhibit the advantages of high speed and applicability to scenarios such as real-time detection. With the continuous development of algorithms, their detection accuracy is gradually improved. However, the disadvantages of the one-stage detection model are low positioning accuracy, insensitivity to complex backgrounds, and occlusion, and these problems may have a certain impact on the target-detection effect of remote sensing images. In 2016, YOLOv1 [13] was proposed by Redmon et al., renowned for its fast detection speed, but the positioning accuracy of object detection is low. In 2017, Redmon et al. introduced YOLOv2 [14], which used DarkNet-19 as its backbone architecture. Despite achieving higher detection accuracy, it is still not friendly to small target recognition. In 2018, Redmon et al. introduced YOLOv3 [15], which used DarkNet-53 with the residual network to extract features from images. In 2019, Bochkovskiy proposed YOLOv4 [16]; he embedded the cross-stage partial (CSP) module into DarkNet-53, improving the accuracy and speed. In 2020, YOLOv5 was proposed with the Focus architecture, and it speeded up the training and improved the detection accuracy. In 2022, YOLOv7 [17] was proposed by Wang et al., which maintained a fast running speed and small memory consumption and achieved ideal results in detection accuracy. Although the YOLO series models have become some of the representative algorithms for object detection, single-stage object-detection methods still have certain limitations in object detection on large images such as RSIs.
In recent years, researchers have achieved notable advancements in the detection of small [18] or rotating object detection [19]. Yang et al. [20] proposed the sampling fusion network (SF-Net) to optimize for dense and small targets. Fu et al. [21] constructed a semantic module and a spatial layout module and then fed both into a contextual reasoning module to integrate contextual information about objects and their relationships. Han et al. [22] proposed the rotation-equivariant detector (ReDet), which explicitly codes rotational equivalence and rotational invariance to predict the direction accurately. Li et al. [23] proposed a model named OrientedRepPoints and a novel adaptive point learning quality assessment and sample allocation method, and they achieved more-accurate performance by accurately predicting the direction. However, while many researchers concentrate on challenges such as small object detection or the detection of rotating objects, they often overlook other common issues in RSIs’ object detection. These issues include enhancing the detection of multi-scale objects, tackling the difficulties of identifying objects concealed in shadows or with colors resembling the ground, and detecting objects with only partial features exposed in RSIs, as illustrated in Figure 1.
Considering the multi-faceted nature and complexity of current RSI problems, this paper proposes an enhanced detector, MHLDet, so as to address the issues mentioned above. Firstly, the VNFE is designed to alleviate the surroundings’ impact and obtain more-robust feature information. In addition, it can concentrate more effectively on crucial regions to address the issue of partial feature detection. Secondly, we propose the SPPE with large receptive fields to fuse better multi-scale features to enhance the network’s ability to detect targets of various scales. Finally, we introduce the CACL to preserve more feature information by minimizing the information loss during downsampling. Therefore, this model holds an advantage in detecting objects with colors close to the ground and in detecting multi-scale targets. In addition, it achieves higher precision and is more lightweight than existing methods. The primary contributions of this paper are outlined as follows:
  • This paper proposes a novel method for object detection in RSIs, which achieves a preferable balance between lightweight and accuracy.
  • We propose the VNFE module for more-effective feature extraction and integrated the SimAM attention mechanism into the backbone to focus on the key features. The VNFE module has fewer parameters, so it can reduce the computational cost more effectively and make the model lighter. In addition, the VNFE module can also enhance the generalization ability of the detector, as it can reduce the risk of overfitting while maintaining high accuracy.
  • This paper introduces a novel SPPE module by optimizing the SPPCSPC of YOLOv7-Tiny [17]. The SPPE is designed to enhance detection accuracy for medium- to large-sized objects while preserving the original advantages. As a result, it effectively fulfills the demands for detecting objects of various scales and addresses the application requirements.
  • The CACL module is designed to minimize the loss of feature information during the downsampling process in the neck network. This module can focus on feature regions with higher discrimination by performing attention calculations on the feature map and enriching the semantic information by fusing feature elements. Consequently, it improves the accuracy and robustness of the model.
The remainder of this study is structured as follows: Section 2 reviews related work on lightweight object detection algorithms and outlines the details of the proposed model. Section 3 introduces the implementation details of the experiments conducted on the SIMD and the UCAS-AOD datasets, and we present the results obtained from the ablation experiments. Section 4 carries out the discussion. Finally, Section 5 provides the conclusions from the study.

2. Materials and Methods

2.1. Related Work

2.1.1. Lightweight Object-Detection Algorithms

In recent years, the one-stage detectors have garnered significant attention due to their fast execution speed and fewer parameters than two-stage detectors. YOLT [24] excels in multi-scale detection, precise localization, and real-time performance, making it particularly suitable for applications requiring high-quality detection of small objects. FCOS [25] regresses the bounding boxes at each position on the feature map to their corresponding positions on the original image. DETR [26] can predict the positions and categories of all objects in an image simultaneously. Zhu et al. proposed Deformable DETR [27], which achieves higher accuracy while reducing the training time. EfficientDet [28] introduced the bi-directional feature pyramid network (BiFPN) to achieve efficient multi-scale feature fusion. VFNet [29] introduced varifocal loss and an IoU-aware classification score to enhance the detection accuracy of dense objects. Aiming to obtain more-accurate features and enhance detection performance, R3DET [30] introduces the feature refinement module to address the misalignment between feature regions and feature maps in one-stage detectors. Moreover, it reduces the overall training memory consumption by approximately two-times.
Based on one-stage detectors, numerous scholars have proposed lightweight models. ShuffleNetV2 [31] stands out for its channel shuffling and efficient group convolutions, offering low computational costs for real-time applications. MobileNetV3 [32] introduces architectural optimizations, achieving improved accuracy while maintaining efficiency. YOLOv3-Tiny [15] is the earliest lightweight model in the YOLO series. By designing a simpler and more-efficient network structure, it achieves faster inference speed with fewer parameters in the YOLO series. With the proposal of the YOLOv7-Tiny network, it has reached a new peak in lightweight object-detection methods. However, there is still scope for further improvement in terms of both the detection accuracy and parameter optimization. Therefore, this paper proposes a new lightweight object-detection method aimed at improving the detection accuracy of medium- to large-sized objects and objects with no obvious features.

2.1.2. Attention Mechanism

The visual attention mechanism is a specialized process within human vision, aiding in swiftly identifying crucial information amidst cluttered surroundings. It establishes the focal point of attention, thus enhancing the efficient processing of visual information. The attention mechanism was applied to computer vision as early as 1998 by Laurent Itti [33], inspired by the neurons of the primate visual system, to merge multi-scale feature maps and select focal regions in order of saliency. In 2014, Google’s DeepMind first applied the method of the attention mechanism on the recurrent neural network (RNN) model for image classification tasks, which attracted more people’s attention [34].
In the field of computer vision, attention mechanisms can be categorized into two types: channel attention and spatial attention. Channel attention is represented by the squeeze-and-excitation network (SENet) model [35]. SENet initially compresses the feature map’s spatial dimensions, subsequently employing network model learning to ascertain the significance of individual feature channels autonomously. This approach can assign varying weight coefficients to each channel, amplifying critical features while attenuating less-relevant ones. Spatial attention, such as spatial transformer networks (STNs) [36], can enable the transformation of data with different spatial deformations. It has the ability to automatically capture important regional features by adjusting and aligning the input data in a spatially adaptive manner.
However, it is essential to note that channel attention is a one-dimensional attention mechanism, which treats different channels differently, but treats all spatial locations equally. Spatial attention is a two-dimensional attention mechanism that treats different locations within the feature map differently while giving equal importance to all channels. SimAM [37] is a 3D parameter-free attention mechanism that enhances the feature extraction capability of neural networks. SimAM attaches 3D attention weights to feature maps, unlike traditional one-dimensional or two-dimensional weight attention methods. By considering the interdependencies among channels, spatial locations, and different depth levels, SimAM captures more-comprehensive and fine-grained information from the feature maps. This allows the network to focus on relevant features and improve its ability to extract meaningful representations, enhancing performance in various computer vision tasks.

2.1.3. Partial Convolution

Convolution is a common operation in object detection, extracting feature information by accessing all input channels. Multiple convolutional layers are employed for dimension transformation and feature extraction. However, this approach may encounter computation bottlenecks when dealing with huge input images or using larger convolutional kernels, augmenting the model’s complexity. This solution leads to an increase in the number of parameters and computational requirements. For a feature map with the input  I R C × H × W , using c filters  F R K × K , the FLOPs would be:
h × w × k 2 × c 2
The partial convolution (PConv) mentioned in FastNet [38] exploits the redundancy in the feature map and systematically applies regular convolution (Conv) on only a portion of the input channels without affecting the remaining channels to maintain high FLOPS at reduced FLOPs, as a way to reduce latency and, thus, the parameters and computational effort of the model. The formula for calculating latency is as follows:
L a t e n c y = F L O P s F L O P S
The FLOPs of PConv are:
h × w × k 2 × c p 2
where  c p = c / 4 , and the FLOPs of PConv are only  1 / 16  of the conventional convolution. Therefore, in the proposed VNFE module, we utilized PConv to effectively reduce the number of model parameters.

2.2. Proposed Method

The overall architecture of our MHLDet is depicted in Figure 2. In the backbone network, we embedded the VNFE module and SimAM to enhance the feature extraction ability while reducing the model’s parameters. We introduced the SPPE module within the neck network that executes the pyramid pooling operations on the feature maps, which are obtained from the backbone network. This process aims to acquire multi-scale feature representations. The CACL module has been proposed to enhance feature expression capabilities and reduce information loss. Table 1 summarizes the main parameters of the proposed method.
In the backbone,  I n p u t R 1024 × 1024 × 3  represents the input images. It is downsampled by two CBL modules (consisting of convolutional layers, batch normalization, and the LeakyRELU activation function [39]); the convolution kernel size is  3 × 3 , and the stride is set to 2. Moreover, the SimAM attention mechanism is embedded after max pool to help the network focus on key features, which can better represent important information in the input images to enhance the performance and the expression ability of the model. Then, we used the VNFE module to make the feature mapping improved, accelerate the training and convergence speed, and make the model more lightweight. The SPPE module is used to extract feature vectors of objects with multiple scales and expand their receptive fields. Feature maps of size  128 × 128 64 × 64 , and  32 × 32  are extracted from the backbone P3, P4, and P5 layers, respectively, and sent to the neck network, where feature fusion is performed. As we all know, low-level features encompass local information such as edges and textures of objects, providing more spatial information, but limited semantic information. On the other hand, high-level features tend to possess richer semantic information. To address the loss of feature information due to downsampling and to make the neural network more robust and accurate for recognition, the features of the P3, P4, and P5 layers are fused by the CACL module.

2.2.1. Validity-Neat Feature Extract Module

For the original image input to the model, excessive convolutional and pooling layers in the backbone network will not only lead to feature redundancy but also lead to computational redundancy. We propose using the VNFE module to replace the regular convolutional layers of the original backbone network.
As shown in Figure 3, the PBL module plays a role in downsampling the feature map and extracting high-level features. Specifically, the PConv within this module extracts spatial features from the input channels, leaving the other channels unaltered. This strategic approach reduces redundant computations and minimizes memory usage, contributing to the overall lightweight nature of the model. Furthermore, SimAM is integrated into the second branch of the VNFE module. It aims to assist the network in emphasizing key features and appending crucial information onto the feature maps without introducing extra parameters. The formula is as follows:
Y 1 = P 3 , 1 ( P 3 , 1 ( F 1 , 1 ( X ) ) )
Y 2 = P 3 , 1 ( ( F 1 , 1 ( X ) ) + ( S i m A M ( X ) ) )
Y 3 = F 1 , 1 ( Y 1 Y 2 )
where X represents the feature input,  F 1 , 1  denotes a convolution operation with a filter size  1 × 1  and stride of 1,  P 3 , 1  is a PConv operation with a filter size of  3 × 3  and stride of 1, SimAM represents the attention operation, and ⊕ is the concatenate operation.
SimAM attention distinguishes the importance of neurons by defining an energy function for each neuron. Specifically, an energy function is defined for each neuron using the following formula:
e t ( w t , b t , y , x i ) = 1 M 1 i = 1 M 1 ( 1 ( ω t x i + b t ) ) 2 + ( 1 ( ω t t + b t ) ) 2 + λ ω t 2
where t denotes the target neuron while  x i  denotes the other neurons in a single channel of the input feature  X R ( C × H × W ) , i is the index in the spatial dimension, and  M = H × W  is the number of all neurons in a single channel.  ω t  and  b t  are the weights and biases of the transformation, which can be expressed as follows:
ω t = 2 ( t μ t ) ( t μ t ) 2 + 2 σ t 2 + 2 λ
b t = 1 2 ( t + μ t ) ω t
μ t  and  σ t 2  are the mean and variance of all neurons in this channel, except t:
μ t = 1 M 1 i = 1 M 1 x i
σ t 2 = 1 M 1 i = 1 M 1 ( x i μ t ) 2
Assuming the same distribution among all the pixels in a single channel, we can calculate the mean and variance of all the neurons, which allows us to derive the minimum energy formula:
e t = 4 ( σ ^ 2 + λ ) ( t μ ^ ) 2 + 2 σ ^ 2 + 2 λ
Equation (12) indicates that the energy value of  e t  reflects the significance of the contrast between neuron t and nearby neurons, and a lower energy value indicates a more-pronounced difference and higher importance of neuron t in the context of nearby neurons.  λ  is the regularization coefficient. The entire refinement phase of SimAM can be calculated as follows:
S i m A M ( X ) = s i g m o i d ( 1 E ) × X
where E groups all  e t  from the spatial and channel dimensions. Besides, we used sigmoid activation to constrain excessive E values.

2.2.2. Spatial Pyramid Pooling Enforce Module

The current spatial pyramid pooling method still has potential for improvement when detecting medium- to large-sized targets. In order to better enhance the detection ability of the model for multi-scale targets, we learned from the SPPCSPC [17] module and propose a new spatial pyramid pooling enforce (SPPE) module. In the first branch, multi-scale feature information is captured by constructing pyramid levels at different scales. Pooling operations at different scales enable the network to perceive and process target objects of different scales and sizes of the input image, thereby improving the adaptability and generalization ability of the model. The larger pooling operation can retain the overall features and spatial layout of the target object. By increasing the convolution kernel size, the model can cover more details and context information of the target area, which is helpful in detecting medium- to large-sized targets more accurately. Therefore, we changed the connection mode of the maximum pooling layers. The receptive fields increased to 5, 13, and 25, respectively, so as to enhance the detection capability for multi-scale objects. The receptive field can be calculated using the following formula:
r 0 = 1 , r 1 = k 1 , r n = r n 1 + ( k n 1 ) i = 1 n 1 s i . ( n 2 )
The input is passed through the CBL and PBL in the second branch, respectively. The number of output channels is reduced by half compared to the original without changing the feature map size. These channels are subsequently element by element multipliedto increase the amount of information in each dimension of the feature map, which can supplement the feature detail information to obtain a richer and more-comprehensive feature representation. In addition, it makes up for the attenuation of the small target detection ability due to the increase of the receptive field. Finally, the two branches are concatenated in the channel dimension. Then, the number of output channels of the SPPE was adjusted to be consistent with the input through a CBL with a  1 × 1  convolutional kernel size. The SPPE structure is illustrated in Figure 4.

2.2.3. Convolution Aggregation Cross Layer

Using a CBL with only a 3 × 3 convolution for downsampling can result in information loss and decrease feature quality. To minimize the information loss caused by downsampling, we embedded the CACL module within the network, allowing the feature map to retain more feature information. In the first branch of the CACL, compared with the max pooling layer, using the average pooling layer can better preserve the spatial information of the feature map and effectively utilize the global information. After that, a  1 × 1  convolution is used to deepen the network, introduce more nonlinearity, and enhance the neural network’s expressive capability without expanding the receptive field. The feature map’s channel count is doubled as well. Then, the network model focuses on the feature regions with higher discrimination by using the SimAM attention mechanism.
We used a CBL module with a convolution kernel size of  3 × 3  and a stride of 2 in the second branch to reduce the feature map size through downsampling. This downsampling process helps maintain crucial feature information while enhancing the model’s robustness.
By adding the two branches, the module can fuse different feature expressions, improve the diversity of features and the expression ability, and enhance the representation ability and detection performance. Its structure is illustrated in Figure 5, and the formula is as follows:
O = S i m A M ( F 1 , 1 ( A v g P o o l ( X ) ) ) + F 3 , 2 ( X )
where O and X represent the output and input feature maps of this module, SimAM represents the attention operation,  F 3 , 2  represents a convolution operation with filter size  3 × 3  and a stride of 2, and AvgPool represents the average pooling operation.

3. Results

3.1. Datasets

We conducted the experiments using the SIMD [40] and UCAS-AOD [41] datasets to evaluate the effectiveness. Figure 6 shows some examples of the datasets this paper used.
The SIMD dataset is a recently released high-resolution remote sensing dataset. It is characterized by its high resolution and the inclusion of multi-scale imagery. This dataset contains 5000 high-resolution images with 15 categories, Car, Truck, Van, Long Vehicle, Bus, Airliner, Propeller Aircraft, Trainer Aircraft, Chartered Aircraft, Fighter Aircraft, Others, Stair Truck, Pushback Truck, Helicopter, Boat, respectively. There are a total of 45,096 instances, and the image size is 1024 × 768 pixels. We used 4000 images for training and 1000 images for validation according to the division of the original dataset.
The UCAS-AOD dataset is specifically designed for the detection of airplanes and cars. The image size is 1280 × 659 pixels, and it contains two categories of airplane and car. The objects in the images are evenly distributed in terms of their orientations. In this experiment, we removed the negative example images that did not contain any example images and only kept the images that contained the two classes of airplane and car. The training set, validation set, and test set of the UCAS-AOD dataset were randomly divided with a ratio of 7:2:1.

3.2. Evaluation Criteria

To evaluate the performance of our proposed detector, we used several widely used evaluation metrics. These metrics included the precision (P), recall (R), average precision ( A P ), and mean average precision ( m A P ). The formulas for precision and recall are as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
A P = 0 1 P ( R ) d R
where  T P  is true positive and is the number of positive samples that were correctly predicted to be positive.  F P  is false positive, which is the number of false positives that were incorrectly predicted.  F N  is false negative and is the number of positive samples that were incorrectly predicted as negative.  A P  is the area under the precision–recall curve (P–R curve) and ranges from 0 to 1.  m A P  is a commonly used comprehensive evaluation metric in object detection, which comprehensively considers the precision and recall of different categories. We calculated the area under the P–R curve for each class and, then, averaged the  A P  across all classes to obtain the  m A P . The  m A P  provides an overall evaluation of model performance and can be used to compare the strengths and weaknesses of different models.
m A P = 1 n i = 1 n A P i
where n represents the total number of categories in the detection task.
The intersection over union ( I o U ), is a metric used to measure the degree of overlap between the predicted bounding box ( B p ) and the true bounding box ( B g t ) in the object-detection task. It is defined by computing the ratio between the intersection area of the predicted bounding box and the true bounding box and the area of their union. The calculation formula is as follows:
I o U = a r e a ( B p B g t ) a r e a ( B p B g t )
In addition, we followed the COCO evaluation metrics [42], including the mAP ( I o U  threshold = 0.5),  m A P 0.75 , and  m A P 0.5 : 0.95 , and it also includes  A P s  (for small objects, area <  32 2 ),  A P m  (for medium objects,  32 2  < area <  96 2 ), and  A P l  (for large objects,  96 2  < area).
These metrics can help us understand the efficiency of the model in practical applications. Combining these evaluation metrics, we can comprehensively assess the performance of our proposed method in regard to accuracy, efficiency, and practicality and compare it with current methods.

3.3. Parameter Setting

The experimental environment was configured as follows: the computer was equipped with an NVIDIA GeForce RTX 3070 graphics card (8GB); the CPU was an Intel Core i7-10700K; the operating system was Ubuntu 20.04.4 LTS. We used the PyTorch deep learning framework (Version 1.13.1) for model development and training in Python 3.8 and CUDA 11.4.
The input image size was  1024 × 1024 ; the batch size was four; the optimization was stochastic gradient descent (SGD); the momentum was 0.937; the weight decay was set to 0.0005. To improve the convergence and training efficiency of the model, we performed three epochs of warm-up training and set the warm-up momentum to 0.8 initially. The initial train parameters are shown in Table 2.

3.4. Experimental Results

3.4.1. Experimental Results on the SIMD Dataset

This paper took YOLOv7-Tiny [17] as the baseline and conducted experiments on the SIMD dataset to validate the effectiveness of our method and compare it with other one-stage detectors. The detectors for comparison were the anchor-free detector FCOS [25], TOOD [43], PP-Picodet-m [44], YOLOX-S [45], YOLOv6-N, YOLOv6-S [46], and YOLOv8-N [47], and the anchor-based detectors were RetinaNet [12], YOLOv5-N, YOLOv5-S [48], YOLO-HR-N [49], and YOLOv7-Tiny. We did not compare with two-stage detectors. In our experiments, we focused on proposing a lightweight model designed to meet the needs of scenarios with limited computational resources, and our approach focused more on reducing the complexity and computational overhead of the model while maintaining high performance compared to traditional two-stage models. Our experiments on the SIMD dataset used the K-Means++ [50] algorithm to generate a new anchor, and the experimental data after using the new anchor are shown in Table 3. In addition, to display the results more directly, this paper assigned an abbreviation name (AN) to each category in the dataset, as shown in Table 4.
Table 5 shows the detection accuracy results of our detector and other detectors in various categories on the SIMD dataset. It is evident that our method achieved optimal accuracy across the majority of categories, with the mAP reaching 84.7%. As shown in the table, our mAP was 2.4% higher than YOLOv7-Tiny. Additionally, it was nearly 10% higher than YOLOv5-N.
There are 15 categories in the SIMD dataset, and our detection efficiency in most categories reached an optimal level. In dense scenes, such as Car (CR) and Long Vehicle (LV), our MHLDet achieved the best performance. At the same time, it had a significant improvement compared with other models on medium- to large-sized targets such as Airliner (ALR), Chartered Aircraft (CA), and Fighter Aircraft (FA). The observed improvement can be attributed to the proposed VNFE module and the embedded SimAM attention mechanism, which effectively leverages the feature information within the backbone and enhances the feature representation, while the inclusion of the SPPE module expands the receptive fields, making it more conducive for detecting medium- to large-sized objects, and the CACL module reduced the feature loss in the process of downsampling. The experimental results further validated the effectiveness and robustness of the proposed method. Table 6 shows the performance comparison of MHLDet with other one-stage methods.
On the SIMD dataset, our MHLDet detector achieved an 84.7% mAP with 5.28 m model parameters. As shown in Figure 7, the proposed method outperformed all compared one-stage methods, achieving the highest detection accuracy with fewer parameters. This excellence enabled our model to excel in resource-constrained environments. Compared with the baseline, the parameters of the model proposed by us decreased by 12.7%, and the mAP increased by 2.4%; the  m A P 0.75  and  m A P 0.5 : 0.95  also increased by 2.9% and 2.6%, respectively. On medium- to large-sized objects,  A P m  and  A P l  increased by 2.4% and 1.3%, respectively, compared to the baseline. In particular, the proposed model achieved the highest performance scores the mAP,  m A P 0.75 m A P 0.5 : 0.95 , and small-, medium-, and large-scale AP values, while maintaining fewer parameters. Compared with the anchor-free detectors PP-PicoDet-m, YOLOv6-N, and YOLOv8-N, these detectors’ parameters were slightly fewer than MHLDet, but our mAP showed an increase of 11.8%, 10.3%, and 3.1%. In comparison to the anchor-based detectors, YOLOv5-N had the fewest parameters, but we were 9% and 8.9% higher in the mAP and  m A P 0.5 : 0.95 , respectively. In the aspect of multi-scale object detection, our  A P m  and  A P l  were 12.3% and 10.1% higher than their counterparts.
By comparing the  A P s A P m , and  A P l  with each model, we found that the gains of our model mainly came from medium objects and large objects. Although the accuracy on medium objects was the same as that of YOLOv5-S, the accuracy on  A P s  and  A P l  improved by 8.9% and 4.1%, respectively. Despite the MHLDet model having 3.5 M more parameters compared to YOLOv5-N, it remains categorized as a lightweight model.
Figure 8 depicts the visual comparison of the heat map of the baseline with our proposed method, and it can be seen that our proposed method excels in predicting objects with only partial features and objects that are close to the surface color. To be specific, the heatmap’s colors transition from blue to red on the images, symbolizing the probability of target presence increasing from low to high. Blue areas indicate the potential absence of targets, while red regions signify a high likelihood of target presence. Figure 8(b1–b3) are the results of YOLOv7-Tiny, and Figure 8(c1–c3) are the results of our proposed model. From Figure 8(b1,c1), it can be observed that we also had a heatmap representation on the ship just below, where only a portion of the features was exposed. As shown in Figure 8(b2,c2), YOLOv7-Tiny overlooked the middle vehicle, where our model correctly annotated it. From Figure 8(b3,c3), it is evident that our model provided more-accurate and continuous annotations for vehicles in shadow. As a result, our model can accurately depict the distribution of data. The visualization detection results are shown in Figure 9.
To visually illustrate the improvement achieved by our proposed model, this paper presents the detection results for some typical scenes. Figure 9(b1–b6) are the results of YOLOv7-Tiny, and Figure 9(c1–c6) are the results of our proposed model. From Figure 9(b1,c1), our model performed significantly better than YOLOv7-Tiny in terms of detecting small objects. From Figure 9(b2–b4,c2–c4), it can be observed that our model detected more objects that closely matched the color of the ground. As shown in Figure 9(b5,c5), our model detected the object that exhibited only partial features. In addition, our model detected a higher number of densely arranged objects in Figure 9(b6,c6).

3.4.2. Experimental Results on the UCAS-AOD Dataset

Table 7 shows the accuracy comparison of our method with other methods on the UCAS-AOD dataset. It is evident from the table that our proposed method achieved the highest mAP among all the compared methods. Our method demonstrated a 4.6% improvement over the second-best method for small and 2% for large objects.
Based on the data presented in Table 7, our proposed method improved in all scales compared to other methods, and it can be competent for the vast majority of object-detection tasks by the excellent performance compared to current state-of-the-art methods. Figure 10 shows the visualization results on the UCAS-AOD dataset, further supporting the effectiveness of our proposed method.

3.5. Ablation Study

To verify the effectiveness of our proposed detector, we conducted ablation studies on the SIMD dataset using the same hyperparameters and parameter settings for each module to ensure a fair and unbiased comparison. We used the precision, recall, mAP, Params, and floating point operations (FLOPs) to verify the availability of the module we proposed. It can be seen in Table 8 and Figure 11 that each of our proposed modules had some improvements over the baseline.
It can be observed from the second and third row of Table 8 that, with the help of SimAM, the mAP improved from 82.3% to 83.6%. Furthermore, with the help of the VNFE, the mAP increased to 83.7%, and the Params and FLOPs decreased by 18% and 10.5%, respectively. This was caused by the absence of semantic information in the backbone before the use of both of them, resulting in less-prominent features. The addition of SimAM enhanced the features of objects to be detected, while VNFE effectively reduced the parameters when extracting the features, which also explains the reason the mAP value improved when the parameters were reduced.
In order to prove the SPPE, we replaced the original SPPCSPC with the SPPE, with the rest of the model remaining the same, from Table 8. In the fourth row, the mAP improved from 82.3% to 83.7%. In the SPPE, we changed the connection mode of the maximum pooling layer and increased its receptive fields from the original 5, 9, and 13 to 5, 13, and 25, so as to enhance the model’s detection performance for multi-scale objects.
The validity of the CACL module can be seen from the fifth row in Table 8. The mAP increased from 82.3% to 83.4%. The CACL module effectively utilized the global information and preserved the crucial details in the feature map during downsampling, which led to the enhanced robustness of the model.
From the sixth and seventh rows of Table 8, it can be observed that stacking multiple modules progressively yielded increasing gains, reaching 84.1% and 84.4%, surpassing the results obtained from using a single module. Additionally, as the achieved mAP continued to rise, the increase in the parameter count also remained within an acceptable range.
Finally, in the eighth row of Table 8, it is evident that the mAP value of applying all the improved modules surpassed that of any individual module alone. It balanced the precision and light weight, enabling higher accuracy even with limited computing resources.

4. Discussion

The comparison of the first and eighth rows in Table 8 reveals that, while precision decreased by 0.2%, both the recall and mAP increased by 3% and 2.4%, respectively. This indicates an enhanced ability of the proposed detector to detect objects within remote sensing images, reducing instances of missed detections and facilitating a more-comprehensive recognition of critical features and objects. Furthermore, there was a reduction in the number of parameters and FLOPs by 0.77 m and 1.1 G, respectively. In summary, depending on the specific application and requirements, trading off a slight decrease in precision for an improved recall and mAP can be a justifiable choice.
In addition to the above successes, the proposed detector had certain limitations. MHLDet exhibits suboptimal performance on unmanned aerial vehicle (UAV) images. UAVs can capture images from various poses and viewpoints, leading to variations in the appearance of the same target across different images. This highlights the need for improving the model’s generalization ability. Furthermore, when detecting a multitude of rotated objects in RSIs, it can result in horizontal detection boxes containing background regions and overlapping detection boxes, which adversely affects precision. Therefore, the detection performance of MHLDet in this scenario needs to be improved.

5. Conclusions

This paper proposes a more-lightweight object detector, MHLDet, based on YOLOv7-Tiny. We integrated SimAM and our proposed VNFE into the backbone network to enhance the model’s feature-extraction capability. The improved backbone network also helps the model focus more on critical regions in the feature maps to alleviate the challenge of detecting objects affected by shadows. It can be observed in Figure 9 that MHLDet detects more objects that closely resemble the ground color. Furthermore, the SPPE fuses multi-scale features, enabling the network to handle targets of various scales effectively. Finally, we used the CACL module for downsampling instead of using regular convolution in order to reduce the feature map size and preserve more feature information. The experimental results on the SIMD dataset demonstrated the efficacy of our proposed model. It achieved a reduction in parameters by 12.7% while improving the mAP by 2.4% compared to the baseline and outperformed the majority of the current object-detection models. In our future work, we will focus on dense and rotated objects and enhance the model’s detection performance for these objects while maintaining a lightweight design.

Author Contributions

Conceptualization, L.Z.; methodology, H.Z.; software, H.Z.; validation, L.Z., H.Z., Z.L. and K.C.; formal analysis, Y.L.; investigation, K.C.; resources, X.Z.; data curation, K.C.; writing—original draft preparation, L.Z. and H.Z.; writing—review and editing, L.Z. and Y.L.; visualization, Z.L.; supervision, Z.L.; project administration, H.Z.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Basic Research Program of China (Grant No. 2019YFE0126600); the Major Project of Science and Technology of Henan Province (Grant No. 201400210300); the Key Scientific and Technological Project of Henan Province (Grant No. 212102210496); the Key Research and Promotion Projects of Henan Province (Grant Nos. 212102210393 and 202102110121); the Kaifeng Science and Technology Development Plan (Grant No. 2002001); the National Natural Science Foundation of China (Grant No. 62176087); and the Shenzhen Science and Technology Innovation Commission (SZSTI), the Shenzhen Virtual University Park (SZVUP) Special Fund Project (Grant No. 2021Szvup032).

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

We sincerely thank the anonymous Reviewers for the critical comments and suggestions for improving the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hu, L.; Ni, Q. IoT-driven automated object detection algorithm for urban surveillance systems in smart cities. IEEE Internet Things J. 2017, 5, 747–754. [Google Scholar] [CrossRef]
  2. Shan, Y.; Dai, X.; Li, W.; Yang, Z.; Wang, Y.; Qu, G.; Liu, W.; Ren, J.; Li, C.; Liang, S.; et al. Detecting Spatial-Temporal Changes of Urban Environment Quality by Remote Sensing-Based Ecological Indices: A Case Study in Panzhihua City, Sichuan Province, China. Remote Sens. 2022, 14, 4137. [Google Scholar] [CrossRef]
  3. Tseng, H.H.; Yang, M.D.; Saminathan, R.; Hsu, Y.C.; Yang, C.Y.; Wu, D.H. Rice seedling detection in UAV images using transfer learning and machine learning. Remote Sens. 2022, 14, 2837. [Google Scholar] [CrossRef]
  4. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  5. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
  6. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  7. Liu, S.; Zha, J.; Sun, J.; Li, Z.; Wang, G. EdgeYOLO: An Edge-Real-Time Object Detector. arXiv 2023, arXiv:2302.07483. [Google Scholar]
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  9. He, K.; Gkioxari, G.; Dollár, P. Mask r-cnn. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  10. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  11. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  12. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  13. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  14. Redmon, J.; Farhadi, A. Yolo9000: Better, faster, stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  15. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  16. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  17. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  18. Srinivas, A.; Lin, T.Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck transformers for visual recognition. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16514–16524. [Google Scholar]
  19. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2844–2853. [Google Scholar]
  20. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8231–8240. [Google Scholar]
  21. Fu, K.; Li, J.; Ma, L.; Mu, K.; Tian, Y. Intrinsic relationship reasoning for small object detection. arXiv 2020, arXiv:2009.00833. [Google Scholar]
  22. Han, J.; Ding, J.; Xue, N.; Xia, G.S. Redet: A rotation-equivariant detector for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2786–2795. [Google Scholar]
  23. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented RepPoints for Aerial Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1819–1828. [Google Scholar]
  24. Van Etten, A. You only look twice: Rapid multi-scale object detection in satellite imagery. arXiv 2018, arXiv:1805.09512. [Google Scholar]
  25. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  26. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  27. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  28. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  29. Zhang, H.; Wang, Y.; Dayoub, F.; Sunderhauf, N. Varifocalnet: An IoU-aware dense object detector. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8510–8519. [Google Scholar]
  30. Yang, X.; Liu, Q.; Yan, J.; Li, A. R3det: Refined single-stage detector with feature refinement for rotating object. arXiv 2021, arXiv:1908.05612. [Google Scholar] [CrossRef]
  31. Ma, N.N.; Zhang, X.Y.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 122–138. [Google Scholar]
  32. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  33. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]
  34. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  35. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  36. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  37. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Graz, Austria, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  38. Chen, J.; Kao, S.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t Walk: Chasing Higher FLOPS for Faster Neural Networks. arXiv 2023, arXiv:2303.03667. [Google Scholar]
  39. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier Nonlinearities Improve Neural Network Acoustic Models. Proc. ICML 2013, 30, 3. [Google Scholar]
  40. Haroon, M.; Shahzad, M.; Fraz, M.M. Multisized object detection using spaceborne optical imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3032–3046. [Google Scholar] [CrossRef]
  41. Zhu, H.; Chen, X.; Dai, W.; Fu, K.; Ye, Q.; Jiao, J. Orientation robust object detection in aerial images using deep convolutional neural network. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3735–3739. [Google Scholar]
  42. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  43. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-Aligned One-Stage Object Detection. In Proceedings of the 2021 IEEE International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3490–3499. [Google Scholar]
  44. Yu, G.; Chang, Q.; Lv, W.; Xu, C.; Cui, C.; Ji, W.; Dang, Q.; Deng, K.; Wang, G.; Du, Y.; et al. PP-PicoDet: A Better Real-Time Object Detector on Mobile Devices. arXiv 2021, arXiv:2111.00902. [Google Scholar]
  45. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  46. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.029726. [Google Scholar]
  47. Ultralytics. YOLOv8. Available online: https://github.com/ultralytics/ultralytics (accessed on 1 January 2023).
  48. Ultralytics. YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 18 June 2022).
  49. Wan, D.; Lu, R.; Wang, S.; Shen, S.; Xu, T.; Lang, X. YOLO-HR: Improved YOLOv5 for Object Detection in High-Resolution Optical Remote Sensing Images. Remote Sens. 2023, 15, 614. [Google Scholar] [CrossRef]
  50. Arthur, D.; Vassilvitskii, S. K-means++ the advantages of careful seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; pp. 1027–1035. [Google Scholar]
Figure 1. The object-detection issues in remote sensing images from the UCAS-AOD and the SIMD datasets. (a) represents the multi-scale objects of RSIs; (b) represents the objects hidden in shadow in RSIs; (c) represents the objects with only partial features exposed in RSIs.
Figure 1. The object-detection issues in remote sensing images from the UCAS-AOD and the SIMD datasets. (a) represents the multi-scale objects of RSIs; (b) represents the objects hidden in shadow in RSIs; (c) represents the objects with only partial features exposed in RSIs.
Remotesensing 15 04625 g001
Figure 2. The overall structure of the MHLDet. We embedded the SimAM attention mechanism and VNFE module in the backbone and used the SPPE and CACL structure in the neck network. H, W, and C represent height, width, and channels, respectively. Cls represents the number of class predictions. Pre presents the bounding box predictions.
Figure 2. The overall structure of the MHLDet. We embedded the SimAM attention mechanism and VNFE module in the backbone and used the SPPE and CACL structure in the neck network. H, W, and C represent height, width, and channels, respectively. Cls represents the number of class predictions. Pre presents the bounding box predictions.
Remotesensing 15 04625 g002
Figure 3. Structure of the VNFE module.
Figure 3. Structure of the VNFE module.
Remotesensing 15 04625 g003
Figure 4. Structure of the SPPCSPC and SPPE. (a) represents the SPPCSPC; (b) represents the SPPE.
Figure 4. Structure of the SPPCSPC and SPPE. (a) represents the SPPCSPC; (b) represents the SPPE.
Remotesensing 15 04625 g004
Figure 5. Structure of the CACL module.
Figure 5. Structure of the CACL module.
Remotesensing 15 04625 g005
Figure 6. Some examples of the remote sensing images utilized for training and evaluating our method from the SIMD and UCAS-AOD datasets.
Figure 6. Some examples of the remote sensing images utilized for training and evaluating our method from the SIMD and UCAS-AOD datasets.
Remotesensing 15 04625 g006
Figure 7. Comparison of the mAP and parameters of different detectors on the SIMD dataset.
Figure 7. Comparison of the mAP and parameters of different detectors on the SIMD dataset.
Remotesensing 15 04625 g007
Figure 8. Heatmap visualization results on the SIMD dataset. (a1a3) represents the original images; (b1b3) represents the heatmap visualization results of YOLOv7-Tiny; (c1c3) represents the heatmap visualization results of the method we proposed.
Figure 8. Heatmap visualization results on the SIMD dataset. (a1a3) represents the original images; (b1b3) represents the heatmap visualization results of YOLOv7-Tiny; (c1c3) represents the heatmap visualization results of the method we proposed.
Remotesensing 15 04625 g008aRemotesensing 15 04625 g008b
Figure 9. The visualization results of baseline and MHLDet on the SIMD dataset. (a1a6) represents the original images; (b1b6) represents the visualization results of YOLOv7-Tiny; (c1c6) represents the visualization results of the method we proposed. The yellow boxes show the differences between YOLOv7-Tiny and MHLDet.
Figure 9. The visualization results of baseline and MHLDet on the SIMD dataset. (a1a6) represents the original images; (b1b6) represents the visualization results of YOLOv7-Tiny; (c1c6) represents the visualization results of the method we proposed. The yellow boxes show the differences between YOLOv7-Tiny and MHLDet.
Remotesensing 15 04625 g009aRemotesensing 15 04625 g009b
Figure 10. Comparison of different visualization results on the UCAS-AOD dataset. (a1a6) represents the original image; (b1b6) represents the visualization results of YOLOv7-Tiny; (c1c6) represents the visualization results of the method we proposed. The yellow boxes show the differences between YOLOv7-Tiny and MHLDet.
Figure 10. Comparison of different visualization results on the UCAS-AOD dataset. (a1a6) represents the original image; (b1b6) represents the visualization results of YOLOv7-Tiny; (c1c6) represents the visualization results of the method we proposed. The yellow boxes show the differences between YOLOv7-Tiny and MHLDet.
Remotesensing 15 04625 g010aRemotesensing 15 04625 g010b
Figure 11. Comparison of experimental mAP for ablation. A: SimAM; B: VNFE; C: SPPE; D: CACL; E: SimAM + VNFE; F: SimAM + VNFE + SPPE; G: SimAM + VNFE + SPPE + CACL.
Figure 11. Comparison of experimental mAP for ablation. A: SimAM; B: VNFE; C: SPPE; D: CACL; E: SimAM + VNFE; F: SimAM + VNFE + SPPE; G: SimAM + VNFE + SPPE + CACL.
Remotesensing 15 04625 g011
Table 1. Parameters of MHLDet.
Table 1. Parameters of MHLDet.
NumberModuleFromNOutputParams
Input 1024 × 1024 × 3
0CBL−11 512 × 512 × 32 928
1CBL−11 256 × 256 × 64 18,560
2ELAN−11 256 × 256 × 128 31,104
3max pool−11 128 × 128 × 128 0
4SimAM−11 128 × 128 × 64 0
5ELAN−11 128 × 128 × 128 115,456
6max pool−11 64 × 64 × 128 0
7VNFE−11 64 × 64 × 256 242,432
8max pool−11 32 × 32 × 256 0
9VNFE−11 32 × 32 × 512 966,144
10SPPE−11 32 × 32 × 256 936,448
11CBL−11 32 × 32 × 128 33,024
12Upsample−11 64 × 64 × 128 0
13CBL71 64 × 64 × 128 33,024
14Concat(−1,12)1 64 × 64 × 256 0
15ELAN−11 64 × 64 × 128 140,032
16CBL−11 64 × 64 × 64 8320
17Upsample−11 128 × 128 × 64 0
18CBL51 128 × 128 × 64 8320
19Concat(−1,17)1 128 × 128 × 128 0
20ELAN−11 128 × 128 × 64 35,200
21CACL−11 64 × 64 × 128 82,432
22Concat(−1,15)1 64 × 64 × 256 0
23ELAN−11 64 × 64 × 128 140,032
24CACL−11 32 × 32 × 256 328,704
25Concat(−1,10)1 32 × 32 × 512 0
26ELAN−11 32 × 32 × 256 558,592
27CBL201 128 × 128 × 128 73,984
28CBL231 64 × 64 × 256 295,424
29CBL261 32 × 32 × 512 1,180,672
30Detect 1355,016
Total 5,283,848
Table 2. The initialization training parameters on the SIMD dataset.
Table 2. The initialization training parameters on the SIMD dataset.
Input SizeBatch SizeMomentumWarm-Up MomentumWeight DecayLearning Rate
1024 × 1024 40.9370.80.00050.01
Table 3. The results of YOLOv7-Tiny with and without the K-Means++ algorithm on the SIMD dataset.
Table 3. The results of YOLOv7-Tiny with and without the K-Means++ algorithm on the SIMD dataset.
AlgorithmmAP (%)mAP0.5:0.95 (%)Params (m)FLOPs (G)
YOLOv7-Tiny82.364.26.0513.3
YOLOv7-Tiny+K-Means++82.764.86.0513.3
Table 4. Categories and their corresponding names on the SIMD dataset.
Table 4. Categories and their corresponding names on the SIMD dataset.
CategoryANCategoryANCategoryAN
CarCRAirlinerALOthersOT
TruckTRPropeller AircraftPAStair TruckST
VanVANTrainer AircraftTAPushback TruckPT
Long VehicleLVChartered AircraftCAHelicopterHC
BusBUSFighter AircraftFABoatBO
Table 5. The detection accuracy of each category on the SIMD dataset.
Table 5. The detection accuracy of each category on the SIMD dataset.
ModelCRTRVANLVBUSALPATACAFAOTSTPTHCBOmAP
TOOD [43]63.758.957.256.767.983.078.478.578.976.510.924.115.230.875.574.3
RetinaNet [12]66.554.657.952.962.384.772.676.677.084.67.221.65.333.473.371.7
FCOS [25]69.558.261.752.559.979.170.773.972.182.516.226.29.142.072.575.9
YOLOv5-N [48]94.283.284.583.291.896.385.295.292.988.823.144.615.558.397.975.7
YOLOv5-S [48]94.485.286.783.492.796.494.896.596.392.633.854.247.788.998.482.8
YOLOv7-Tiny [17]94.585.686.286.393.797.696.695.994.299.528.150.548.378.298.682.3
Ours94.785.787.587.294.797.696.197.895.894.630.154.558.896.898.984.7
Bold text is the optimal value for the column.
Table 6. Comparison with state-of-the-art lightweight detectors on the SIMD dataset.
Table 6. Comparison with state-of-the-art lightweight detectors on the SIMD dataset.
MethodmAPmAP0.75 (%)mAP0.5:0.95 (%)APs (%)APm (%)APl (%)Params (m)FLOPs (G)
Anchor-free
FCOS (2019) [25]75.967.256.45.947.461.231.87201.9
TOOD (2021) [43]74.368.857.112.349.463.331.98188.7
PP-PicoDet-m (2021) [44]72.967.756.715.750.361.63.43-
YOLOX-S (2021) [45]80.374.862.712.259.567.88.9468.5
YOLOv6-N (2022) [46]74.469.058.58.249.363.14.711.4
YOLOv6-S (2022) [46]78.973.162.78.557.869.218.545.3
YOLOv8-N (2023) [47]81.6-65.9---3.018.2
Anchor-based
YOLO-HR-N (2022) [49]83.0-64.0---3.344.4
RetinaNet (2017) [12]71.765.555.45.447.761.536.39215.5
YOLOv5-N (2020) [48]75.769.357.915.149.662.01.784.3
YOLOv5-S (2020) [48]82.874.964.411.261.968.07.0616.1
YOLOv7-Tiny (2022) [17]82.375.464.219.359.570.86.0513.3
MHLDet (ours)84.778.366.820.161.972.15.2812.2
Table 7. The detection accuracy of different methods on the UCAS-AOD dataset.
Table 7. The detection accuracy of different methods on the UCAS-AOD dataset.
MethodAPs (%)APm (%)APl (%)mAP
YOLOv5-N [48]17.254.550.195.9
YOLOv6-N [46]11.453.844.293.63
YOLOv6-S [46]9.557.449.194.9
YOLOv7-Tiny [17]18.257.448.295.4
MHLDet (Ours)23.857.652.196.4
Table 8. The experimental results of the ablation of each module.
Table 8. The experimental results of the ablation of each module.
MethodSimAMVNFESPPECACLmAPPrecisionRecallParams (m)FLOPs (G)
Baseline 82.377.379.56.0513.3
A 83.684.178.66.0513.3
B 83.776.482.74.9611.9
C 83.776.682.76.0913.4
D 83.477.282.06.3313.5
E 84.176.383.84.9611.9
F 84.477.682.45.2412.1
G84.777.182.55.2812.2
Bold text is the optimal value for the column.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, L.; Zhao, H.; Liu, Z.; Cai, K.; Liu, Y.; Zuo, X. MHLDet: A Multi-Scale and High-Precision Lightweight Object Detector Based on Large Receptive Field and Attention Mechanism for Remote Sensing Images. Remote Sens. 2023, 15, 4625. https://doi.org/10.3390/rs15184625

AMA Style

Zhou L, Zhao H, Liu Z, Cai K, Liu Y, Zuo X. MHLDet: A Multi-Scale and High-Precision Lightweight Object Detector Based on Large Receptive Field and Attention Mechanism for Remote Sensing Images. Remote Sensing. 2023; 15(18):4625. https://doi.org/10.3390/rs15184625

Chicago/Turabian Style

Zhou, Liming, Hang Zhao, Zhehao Liu, Kun Cai, Yang Liu, and Xianyu Zuo. 2023. "MHLDet: A Multi-Scale and High-Precision Lightweight Object Detector Based on Large Receptive Field and Attention Mechanism for Remote Sensing Images" Remote Sensing 15, no. 18: 4625. https://doi.org/10.3390/rs15184625

APA Style

Zhou, L., Zhao, H., Liu, Z., Cai, K., Liu, Y., & Zuo, X. (2023). MHLDet: A Multi-Scale and High-Precision Lightweight Object Detector Based on Large Receptive Field and Attention Mechanism for Remote Sensing Images. Remote Sensing, 15(18), 4625. https://doi.org/10.3390/rs15184625

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop