Next Article in Journal
Cloud/VPN-Based Remote Control of a Modular Production System Assisted by a Mobile Cyber–Physical Robotic System—Digital Twin Approach
Previous Article in Journal
Diagnosis of Reverse-Connection Defects in High-Voltage Cable Cross-Bonded Grounding System Based on ARO-SVM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism

School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(2), 589; https://doi.org/10.3390/s25020589
Submission received: 30 December 2024 / Revised: 15 January 2025 / Accepted: 16 January 2025 / Published: 20 January 2025
(This article belongs to the Section Remote Sensors)

Abstract

:
With the rapid development of AI algorithms and computational power, object recognition based on deep learning frameworks has become a major research direction in computer vision. UAVs equipped with object detection systems are increasingly used in fields like smart transportation, disaster warning, and emergency rescue. However, due to factors such as the environment, lighting, altitude, and angle, UAV images face challenges like small object sizes, high object density, and significant background interference, making object detection tasks difficult. To address these issues, we use YOLOv8s as the basic framework and introduce a multi-level feature fusion algorithm. Additionally, we design an attention mechanism that links distant pixels to improve small object feature extraction. To address missed detections and inaccurate localization, we replace the detection head with a dynamic head, allowing the model to route objects to the appropriate head for final output. We also introduce Slideloss to improve the model’s learning of difficult samples and ShapeIoU to better account for the shape and scale of bounding boxes. Experiments on datasets like VisDrone2019 show that our method improves accuracy by nearly 10% and recall by about 11% compared to the baseline. Additionally, on the AI-TODv1.5 dataset, our method improves the mAP50 from 38.8 to 45.2.

1. Introduction

As object detection technology advances, unmanned aerial vehicles (UAVs) are increasingly being applied in diverse areas including traffic monitoring [1], power inspections [2], crop analysis [3], and disaster relief. For instance, in the field of traffic monitoring, UAVs can perform aerial surveillance, unconstrained by road limitations, with advantages such as their high speed, high degree of freedom of movement, and wide field of view. However, most studies have focused on the analysis of ground-based monitoring images, and object detection algorithms for UAV-perspective images have not yet been sufficiently explored. The main reasons are the challenges presented by UAV-perspective images, including scale variation, class imbalance, uneven sample distribution, a high number of objects predominantly small in size, and the difficulty of balancing the high computational demands of high-resolution UAV images with the limited computing power of current low-power chips. Compared to ground-level natural images, UAV-perspective images often require high-resolution imagery to extract features effectively due to the large field of view, uneven lighting, and dense object distribution. This significantly increases the computational and memory requirements of object detection algorithms. Employing unoptimized general object detection algorithms requires models with extensive parameters and significant computational power to maintain detection accuracy, leading to prohibitive computational and memory expenses. Consequently, enhancing the performance of object detectors while operating under resource limitations is a key challenge in UAV image object detection.
Object detection is a fundamental technology in computer vision, designed to allow computers to identify and pinpoint objects in images. This process involves not only recognizing the objects but also determining their locations within the image and classifying them into specific categories. In the early 21st century, most object detection used traditional algorithms, which generally involved extracting candidate boxes, extracting features from these candidate boxes, and then classifying them using a classifier. However, the performance and stability of these methods were generally poor. As a result, deep learning methods are now widely employed for object detection. Since the introduction of AlexNet [4] in 2012, the success of deep neural networks has been proven, marking the beginning of an era dominated by deep learning-based detection techniques. In today’s computer vision domain, deep learning-based object detection methods are primarily categorized into two types: single-stage and two-stage object detection methods. The RCNN [5,6,7] series, which first combined deep learning with object detection, is a typical two-stage detector. It generates numerous region proposals in the image after feature extraction, performs a simple foreground–background judgment, and then feeds the proposed regions into the network for further classification and the refinement of the detection boxes. Although two-stage object detection has certain advantages in terms of accuracy, it has significant issues in model complexity and real-time performance. The single-stage object detection method, on the other hand, directly classifies and locates features extracted from the input image via neural networks, bypassing the candidate box generation stage. Despite the resulting problem of imbalanced positive and negative samples, the good detection speed and real-time performance of the single-stage object detection method have led to its widespread use in daily production and life. Representative single-stage object detection algorithms include SSD [8], YOLO [9], and RetinaNet [10]. Additionally, since ViT [11] directly applied the Transformer encoder to image feature extraction, achieving cross-domain migration from natural language processing to the vision field, it has provided a new approach to handling visual tasks. DETR [12] represents a bold attempt to combine Transformers with object detection, eliminating the need for predefined anchors in the image and the use of NMS for post-processing. As an end-to-end object detector, DETR has achieved impressive results in object detection tasks. However, DETR also has significant drawbacks, including high computational complexity and the need for large amounts of data and computational resources for training. Given the same computational costs, it does not have a significant advantage over some CNN networks. However, in some natural scenes, object scales vary, shapes are irregular, and UAV-based object detection frequently involves focusing on small objects. Consequently, the mainstream object detection algorithms discussed earlier are not directly suitable for UAV aerial detection tasks.
Recently, researchers have dedicated substantial efforts to enhancing the accuracy of detecting small objects [13], addressing this challenge through dataset augmentation methods, including segmenting images to enhance the representation of small objects in each segment and inserting small objects into different backgrounds. This approach increases the number of small object samples and improves detector performance. Deng et al. [14] introduced an ultra-high-resolution, extended feature pyramid network (FPN) tailored to small object detection, along with the FTT module, which enhances feature extraction while minimizing detail loss in the feature map. Gong et al. [15] suggested that while the top-down connections between adjacent layers in FPN benefit small object detection, they also present some limitations. They introduced a novel concept—fusion factor—to manage the information flow from deeper to shallower layers, thereby optimizing FPN for detecting small objects. Noh et al. [16] enhanced detection performance by utilizing high-resolution target features as supervisory signals and aligning the relative receptive fields of input and target features. Additionally, QueryDet [17] initially estimates the approximate positions of small objects on a low-resolution feature map and subsequently creates a sparse feature map from the high-resolution features of these estimated positions to achieve precise detection results. This approach maximizes the use of high-resolution feature maps while avoiding unnecessary computations in background regions. Regarding loss functions, since IoU [18] (Intersection over Union) is highly sensitive to bounding box deviation for small objects, Wang et al. [19] developed a new metric tailored to the characteristics of small targets. They reformulated the bounding box as a two-dimensional Gaussian distribution, converting the Intersection over Union (IoU) between predicted and ground truth boxes into a measure of similarity between the two distributions. This similarity was then normalized to create a new evaluation metric, NWD (Normalized Wasserstein Distance), which quantifies the similarity between the distributions.
Inspired by the aforementioned research, we propose a UAV-perspective-based object detection model. This model uses YOLOv8 as the backbone network and introduces modifications to its structure and modules to enhance detection performance without consuming too many resources, ensuring a real-time capability. This paper’s primary contributions are as follows:
  • We replaced the PAN (Path Aggregation Network) in YOLOv8 with HSPAN to better fuse features at different scales. And we incorporated a feature-enhanced attention mechanism called Content Anchor Attention, which strengthens central features and captures contextual relationships between edge pixels, thereby enhancing the model’s classification and localization abilities.
  • We augmented the YOLOv8 detection head with Dynamic Head to improve the detection head’s scale awareness, spatial awareness, and task awareness for small objects in images. This enhancement increased the model’s ability to perceive small objects, optimizing the detector’s performance.
  • During model training, we introduced the Slideloss classification loss function and the ShapeIoU localization loss function. These functions strengthen the model’s learning of difficult samples, address issues arising from sensitivity of IoU to small object shifts, and accelerate the model’s convergence.
The rest of this paper is organized as follows. Section 2 presents some work related to this study. Section 3 presents the framework of the deep learning network used and the methods proposed in this work. In Section 4, the experimental results are provided and analyzed. Finally, the paper is concluded in Section 5.

2. Related Work

2.1. Yolo Algorithm

Since its release, the YOLO object detector has been renowned for its fast detection speed and has achieved an excellent balance between detection speed and accuracy through successive improvements. Redmon et al. introduced YOLOv1 [9] with the feature extraction network Darknet24 in 2016. YOLOv2 [20] later introduced the anchor boxes mechanism and selected a deeper backbone network for feature extraction. YOLOv3 [21] adopted multi-scale prediction, enabling the detection of objects at different scales. YOLOv4 [22] improved detection performance by using the advanced CSPDarknet. YOLOv5, while maintaining model accuracy, focused on optimizing inference speed, making it ideal for real-time applications. YOLOv6 [23] replaced the SPPF in the backbone with SimSPPF, which better extracts global features. The E-ELAN module in YOLOv7 [24] accelerates model convergence and reduces training time. YOLOv8 removed the anchor box mechanism and adopted the TOOD [25] strategy for positive and negative sample allocation, further optimizing the model. YOLOv9 [26] effectively addresses the problem of information loss in deep neural networks. YOLOv10 [27] boldly eliminates the NMS post-processing operation and directly produces detection results, becoming the first end-to-end detector in the YOLO series. The recently released YOLO11 adopts an improved backbone and neck architecture, enhancing feature extraction capabilities to achieve more accurate object detection and complex task performance. With continuous model improvements, the YOLO series has consistently optimized both accuracy and speed, leading the development of object detectors in the first stage.

2.2. Multi-Scale Feature Fusion

Multi-scale feature fusion can capture the details and characteristics of target objects at different scales, providing varied contextual information from different scales. By integrating information from multiple scales, it can effectively consolidate contextual information and improve the performance of recognition tasks. In computer vision tasks, multi-scale feature fusion can better handle scale variations and provide more accurate location and positional information. As a result, building multi-scale feature representations has consistently been a central research area in object detection. Since the advent of the feature pyramid network (FPN) [28], it has provided an effective solution for multi-scale feature fusion. This architecture enhances the feature extraction capabilities of CNN networks, ensuring that the final output features more accurately capture the dimensional information of the input image. The basic process involves three steps: the bottom-up pathway, which generates features of different dimensions from the bottom up; the top-down pathway, which enhances features from top to bottom; and the lateral connections between the CNN network layers and the final output features, using 1 × 1 convolutions to generate optimal output features. Building on the influence of FPN, PANet [29] adds a downsampling pathway to FPN, incorporating more semantic information into shallow features. NAS-FPN [30] utilizes neural architecture search to discover a new feature pyramid structure within a novel, scalable space that encompasses all cross-scale connections. However, finding the optimal architecture involves significant search costs. The authors of BiFPN [31] observed that the contributions of features at different scales to the fused output are often uneven. To address this, they proposed BiFPN, a simple yet effective weighted bidirectional feature pyramid network that introduces learnable weights to assess the significance of various input features and repeatedly applies top-down and bottom-up multi-scale feature fusion. In contrast, some researchers argue that the simple stacking of BiFPN modules is not optimal, as each BiFPN module operates independently, and deep stacking may risk gradient vanishing. Based on this, GFPN [32] was proposed, designing a new mechanism for skip-layer connection and cross-scale connection to balance accuracy and efficiency during model expansion. Another approach, AFPN [33], introduced a progressive feature pyramid network to support direct interactions between non-adjacent levels, avoiding significant semantic gaps between non-adjacent levels. Similarly, to ensure that the neck can transmit information losslessly during cross-layer fusion, Goldyolo [34] proposed a new collection and distribution mechanism composed of a Feature Alignment Module (FAM), an Information Fusion Module (IFM), and an Information Injection Module (Inject). This mechanism globally fuses multi-scale features and injects global attention into the higher levels, significantly enhancing the neck’s information fusion capability and improving model performance across different object sizes.

2.3. Attention Mechanism

The concept of attention mechanisms originated in the 1990s when cognitive scientists discovered that humans naturally filter out less relevant information and focus on information of interest. This process was termed the “attention mechanism”. In traditional machine learning, attention is exemplified through feature engineering, where input data are transformed into numerical vectors, helping the model select effective, appropriately scaled features. In 2014, Volodymyr Mnih introduced attention mechanisms to the vision domain with the paper “Recurrent Models of Visual Attention” [35]. This was followed by Ashish Vaswani’s 2017 paper “Attention is All You Need” [36], which proposed the Transformer structure. Since then, attention mechanisms have been widely applied in network designs for NLP and CV-related problems. Hu et al. proposed SENet [37] (Squeeze and Excitation Network), which applies attention to the channels of feature maps, enabling the model to learn the weights between different channels. This approach is simple, convenient, and adds only a small amount of computational and parameter overhead. GSoPNet [38] improved the squeeze part of SENet by replacing the original first-order average pooling with second-order pooling, enhancing global statistical modeling compared to SENet. The CBAM [39] attention mechanism is a hybrid attention mechanism that first processes the input feature map through a channel attention module and then through a spatial attention module. This improves the model’s perception ability and performance without increasing network complexity. However, CBAM’s approach to leveraging positional information by reducing the number of channels and using large-size convolutions primarily captures local correlations, failing to model long-range dependencies, which are essential for visual tasks. Coordinate Attention (CA) [40] encodes channel relationships and long-range dependencies by incorporating precise positional information. It achieves this by decomposing global pooling into the height and width directions, capturing cross-channel information as well as direction-aware and position-aware data, which helps the model more accurately locate and identify targets of interest. Triplet Attention [41] includes three parallel branches: two focus on capturing cross-dimensional interactions between channels (C) and spatial dimensions (H or W), while the third branch constructs spatial attention. The outputs of these three branches are averaged to mitigate the separation issue between channel attention and spatial attention. SimAM [42] leverages the property that adjacent pixels in an image have high similarity, while distant pixels have low similarity. It generates attention weights by calculating the similarity between each pixel in the feature map and its neighboring pixels.

2.4. Detection Head

The detection head is the final component of an object detection network, responsible for generating the final detection results, including object categories and locations. The detection head enables features extracted by the backbone network, initially designed for image classification, to be effectively transformed into features suitable for object detection. This further enhances the network’s performance in handling complex scenes and diverse objects, making it a key factor in improving detection accuracy. Early object detectors, such as YOLOv3 and YOLOv4, used a coupled head, where the detection head simultaneously predicted both object categories and bounding box locations. However, this design has some limitations. Song et al. [43] suggested that category prediction and location detection focus on different regions of interest in the image. Category prediction is a classification problem that focuses on the object’s center, while location detection is a regression problem that focuses on the object’s edges. Moreover, combining category and location detection into a single head may cause errors in one task to affect the other. To address this, Wu et al. [44] introduced a decoupled head, employing two separate branches within the detection head: one for predicting categories and the other for predicting locations. The subsequent YOLOX [45] model also adopted this decoupled head, achieving notable improvements in detection performance. Zhuang et al. [46] propose a novel Task-Specific Context Decoupling head. They input feature maps rich in semantic information into the classification branch and high-resolution feature maps containing more edge information into the localization branch to enhance detection performance. Jiang et al. [47] argued that the classification and localization branches in typical decoupled heads lack sufficient interaction. To address this, they proposed a new detection head, G-Head, which enhances the interaction between different tasks and facilitates the process of multi-task learning.

3. Method

YOLOv8 is one of the most advanced object detection algorithms available today, employing state-of-the-art methods in positive and negative sample selection strategies, feature extraction, and multi-scale feature fusion, making it suitable for most everyday scenarios. However, due to the characteristics of UAV perspective images, such as the prevalence of small targets, dense and uneven distribution, YOLOv8 struggles to achieve satisfactory detection performance in these images, rendering it inadequate for UAV aerial scenes. To address these issues, we have optimized the YOLOv8s base model by focusing on multi-scale feature fusion, detection heads, attention mechanisms, and loss functions. Our main improvement strategies are as follows: First, we introduced the concept of the High-level Screening-feature Pyramid Network (HS-FPN) [48] to enhance the neck of YOLOv8 that integrates multi-scale features, and then upgraded it to HS-PAN, similar to PANet, to better extract semantic information from input images. Additionally, we replaced the original channel attention mechanism with Context Anchor Attention [49], which is more suitable for small targets, thereby making the model more focused on small object features. Furthermore, we integrated an additional detection head after the 160 × 160 feature map in the original YOLOv8, and replaced all detection heads with Dynamic Head [50], which possesses scale awareness, spatial awareness, and task awareness, thus enhancing the detection heads’ perception of small objects. Finally, we improved the classification and localization loss functions during model training to ensure the better convergence of the model. The overall structure of our proposed model is shown in Figure 1.

3.1. CAA-HSPAN (Context Anchor Attention High-Level Screening-Feature Path Aggregation Network)

In UAV-perspective images, the objects to be detected exist at different scales. Even objects of the same category can vary in size and features due to distance variations within the same viewpoint, presenting a significant challenge for successful detection. To address the multi-scale issue inherent in UAV perspective images, we designed the Context Anchor Attention High-level Screening-feature Path Aggregation Network (CAA-PAN) to achieve multi-scale feature fusion, as detailed in Figure 2. This network structure is an improvement based on HS-FPN, and experimental results demonstrate that this design enables the model to capture features in UAV images more comprehensively, thereby improving the model’s performance.

3.1.1. Feature Enhancement Module

The feature selection module of HS-FPN consists of the CA module and Dimensional Matching (DM). The CA module focuses on extracting the most representative information from each channel. However, we argue that extracting valuable spatial information without adding extra parameters is more advantageous for object detection in UAV-perspective images. Therefore, we replaced the original Channel Attention (CA) module with the Context Anchor Attention (CAA) module. The CAA module enhances central features while grasping contextual interdependencies among distant pixels, which is highly beneficial for small object localization. The CAA module first performs an average pooling and a 1 × 1 convolution operation on the input feature map to obtain local regional features:
F p o o l = C o n v 1 × 1 ( P a v g ( X ) ) ,
where P a v g denotes average pooling and X represents the input feature map. Then, two depthwise strip convolutions are used to replace the standard depthwise convolution.
F w = D W C o n v 1 × k b ( F p o o l ) ,
F w = D W C o n v 1 × k b ( F p o o l ) ,
There are three main reasons for using depthwise strip convolutions. First, strip convolutions are a lightweight convolution method. Compared to standard 2D depthwise convolutions, using two 1D strip convolutions can significantly reduce computational costs while achieving similar effects. Second, strip convolutions enhance the recognition and extraction of elongated target features, such as cars and pedestrians. Third, strip convolutions can connect distant pixels without significantly increasing computational complexity.
Finally, the attention weights are created using a Sigmoid function and then multiplied with the input feature map.
A = S i g m o i d ( C o n v 1 × 1 ( F h ) ) ,
F a t t = A × X ,
Subsequently, feature fusion is performed, but first, dimensional matching is needed for the feature maps of different scales due to their varying numbers of feature channels. Here, the DM module uses 1 × 1 convolutions to adjust the number of channels for each scale’s feature map to 256.

3.1.2. Feature Fusion Module

Features from the backbone network typically exhibit two key traits: high-level features are rich in semantic information but lack precise target location details, while low-level features provide accurate location information but are less effective in semantic classification. To address these issues, a common approach is to use a feature pyramid structure, upsample the high-level feature maps, and then add or concatenate them with low-level feature maps. Subsequently, the resulting feature maps are downsampled and again added or concatenated with the previous feature maps. However, this approach lacks a feature selection phase, simply performing elementwise addition or channel concatenation directly. Therefore, we designed the USFF and DSSF modules. The USFF module employs feature maps with rich semantic information as weights to filter out irrelevant semantic details from the low-level feature maps. Conversely, the DSSF module uses feature maps containing spatial location information as weights to improve the spatial accuracy of high-level features. As shown in Figure 3a, given a high-level f h i g h R C × H × W feature map and a low-level f l o w R C × 2 H × 2 W feature map, the high-level feature map is first subjected to a transpose convolution operation with a stride of 2, doubling the height and width of the feature map to f h i g h R C × 2 H × 2 W .
Next, the CAA module is applied to transform the features into corresponding weights, which are then used to filter the low-level feature maps. Finally, the filtered feature maps are fused with the high-level features f o u t R C × 2 H × 2 W , enhancing the model’s feature representation capability; the DSSF module in Figure 3b operates similarly to the USFF module, with the key difference being that larger feature maps are downsampled before being processed by the CAA module to filter the smaller feature maps. The process of feature selection and fusion is as follows:
f 1 = T C o n v ( f ) ,
f o u t = C A A ( f 1 ) × f 2 + f 1 ,

3.2. Dynamic Head

To improve the network’s capability in identifying small-scale objects, we added a detection head after the p2 layer, a 4× downsampled feature map. This is because, when downsampling small objects by 4×, most of the features of the small objects have not completely disappeared from the feature map. The 4× downsampled feature map can well preserve the spatial and semantic features of small objects, which is very beneficial for UAV-perspective object detection tasks. In addition, we replaced the four detection heads with dynamic heads, effectively improving the performance of the object detector. The detailed structure of the Dyhead is shown in Figure 4.
For object detection in UAV perspectives, an image often contains multiple objects of different scales, so the head should be scale-aware. Additionally, since objects can have different shapes, angles, and positions in different views, the head should be spatial-aware. Moreover, because objects can be expressed in various forms, such as a bounding box, center, points, etc., each with different target functions and optimization methods, the head should be task-aware. The dynamic head can unify scale awareness, spatial awareness, and task awareness. Suppose the feature map output from the backbone is level*space*channel. The dynamic head uses different attention mechanisms across different dimensions: attention in the level dimension can learn the relative importance of semantic layers and enhance object features in UAV-perspective images based on scale; attention in the spatial dimension can learn distinctive feature representations in space; and attention in the channel dimension can improve the feature representation capability of different channels. For the multi-scale L-layer feature maps F i n = { F i } i = 1 L output from the backbone, they are first resized to the size of the middle-scale feature map, resulting in a 4D tensor F R L × H × W × C , where L represents the number of feature maps input to the detection head, and H, W, C represent the height, width, and feature channels of the middle-scale feature map, respectively. Then, define S = H × W , and reshape the 4D feature into a 3D feature F R L × S × C . Finally, apply self-attention to it, with the general formula as follows:
W ( F ) = π ( F ) · F ,
π ( · ) is an attention function. Nonetheless, directly learning the attention function across all dimensions proves to be computationally demanding and necessitates considerable memory resources. Consequently, the process is divided into three sequential attentions, with each sub-attention dedicated to a single dimension:
W ( F ) = π C ( π S ( π L ( F ) · F ) · F ) · F ,
In the formula, π L ( · ) , π S ( · ) ,   π C ( · ) are different attention mechanisms applied to the L , S , and C dimensions, respectively.

3.2.1. Scale-Aware Attention

Utilize scale-aware attention to adaptively integrate features of varying scales according to their semantic significance.
π L ( F ) · F = σ ( f ( 1 S C S , C F ) ) · F ,
f ( · ) is a linear function implemented using a 1 × 1 convolution, and σ ( x ) = max ( 0 , min ( 1 , x + 1 2 ) ) is a hard sigmoid function.

3.2.2. Spatial-Aware Attention

Using spatial-aware attention, the network is guided to focus on abrupt changes in spatial positions and feature levels. Given the large S dimension, this involves two steps: first, using deformable convolutions to enable the attention to learn the coefficient features, and subsequently aggregating features at corresponding locations across different scale feature maps.
π S ( F ) · F = 1 L l = 1 L k = 1 K ω l , k · F ( l ; p k + Δ p k ; c ) · Δ m k ,
Here, K is the number of sparse sampling locations, p k + Δ p k is the sampling position after learning the offset, allowing the attention mechanism to focus on abrupt changes, and Δ m k represents the importance learned at position p k .

3.2.3. Task-Aware Attention

By employing task-aware attention, we can dynamically control each channel for different tasks, using the most appropriate channels for predictions.
π C ( F ) · F = max ( α 1 ( F ) · F C + β 1 ( F ) , α 2 ( F ) · F C + β 2 ( F ) ) ,
F C is the slice on the C -th channel, [ α 1 , α 2 , β 1 , β 2 ] = θ ( · ) are hyperparameters that learn to control the activation threshold; initially, global average pooling is applied across the L * S dimensions to reduce dimensionality. This is followed by two fully connected layers, a normalization layer, and a sigmoid activation function, normalizing to [−1, 1].

3.3. Loss Function

In aerial drone scenarios for object detection tasks, small objects comprise a significant proportion. An appropriately designed loss function can substantially improve the model’s detection performance. YOLOv8 employs DFL [51] and CIoU [52] losses for bounding box regression, alongside BCE loss for classification. However, BCE loss does not effectively balance between challenging and simpler samples. To address this, we introduce SlideLoss [53], which differentiates between hard and easy samples based on the Intersection over Union (IoU) [18] between predicted and ground truth boxes. To reduce hyperparameters, the loss uses the average IoU of all bounding boxes as a threshold µ ; samples with classification scores below µ are considered hard, and those above µ are considered easy. Moreover, ambiguous samples near classification boundaries often incur large losses, which SlideLoss guides the model to optimize and effectively utilize for training. Given their scarcity, these samples receive greater weight during loss computation. The specific process involves determining hard and easy samples using parameter µ , followed by a weighting function that emphasizes samples with ambiguous classifications.
f ( x ) = { 1 ,                 x μ 0.1 e 1 μ ,   μ < x < μ 0.1 e 1 x ,                     x μ ,
In object detection, bounding box regression losses are typically measured using IoU and its various variants, including GIoU [54], DIoU [55], CIoU, EIoU [56], and SIoU [57], among others. Although these methods account for factors such as the distance, shape, and angle between predicted and ground truth boxes, they often neglect the effects of the bounding box’s shape and scale itself. In Figure 5a, although the deviation between A and B is identical to that between C and D, the Intersection over Union (IoU) values differ. Similarly, while the deviation between C and D remains consistent, the IoU values vary. However, the difference in IoU values between C and D is less pronounced compared to the difference between A and B. In Figure 5b, a similar pattern is observed between A and B, as well as between C and D. In contrast, Figure 5a illustrates that the variation in IoU between A and B is attributable to differences in the shapes of the ground truth boxes, with deviations corresponding, respectively, to the directions of the long and short sides. It can be seen that deviations along the direction of the long side have a smaller impact on IoU compared to deviations along the direction of the short side. In Figure 5b, it is observed that the shape of the ground truth affects the variation in IoU values during regression. Here, we use Shapeiou [58] to address the above-mentioned issues. Its computational formula is as follows:
I o U = | B B g t | | B B g t | ,
w w = 2 × ( w g t ) s c a l e ( w g t ) s c a l e + ( h g t ) s c a l e ,
h h = 2 × ( h g t ) s c a l e ( w g t ) s c a l e + ( h g t ) s c a l e ,
d i s t a n c e s h a p e = h h × ( x c x c g t ) 2 / c 2 + w w × ( y c y c g t ) 2 / c 2 ,
Ω s h a p e = t = w , h ( 1 e ω t ) θ , θ = 4 ,
{ ω w = h h × | w w g t | m a x ( w , w g t ) ω h = w w × | h h g t | m a x ( h , h g t ) ,
where scale is a parameter related to the dataset, and w w and h h are the weight coefficients in the horizontal and vertical directions, respectively, which are related to the shape of the ground truth. The bounding box regression loss is as follows:
L S h a p e I o U = 1 I o U + d i s t a n c e s h a p e + 0.5 × Ω s h a p e ,
Ultimately, we use SlideLoss and ShapeIoU+DFL loss as the classification loss and localization loss for our model, respectively. Experimental results demonstrate that this approach can improve the performance of the object detector.

4. Experiments

This chapter primarily introduces the dataset used in this study, the experimental environment, and the training strategies. Additionally, it presents the evaluation metrics related to the experimental results.

4.1. Datasets

The VisDrone2019 dataset [59], developed by Tianjin University and the AISKYEYE data mining team, is a prominent resource for aerial drone imagery. It encompasses scenes from over ten cities across China, captured using various drones under diverse angles, scenarios, and tasks. This results in a dataset rich in information, covering different environments (urban and rural), object types (pedestrians, vehicles, bicycles, etc.), and densities (sparse to crowded). The dataset includes 10,209 static images (6471 for training, 548 for validation, and 3190 for testing), with more than 2.6 million manually annotated bounding boxes. Additionally, it provides crucial attributes such as scene visibility, object categories, and occlusion conditions to enhance data utilization. The VisDrone2019 dataset is a high-quality dataset specifically designed for object detection from UAV perspectives. It effectively represents the scenarios that UAVs typically encounter in daily use, aligning with the content and issues addressed in this study. Therefore, we conducted comparative and ablation experiments on this dataset.
The AI-TODv1.5 dataset [60] is derived from several publicly available large-scale aerial image datasets, including the DOTA-v1.5, xView, VisDrone2018-Det, Airbus Ship, and DIOR datasets. The dataset is created by dividing the original images into 800 image patches with a 200-pixel overlap. Images smaller than 800 pixels are padded with zeros. This procedure results in 11,214 training images and 2804 validation images. The dataset encompasses eight categories: airplane, bridge, oil tank, ship, swimming pool, vehicle, person, and windmill. The average size of the objects in AI-TODv1.5 is just 12.7 pixels, presenting a substantial challenge for object detection. Compared to the VisDrone2019 dataset, the AI-TODv1.5 dataset contains objects with smaller sizes and requires the detection of objects on the ground or on water surfaces from higher altitudes, making it more challenging. Therefore, to further assess the effectiveness of our proposed model, we conducted ablation experiments using this dataset.
To evaluate the generalization performance of the proposed model, we also used the PASCAL VOC dataset to assess our model’s performance and compare it with other object detection algorithms. Compared to the two small object datasets mentioned above, the Pascal VOC dataset is a conventional object detection dataset, primarily consisting of ground-level images and featuring a wider range of detection categories. Experiments on this dataset can assess the generalization performance of the proposed model. This dataset combines the PASCAL VOC 2007 and 2012 sets and is specifically designed for object detection tasks, containing 20 object categories. For our experiments, we used 16,551 images from the VOC 2007 and VOC 2012 datasets for training and 4952 images from the VOC 2007 dataset for testing.

4.2. Experimental Environment and Training Strategies

The hardware and environment parameters we used during the model training phase are shown in the following Table 1.
We applied consistent hyperparameters across the training, validation, and testing phases of our model. The chosen optimizer was SGD, configured with a momentum of 0.937 and a weight decay of 5 × 10−4. We implemented a warm-up strategy during the initial three epochs of training. For the datasets mentioned, the learning rate decay factor (gamma) was set to 0.01.
To evaluate the detection performance of our proposed improved model, we used several metrics including precision, recall, m A P 0.5 , m A P 0.5 : 0.95 , the number of model parameters, model size, and detection speed. The following terms were utilized in the formulas for some of these evaluation metrics: T P (true positive, correctly predicted as positive), F P (false positive, incorrectly predicted as positive), and F N (false negative, incorrectly predicted as negative). Intersection over Union (IoU) measures the ratio of the overlap area to the union area of the predicted bounding box and the ground truth box. Precision is defined as the ratio of true positive predictions to the total number of detected samples, and it is calculated using the following formula:
P r e c i s i o n = T P T P + F P ,
Recall is the ratio of the number of correctly predicted positive samples to the actual number of positive samples, calculated as shown in the following formula:
R e c a l l = T P T P + F N ,
Average precision ( A P ) is the area under the precision–recall curve, calculated as shown in the following formula:
A P = 0 1 P r e c i s i o n ( R e c a l l ) d ( R e c a l l ) ,
Average precision ( m A P ) is the weighted average of A P values across all sample categories, used to measure the detection performance of a model across all classes. The formula is as follows:
m A P = 1 N i = 1 N A P i ,
In Formula (24), A P i represents the A P value for the category index i , where N denotes the number of categories in the training dataset (in this context, N is 10). m A P 0.5 indicates the average precision when the IoU threshold for the detection model is set to 0.5, while m A P 0.5 : 0.95 denotes the average precision when the IoU threshold ranges from 0.5 to 0.95 (in increments of 0.5). GFLOPs measures the computational complexity during model training, representing billions of floating-point operations per second. Params refers to the number of parameters in the model, indicating the consumption of computational memory resources. FPS measures the number of images the model can detect per second, directly related to the resolution of the input images during detection. Generally, higher input image resolutions result in lower FPS under the same configuration.

4.3. Experimental Results on Visdrone2019 Dataset

To validate the effectiveness of our improved model, we performed comparative experiments with several state-of-the-art object detection algorithms using the VisDrone2019 dataset’s test set. The experiments used an input image resolution of 640 × 640, and the models were trained for 300 epochs. The comparative results of different models on the VisDrone2019 dataset are presented in Table 2.
Table 2 shows the detection results of our improved method compared to other object detection models. Our model utilizes CSPDarknet53 as the backbone network, CAAHSPAN as the neck, and Dyhead as the detection head, with an input image resolution of 640 × 640. Ultimately, it achieved an mAP50 of 41.4%, attaining the highest accuracy across each detected category. Notably, there was a significant improvement in detecting small objects, such as bicycles. Additionally, our model is an improvement based on the YOLOv8s model. As shown in the Table 2, the improved model not only surpasses the YOLOv8s model but also outperforms theYOLOv8m model while having fewer parameters and lower computational complexity (our proposed model has 10.1 million parameters and 72.7 GFLOPs, compared to YOLOv8m’s 25.9 million parameters and 79.1 GFLOPs).
We plotted the precision–recall (P-R) curves for each category, as illustrated in Figure 6. The AP for each class, listed in the table, corresponds to the area under the respective P-R curve. Compared to YOLOv8s and YOLOv8m, the proposed model achieved the highest mAP50, with the most significant improvements in the mAP for the “people” and “motor” categories. In Figure 7, we also present the confusion matrix based on our model’s results on the VisDrone2019 dataset. The prediction was conducted with an Intersection over Union (IoU) threshold of 0.5 and a confidence threshold of 0.25. The confusion matrix displays the classification outcomes for each category, with rows representing the model’s predictions and columns indicating the actual categories. The values along the diagonal indicate the proportion of correct classifications made by the model. The higher the probability on the diagonal, the better the model predicts that category. From Figure 7, it can be seen that most people were misclassified as background. This is due to the dense distribution of targets in this dataset, with most people crowded together, resulting in significant occlusion and small target sizes. When the model downsamples to extract features, it might misclassify these targets as background noise. Additionally, most bicycles, tricycles, and auto rickshaws were not correctly predicted, resulting in low AP values for these classes. Firstly, this is because the number of training samples for these classes is relatively small compared to other categories, limiting the model’s learning of these categories’ information. Secondly, these are small objects, making it challenging to extract useful information in environments with occlusion and dense targets, thus hindering the detection of these objects.

4.4. Experimental Results on the PASCAL VOC Dataset

To further assess the effectiveness of various networks, we carried out comparative experiments using the PASCAL VOC dataset. We used the training and validation images from PASCAL VOC 2012 along with the training images from VOC2007 as the training set, and validated the results on the VOC2007 test set.
Table 3 presents the comparative results of our proposed method and other popular object detectors. As can be seen from the table, our proposed model significantly outperforms two-stage object detection algorithms, such as Faster R-CNN. Additionally, among single-stage object detection algorithms, our model achieves the highest mAP50. Our model is an improved version of YOLOv8s, and it shows a slight reduction in the number of parameters, decreasing by nearly 1 million. Compared to YOLOv8s’s mAP50 of 81.5, our improved algorithm reaches an mAP50 of 82.4, an increase of nearly one point. Furthermore, in comparisons with the newly introduced YOLOv9s and YOLOv10s, our proposed algorithm still outperforms them, particularly YOLOv10s, where our method shows an improvement of nearly two points. Compared to YOLOv4 and YOLOv8m, these two models have significantly more parameters than the model we proposed. Additionally, our model’s advantages were not fully realized on conventional datasets, resulting in slightly lower performance metrics compared to these two models. Moreover although there is a decrease in FPS (frames per second), the 52.2 FPS achieved by our proposed algorithm is sufficient to meet the needs of daily life and production work.

4.5. Ablation Experiments

In this section, we conducted experiments on all modules of the model using the VisDrone2019 validation set to evaluate their impact on model performance, followed by a combination and step-by-step analysis of each module. Additionally, the loss function was evaluated. The ablation studies used the YOLOv8s algorithm as a baseline, with mAP, model size, parameter count, and computational complexity as evaluation metrics.

4.5.1. Ablation Experiments on Visdrone2019 Val Dataest

In our experiments, we made slight improvements to the positive and negative sample assignment strategy of YOLOv8. YOLOv8 uses the TaskAlignedAssigner matching strategy, which selects positive samples based on the weighted scores of classification and regression. The formula is as follows:
a l i g n m e t r i c = s α u β ,
Here, s is the predicted class score, u is the CIoU between the predicted box and the ground truth box, α and β  are weight hyperparameters. The product of these terms measures the matching degree, with the align-metric approaching 1 indicating a higher match between the predicted box and the ground truth box. For each ground truth box, we sort the align-metrics and select the top k predicted boxes as positive samples. By default, k is set to 10, but for the VisDrone2019 dataset, we modified k to 5. This adjustment was made because the targets in this dataset are densely packed, leading to significant overlap among the ground truth boxes. If k is set to 10, these 10 predicted boxes would match one ground truth box. However, due to the dense distribution of the ground truth boxes, not all 10 predicted boxes would appropriately correspond to the specific ground truth box; they might be better suited for the surrounding overlapping ground truth boxes. By reducing k to 5, the selected predicted boxes are more likely to correspond to the intended ground truth box. Experiments showed that setting k to 5 improved the model’s performance.
In the experiment, the input image resolution for the model was set to 640 × 640. The ablation test results are presented in Table 4. As shown in the table, when the topk value in TaskAlignedAssigner was adjusted from 10 to 5, all metrics saw slight improvements without altering the number of parameters or computations. Consequently, for experiments on the VisDrone2019 dataset, k was set to 5.
Additionally, we added a detection head after the 160 × 160 feature map. This layer of the feature map can better retain the spatial location information and semantic information of small targets, making it very suitable for small object detection. Despite a slight increase in computational complexity, the performance improvement is substantial, with m A P 50 increasing from 40.9% to 45.5% and m A P 50 : 95 rising from 24.4% to 27.7%. This indicates that for object detection tasks from a drone’s perspective, adding an extra detection head is both necessary and effective. Therefore, all subsequent experiments will be conducted with this additional detection head.
Subsequently, we replaced the neck part of the network with the CAA-HSPAN structure. Compared to the original PAN structure, m A P 50 increased from 45.5% to 48.7%, and m A P 50 : 95 rose from 27.7% to 29.9%, with both precision and recall improving by 2 to 3 points. This demonstrates that the CAA-HSPAN multi-scale feature fusion module, which integrates the CAA mechanism with HSPAN, can adaptively fuse information between deep and shallow feature maps and integrate contextual information, thus better extracting useful information. When fusing two feature maps, the CAA module is used to help the model focus more on the target regions in the input image, resulting in some computational overhead. Although the number of parameters decreased by 1 million, the computational complexity increased significantly, adding nearly 30 GFLOPs. However, this is still less than the computational complexity of YOLOv8m. The real-time performance of the model was also affected, with a decrease in FPS.
Next, we replaced the classification and localization losses in YOLOv8 with Slide Loss and ShapeIoU, respectively. This change improved mAP50 from 48.7% to 49.6%, with a 2-point increase in precision, without adding computational complexity or parameter count. This indicates that by changing the loss functions, the model pays more attention to hard-to-distinguish samples and effectively utilizes them during training, allowing the model to perform better in both classification and localization, thereby enhancing the performance of the detector.
Finally, we replaced all detection heads in our proposed model with Dyhead. This resulted in a slight increase in both the parameter count and computational complexity, with the parameter count increasing by 0.6 M and the computational complexity rising from 69.4 GFLOPs to 72.7 GFLOPs. However, this is still less than the computational complexity of YOLOv8m. After using Dyhead, m A P 50 increased from 49.4% to 52.2%, and m A P 50 : 95 improved from 30.3% to 31.8%, with the recall rate also increasing by nearly 2 points. This indicates that Dyhead can capture feature information at different scales, enabling the model to classify and localize targets accurately and efficiently. It improves the performance of object detection tasks from a drone’s perspective while maintaining a relatively small computational load.

4.5.2. Ablation Experiments on AI-TOD Val Dataset

To fully verify the effectiveness of each module proposed in this paper, we also conducted ablation studies on the AI-TOD dataset. The information and results for each model are shown in Table 5.
From the table, it can be seen that when we added a detection head after the P2 layer, the model’s performance improved significantly, with increases in precision, recall, and m A P . Specifically, m A P 50 increased from 38.8% to 41.7%, although FPS decreased. Furthermore, each time we incorporated one of our proposed modules into the model, all evaluation metrics showed improvements. When all the improvements were integrated into the network, the model achieved the best performance, with the highest precision and an m A P 50 increase to 45.2%. Although the final model experienced a decrease in FPS to 52.2, it remains sufficient to meet the demands of everyday production and work needs.

4.6. Visualization

To more intuitively demonstrate the effectiveness of our proposed method in everyday scenarios, we selected some representative images from the VisDrone2019 test set for detection and conducted a visualization analysis. The visualization results and comparisons are shown in Figure 8.
In Figure 8, we present some representative test results. Figure 8a shows the results obtained using the YOLOv8s model, while Figure 8b shows the results predicted by our improved model. We have enlarged some detection results in the images to facilitate a more intuitive comparison. The targets in the images in Figure 8 are very dense, with many distant targets that have indistinct features, and the problem of occluded targets is quite severe.
From the detection results in Figure 8b, it is evident that our proposed algorithm can accurately detect targets even in densely distributed scenarios. It also performs well in identifying targets that are distant, with indistinct features, by recognizing some relatively closer objects. In high-altitude images, ground-level pedestrians and vehicles are typically very small, occupying only a small portion of the pixels. Additionally, different types of targets may overlap. Our proposed model effectively detects these targets, alleviating this issue. Moreover, in shadowed or dimly lit scenes, the baseline model missed many targets, failing to detect most of them. In contrast, our model could still detect most of the targets that the baseline model missed in darker scenes. These results demonstrate that our algorithm has good robustness.
This paper proposes a novel spatial attention-based feature fusion mechanism that effectively integrates information across different scales, enhances contextual information, and maximizes the retention of small object features. Compared to conventional spatial attention mechanisms, our approach achieves these benefits while effectively reducing model parameters and computational complexity, thereby preserving model performance. Additionally, the introduction of the dynamic head allows for perception across channels, spatial dimensions, and tasks, enabling more precise prediction of object categories and positions, thereby enhancing overall model performance. Furthermore, the Slideloss classification loss and ShapeIoU regression loss used during training facilitate faster and better convergence of the model. They improve the accuracy of object classification and the precision of object localization.

5. Conclusions

From an aerial drone perspective, images often exhibit complex scenes with dense clusters of targets, including numerous small objects. This complexity typically results in lower accuracy for conventional object detection algorithms, which consequently fail to meet practical application requirements in drone scenarios. This paper introduces an improved object detection algorithm based on YOLOv8s to address these challenges effectively. Firstly, in the “neck” part of the model, we adopt HS-PAN, an enhancement of HS-FPN with an additional downsampling branch, and introduce the context-aware attention spatial attention mechanism. This mechanism captures important features within image regions, promotes feature fusion across different levels, enhances contextual information, and improves the model’s accuracy in identifying small targets. Secondly, in the detection head, we incorporate channel-aware attention, spatial-aware attention, and task-aware attention to strengthen foreground features, reduce background noise interference in the final detection results, and effectively enhance the model’s ability to extract key information from small targets. Finally, during model training, we utilize Slideloss for classification loss and ShapeIoU for regression loss. These choices enable the model to focus more on challenging samples during training and mitigate the impact of target shape and size on IoU during regression, thereby accelerating convergence. Experiments were conducted on the VisDrone2019, AI-TOD, and PASCAL VOC datasets to evaluate the effectiveness of the proposed model. The results demonstrate that our model outperforms the YOLOv8s baseline model, confirming its effectiveness in detecting small objects within drone imagery.
Typically, higher accuracy often implies more parameters and larger computational requirements, demanding substantial computing resources. Moreover, excessive parameters can adversely affect detection speed, making real-time detection impractical. Therefore, our future main research direction will focus on improving model performance and accelerating detection speed while ensuring computational or parameter constraints are maintained.

Author Contributions

Conceptualization, Z.Q. and J.S.; methodology, Z.Q.; software, Z.Q.; validation, Z.Q.; formal analysis, J.S.; investigation, Z.Q.; resources, Z.Q.; data curation, Z.Q.; writing—original draft preparation, Z.Q.; writing—review and editing, Z.Q.; visualization, Z.Q.; supervision, J.S.; project administration, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Byun, S.; Shin, I.K.; Moon, J.; Kang, J.; Choi, S.I. Road traffic monitoring from UAV images using deep learning networks. Remote Sens. 2021, 13, 4027. [Google Scholar] [CrossRef]
  2. Li, Z.; Zhang, Y.; Wu, H.; Suzuki, S.; Namiki, A.; Wang, W. Design and application of a UAV autonomous inspection system for high-voltage power transmission lines. Remote Sens. 2023, 15, 865. [Google Scholar] [CrossRef]
  3. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. A survey on deep learning-based identification of plant and crop diseases from UAV-based aerial images. Clust. Comput. 2023, 26, 1297–1317. [Google Scholar] [CrossRef] [PubMed]
  4. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  5. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  6. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision–ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  10. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  11. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  12. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  13. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  14. Deng, C.; Wang, M.; Liu, L.; Liu, Y.; Jiang, Y. Extended feature pyramid network for small object detection. IEEE Trans. Multimed. 2021, 24, 1968–1979. [Google Scholar] [CrossRef]
  15. Gong, Y.; Yu, X.; Ding, Y.; Peng, X.; Zhao, J.; Han, Z. Effective fusion factor in FPN for tiny object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1160–1168. [Google Scholar]
  16. Noh, J.; Bae, W.; Lee, W.; Seo, J.; Kim, G. Better to follow, follow to be better: Towards precise supervision of feature super-resolution for small object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 9725–9734. [Google Scholar]
  17. Yang, C.; Huang, Z.; Wang, N. Querydet: Cascaded sparse query for accelerating high-resolution small object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13668–13677. [Google Scholar]
  18. Zhou, D.; Fang, J.; Song, X.; Guan, C.; Yin, J.; Dai, Y.; Yang, R. Iou loss for 2d/3d object detection. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019; IEEE: New York, NY, USA, 2019; pp. 85–94. [Google Scholar]
  19. Wang, J.; Xu, C.; Yang, W.; Yu, L. A normalized Gaussian Wasserstein distance for tiny object detection. arXiv 2021, arXiv:2110.13389. [Google Scholar]
  20. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  21. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  24. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  25. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. Tood: Task-aligned one-stage object detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; IEEE Computer Society: Washington, DC, USA, 2021; pp. 3490–3499. [Google Scholar]
  26. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  27. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  28. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  29. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  30. Ghiasi, G.; Lin, T.Y.; Le, Q.V. Nas-fpn: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7036–7045. [Google Scholar]
  31. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  32. Jiang, Y.; Tan, Z.; Wang, J.; Sun, X.; Lin, M.; Li, H. Giraffedet: A heavy-neck paradigm for object detection. arXiv 2022, arXiv:2202.04256. [Google Scholar]
  33. Yang, G.; Lei, J.; Zhu, Z.; Cheng, S.; Feng, Z.; Liang, R. AFPN: Asymptotic feature pyramid network for object detection. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Oahu, HI, USA, 1–4 October 2023; IEEE: New York, NY, USA, 2023; pp. 2184–2189. [Google Scholar]
  34. Wang, C.; He, W.; Nie, Y.; Guo, J.; Liu, C.; Han, K.; Wang, Y. Gold-YOLO: Efficient object detector via gather-and-distribute mechanism. In Proceedings of the 37th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  35. Mnih, V.; Heess, N.; Graves, A. Recurrent models of visual attention. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 2, pp. 2204–2212. [Google Scholar]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; pp. 5998–6008. [Google Scholar]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  38. Gao, Z.; Xie, J.; Wang, Q.; Li, P. Global second-order pooling convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3024–3033. [Google Scholar]
  39. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  40. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  41. Misra, D.; Nalamada, T.; Arasanipalai, A.U.; Hou, Q. Rotate to attend: Convolutional triplet attention module. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 3139–3148. [Google Scholar]
  42. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  43. Song, G.; Liu, Y.; Wang, X. Revisiting the sibling head in object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11563–11572. [Google Scholar]
  44. Wu, Y.; Chen, Y.; Yuan, L.; Liu, Z.; Wang, L.; Li, H.; Fu, Y. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10186–10195. [Google Scholar]
  45. Ge, Z. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  46. Zhuang, J.; Qin, Z.; Yu, H.; Chen, X. Task-specific context decoupling for object detection. arXiv 2023, arXiv:2303.01047. [Google Scholar]
  47. Jiang, H.; Gu, Q. G-head: Gating head for multi-task learning in one-stage object detection. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; IEEE: New York, NY, USA, 2022; pp. 1–6. [Google Scholar]
  48. Chen, Y.; Zhang, C.; Chen, B.; Huang, Y.; Sun, Y.; Wang, C.; Fu, X.; Dai, Y.; Qin, Y.; Peng, Y.; et al. Accurate leukocyte detection based on deformable-DETR and multi-level feature fusion for aiding diagnosis of blood diseases. Comput. Biol. Med. 2024, 170, 107917. [Google Scholar] [CrossRef] [PubMed]
  49. Cai, X.; Lai, Q.; Wang, Y.; Wang, W.; Sun, Z.; Yao, Y. Poly kernel inception network for remote sensing detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 27706–27716. [Google Scholar]
  50. Dai, X.; Chen, Y.; Xiao, B.; Chen, D.; Liu, M.; Yuan, L.; Zhang, L. Dynamic head: Unifying object detection heads with attentions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7373–7382. [Google Scholar]
  51. Li, X.; Wang, W.; Wu, L.; Chen, S.; Hu, X.; Li, J.; Tang, J.; Yang, J. Generalized focal loss: Learning qualified and distributed bounding boxes for dense object detection. Adv. Neural Inf. Process. Syst. 2020, 33, 21002–21012. [Google Scholar]
  52. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans. Cybern. 2021, 52, 8574–8586. [Google Scholar] [CrossRef] [PubMed]
  53. Yu, Z.; Huang, H.; Chen, W.; Su, Y.; Liu, Y.; Wang, X. Yolo-facev2: A scale and occlusion aware face detector. arXiv 2022, arXiv:2208.02019. [Google Scholar] [CrossRef]
  54. Rezatofighi, H.; Tsoi, N.; Gwak, J.Y.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  55. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  56. Zhang, Y.F.; Ren, W.; Zhang, Z.; Zhen, J.; Liang, W.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  57. Gevorgyan, Z. SIoU loss: More powerful learning for bounding box regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
  58. Zhang, H.; Zhang, S. Shape-IoU: More Accurate Metric considering Bounding Box Shape and Scale. arXiv 2023, arXiv:2312.17663. [Google Scholar]
  59. Du, D.; Zhu, P.; Wen, L.; Bian, X.; Lin, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-DET2019: The vision meets drone object detection in image challenge results. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; IEEE Press: New York, NY, USA, 2019; pp. 213–226. [Google Scholar]
  60. Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.S. Tiny object detection in aerial images. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: New York, NY, USA, 2021; pp. 3791–3798. [Google Scholar]
  61. Zeng, S.; Yang, W.; Jiao, Y.; Geng, L.; Chen, X. SCA-YOLO: A new small object detection model for UAV images. Vis. Comput. 2024, 40, 1787–1803. [Google Scholar] [CrossRef]
  62. Li, Y.; Han, Z.; Xu, H.; Liu, L.; Li, X.; Zhang, K. YOLOv3-lite: A lightweight crack detection network for aircraft structure based on depthwise separable convolutions. Appl. Sci. 2019, 9, 3781. [Google Scholar] [CrossRef]
  63. Su, Z.; Yu, J.; Tan, H.; Wan, X.; Qi, K. MSA-YOLO: A Remote Sensing Object Detection Model Based on Multi-Scale Strip Attention. Sensors 2023, 23, 6811. [Google Scholar] [CrossRef] [PubMed]
  64. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  65. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 6569–6578. [Google Scholar]
  66. Zhang, Z.D.; Tan, M.L.; Lan, Z.C.; Liu, H.C.; Pei, L.; Yu, W.X. CDNet: A real-time and robust crosswalk detection network on Jetson nano based on YOLOv5. Neural Comput. Appl. 2022, 34, 10719–10730. [Google Scholar] [CrossRef]
Figure 1. The architecture of our proposed model features the CSPDarknet53 network as the backbone for feature extraction. In the neck section, we incorporate a multi-scale feature fusion structure enhanced by CAA. In its head part, dynamic head is introduced.
Figure 1. The architecture of our proposed model features the CSPDarknet53 network as the backbone for feature extraction. In the neck section, we incorporate a multi-scale feature fusion structure enhanced by CAA. In its head part, dynamic head is introduced.
Sensors 25 00589 g001
Figure 2. The structure of CAA-HSPAN. It consists of a feature fusion module and a feature enhancement module, where context anchor attention is a type of spatial attention. But in the USFF and DSFF modules, only attention weights are generated and not multiplied element by element with the input feature map.
Figure 2. The structure of CAA-HSPAN. It consists of a feature fusion module and a feature enhancement module, where context anchor attention is a type of spatial attention. But in the USFF and DSFF modules, only attention weights are generated and not multiplied element by element with the input feature map.
Sensors 25 00589 g002
Figure 3. Two feature fusion models. (a) The structure of USFF; (b) The structure of DSFF.
Figure 3. Two feature fusion models. (a) The structure of USFF; (b) The structure of DSFF.
Sensors 25 00589 g003
Figure 4. The overall structure of the dyhead (assuming 3 detection heads) consists of three parts: scale-aware attention, spatial-aware attention, and task-aware attention. The DConv filled in blue is a 3 × 3 deformable convolution with stride = 1, while the DConv filled in green is a 3 × 3 deformable convolution with stride = 2. Sigmoid refers to the hard sigmoid function.
Figure 4. The overall structure of the dyhead (assuming 3 detection heads) consists of three parts: scale-aware attention, spatial-aware attention, and task-aware attention. The DConv filled in blue is a 3 × 3 deformable convolution with stride = 1, while the DConv filled in green is a 3 × 3 deformable convolution with stride = 2. Sigmoid refers to the hard sigmoid function.
Sensors 25 00589 g004
Figure 5. Analysis of Bounding Box Regression Characteristics. (a) The two bounding boxes only have coincidence relationship but no inclusion relationship; (b) The two bounding boxes not only have coincidence relationship but also have inclusion relationship.
Figure 5. Analysis of Bounding Box Regression Characteristics. (a) The two bounding boxes only have coincidence relationship but no inclusion relationship; (b) The two bounding boxes not only have coincidence relationship but also have inclusion relationship.
Sensors 25 00589 g005
Figure 6. The P-R curve of our proposed model and baseline models on the Visdrone2019 dataset. (a) The P-R curve of YOLOv8s; (b) the P-R curve of YOLOv8m; (c) the P-R curve of our proposed model.
Figure 6. The P-R curve of our proposed model and baseline models on the Visdrone2019 dataset. (a) The P-R curve of YOLOv8s; (b) the P-R curve of YOLOv8m; (c) the P-R curve of our proposed model.
Sensors 25 00589 g006
Figure 7. The P-R curve based on our model’s results on the VisDrone2019 test dataset, along with the confusion matrix for the same dataset are presented with an IoU threshold of 0.5 and a confidence thresh old of 0.25.
Figure 7. The P-R curve based on our model’s results on the VisDrone2019 test dataset, along with the confusion matrix for the same dataset are presented with an IoU threshold of 0.5 and a confidence thresh old of 0.25.
Sensors 25 00589 g007
Figure 8. Visualization of results on images from the VisDrone2019 test dataset is presented. In these images, different categories are marked with boxes of different colors.
Figure 8. Visualization of results on images from the VisDrone2019 test dataset is presented. In these images, different categories are marked with boxes of different colors.
Sensors 25 00589 g008
Table 1. Training environment and hardware platform parameters table.
Table 1. Training environment and hardware platform parameters table.
ParametersConfiguration
CPUI5-13600KF
GPUNVIDIA GeForce RTX 3060ti
GPU memory size8 G
Operating systemWin 11
Deep learning architecturePytorch2.0.0 + Cuda11.8
Table 2. Comparison of various methods on the Visdrone2019 test dataset.
Table 2. Comparison of various methods on the Visdrone2019 test dataset.
ModelObject CategorymAP50 (%)
PEDPERBCCarVanTruckTRIATRIBusMO
Fast R-CNN [6]21.415.66.751.729.51913.17.731.420.721.7
Faster R-CNN [7]20.914.87.35129.719.5148.830.521.221.8
CascadeR-CNN [61]22.214.87.654.631.521.614.88.634.921.423.2
RetinaNet [61]137.91.445.519.911.56.34.217.811.813.9
CenterNet [61]22.620.614.659.72421.320.117.437.923.726.2
DMNet [61]28.520.415.956.837.930.122.61447.129.230.3
HRDet+ [61]28.614.511.749.437.135.228.821.943.323.528
ACM-OD [59]30.815.510.352.738.933.226.921.941.424.929.7
CDNet [59]35.619.213.855.842.138.23325.449.529.334.2
HR-Cascade++ [61]32.617.311.154.742.435.332.724.146.528.232.5
MSC-CenterNet [61]33.715.212.155.240.534.129.221.642.227.531.1
YOLOv3-LITE [62]34.523.47.970.831.321.915.36.240.932.728.5
MSA-YOLO [63]33.417.311.276.841.541.414.818.460.93134.7
YOLOv4 [21]24.812.68.664.322.422.711.47.644.321.730.7
YOLOv5s24.215.97.668.730.530.511.114.852.521.827.7
YOLOv8s27.214.69.271.637.139.116.918.556.82932
YOLOv8m30.617.311.674.440.244.520.119.959.333.335.1
YOLOv9s [25]27.515.39.271.437.639.619.120.458.23032.8
YOLOv10s [26]26.416.37.770.735.537.216.51855.628.331.2
YOLO11s27.613.98.5571.737.339.718.41957.429.332.2
Ours (test)39.127.519.479.746.644.826.425.862.741.641.4
Table 3. Comparison of various methods on the VOC2007 test dataset.
Table 3. Comparison of various methods on the VOC2007 test dataset.
ModelBackboneSizeParametersmAP50FPS
Fast R-CNN [6]ResNet600 × 100044.5 M70.01.9
Faster R-CNN [7]ResNet600 × 100040.1 M76.42.4
CascadeR-CNN [64]ResNet600 × 100076.8 M79.62.2
CenterNet [65]ResNet512 × 51212.7 M77.17
RetinaNet [10]ResNet600 × 60033.8 M81.55
YOLO [61]GoogleNet448 × 44862 M63.445
YOLOv2 [61]DarkNet352 × 35267 M76.867
YOLOv3 [20]DarkNet416 × 41661.5 M78.336.9
YOLOv3-LITE [62]DarkNet416 × 41631 M74.071.9
YOLOv4 [21]CSPDarkNet640 × 64039.6 M83.260
MSA-YOLO [63]CSPDarkNet640 × 64012.180.4174.9
CDNet [66]CSPDarkNet640 × 64021.5 M80.3165.3
YOLOv5sCSPDarkNet640 × 6409.13 M79.9210.5
YOLOv8sCSPDarkNet640 × 64011.2 M81.5200.7
YOLOv8mCSPDarkNet640 × 64025.9 M84.580.7
YOLOv9s [25]GELAN640 × 6407.3 M82.2101.1
YOLOv10s [26]CSPDarkNet640 × 6408.07 M80.8172.8
YOLO11sCSPDarkNet640 × 6409.4 M81.6212.2
OursCSPDarkNet640 × 64010.1 M82.452.2
All the algorithms were trained using the combined training set of VOC2007 and VOC2012 and evaluated on the VOC2007 test dataset. Ultimately, the detection accuracy and processing speed of these detectors were compared.
Table 4. The effect of combining different modules in YOLOv8s on the VisDrone2019 test dataset (p2 refers to the additional prediction head).
Table 4. The effect of combining different modules in YOLOv8s on the VisDrone2019 test dataset (p2 refers to the additional prediction head).
MethodmAP50 (%)mAP50:95 (%)P (%)R (%)GFLOPsParams (M)FPS
YOLOv8s39.323.550.938.228.711.1200
YOLOv8s-topk = 540.924.452.139.928.711.1200
YOLOv8s-p2-topk = 545.527.755.943.73710.6146
YOLOv8s-p2-topk = 5-CAAHSPAN48.729.957.246.969.49.5375
YOLOv8s-p2-topk = 5-CAAHSPAN-
Slideloss-ShapeIoU
49.430.359.647.169.49.5375
YOLOv8s-p2-topk = 5-CAAHSPAN-
Slideloss-ShapeIoU-DyHead
52.231.860.949.572.710.152
Table 5. The effect of combining different modules in YOLOv8s on the AI-TODv1.5 validation dataset (p2 refers to the additional prediction head).
Table 5. The effect of combining different modules in YOLOv8s on the AI-TODv1.5 validation dataset (p2 refers to the additional prediction head).
MethodmAP50 (%)mAP50:95 (%)P (%)R (%)GFLOPsParams (M)FPS
YOLOv8s38.816.258.137.728.711.1200
YOLOv8s-p2-topk = 541.718.360.940.43710.6146
YOLOv8s-p2-topk = 5-CAAHSPAN43.319.362.940.769.49.5375
YOLOv8s-p2-topk = 5-CAAHSPAN-Slideloss-ShapeIoU44.119.663.141.969.49.5375
YOLOv8s-p2-topk = 5-CAAHSPAN-Slideloss-ShapeIoU-DyHead45.220.163.943.772.710.152
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Quan, Z.; Sun, J. A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism. Sensors 2025, 25, 589. https://doi.org/10.3390/s25020589

AMA Style

Quan Z, Sun J. A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism. Sensors. 2025; 25(2):589. https://doi.org/10.3390/s25020589

Chicago/Turabian Style

Quan, Zhe, and Jun Sun. 2025. "A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism" Sensors 25, no. 2: 589. https://doi.org/10.3390/s25020589

APA Style

Quan, Z., & Sun, J. (2025). A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism. Sensors, 25(2), 589. https://doi.org/10.3390/s25020589

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop