Next Article in Journal
Improved Order Tracking in Vibration Data Utilizing Variable Frequency Drive Signature
Previous Article in Journal
Relationship Between Cognitive Abilities and Lower-Limb Movements: Can Analyzing Gait Parameters and Movements Help Detect Dementia? A Systematic Review
Previous Article in Special Issue
Efficient Multi-Task Training with Adaptive Feature Alignment for Universal Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Small Target Detection Algorithm Based on Improved Dense Nested U-Net Network

1
Kunming Institute of Physics, Kunming 650223, China
2
National Pilot School of Software, Engineering Research Center of Cyberspace, Yunnan University, Kunming 650091, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 814; https://doi.org/10.3390/s25030814
Submission received: 21 November 2024 / Revised: 24 January 2025 / Accepted: 28 January 2025 / Published: 29 January 2025
(This article belongs to the Special Issue Computer Vision Sensing and Pattern Recognition)

Abstract

:
Infrared weak and small target detection technology has attracted much attention in recent years and is crucial in the application fields of early warning, monitoring, medical diagnostics, and anti-UAV detection.With the advancement of deep learning, CNN-based methods have achieved promising results in general-purpose target detection due to their powerful modeling capabilities; however, CNN-based methods cannot be directly applied to infrared small targets due to the disappearance of deep targets caused by multiple downsampling operations. Aiming at these problems, we proposed an improved dense nesting and attention infrared small target detection method based on U-Net called IDNA-UNet. A dense nested interaction module (DNIM) is designed as a feature extraction module to achieve level-by-level feature fusion and retain small targets’ features and detailed positioning information. To integrate low-level features into deeper high-level features, we designed a bottom-up feature pyramid fusion module, which can further retain high-level semantic information and target detail information. In addition, a more suitable scale and position sensitive (SLS) loss is applied to each prediction scale to help the detector locate the target more accurately and distinguish different scales of the target. With our IDNA-UNet, the contextual information of small targets can be well incorporated and fully exploited by repetitive fusion and enhancement. Compared with existing methods, IDNA-UNet has achieved significant advantages in the intersection over union (IoU), detection probability ( P d ), and false alarm rate ( F a ) of infrared small target detection.

1. Introduction

Accurate target identification is a crucial technological foundation for effective Infrared Search and Track (IRST) operations. It enables the early detection of targets and provides precise warnings, which are critical requirements for air defense systems. However, the targets in infrared images are often characterized by their small size or significant distance from the detector, posing challenges to reliable detection. In images captured by electro-optical (EO) systems, such targets typically occupy only a few or even a single effective pixel, or in some cases, they may appear as sub-pixel features. According to the SPIE international standards, targets that occupy less than 0.12% of the image size are classified as small targets. These targets usually appear as dots or spots with minimal texture or discernible shape features [1]. Moreover, due to atmospheric scattering and absorption, infrared radiation further weakens with distance, making the target’s brightness exceedingly faint. Recent academic studies have shown an increasing focus on infrared faint and sub-pixel target discrimination, highlighting its growing importance in advancing detection methodologies.
In target detection, researchers have proposed several traditional methods in early studies in the field, including filter-based algorithms [2,3], local contrast-based methods [4], and image data structure-based methods [5,6]. However, these conventional approaches predominantly rely on target intensity profiles and local neighborhood characteristics for detection, with their performance constrained by the inherent limitations of manual feature engineering.
The current detection landscape has transitioned from manual feature design to convolutional neural networks (CNNs) as the principal deep learning framework [6,7,8,9]. As the predominant deep learning framework, CNNs have demonstrated remarkable capabilities in extracting robust feature representations, particularly in challenging weak target detection scenarios. Leveraging CNN’s ability to process large-scale data and its powerful model-fitting capabilities, these methods can effectively capture discriminative and salient features of small targets. Typically, these methods focus on the target itself, utilizing multi-layer networks to extract feature information of the target for detection. Enhancing detection accuracy is often achieved by increasing the depth of the neural network. However, it is worth noting that since small targets in images occupy only a few effective pixels, and given the significant differences in shape and texture between visible and infrared targets, infrared small targets lack texture and structural information compared to visible light. Therefore, simply deepening the network does not necessarily improve the detection performance for small targets [10]. On the contrary, due to the scarcity of features in infrared small targets and the limited scale of available datasets, the features of small targets face the risk of being overwhelmed by background features during the detection process after multiple down-sampling operations.
To solve these problems, we proposed an improved dense nesting and attention infrared small target detection algorithm based on U-Net (IDNA-UNet). Our framework utilizes a densely connected interaction network as the core computational backbone to facilitate multi-level feature abstraction from infrared image inputs. In order to avoid using too many pooling operations, we only use a five-layer network structure to expand the receptive field while extracting features. Through progressive feature fusion and interaction, the intermediate nodes can receive the features of themselves and the adjacent two layers of nodes, thereby fully utilizing and integrating the contextual information with small target information. The extracted features include deep features with high-level semantic information and shallow features with rich image contour and grayscale information. The weights of feature channels in traditional convolutional neural networks are the same. However, during the convolution transformation of feature information, the feature areas that each channel should focus on in the infrared image are different, which requires focusing on channels with small target information. Two architectural innovations are implemented: a channel attention module to enhance target-relevant feature channels, and a multi-scale fusion mechanism employing a bottom-up pyramid architecture for optimized feature integration across different levels. The advantage of this is that it can protect the representation of deep, small targets while extracting high-level features. Since the widely used intersection-over-union (IoU) loss lacks sensitivity to target size and position, it is difficult to distinguish targets of different scales and positions for target detection tasks, which greatly limits the performance of the detector. A novel and effective scale- and position-sensitive loss function is used to improve the existing problems. In summary, the contributions of this paper are summarized as follows:
  • Based on the encoder and decoder architecture of U-Net, we proposed an improved densely nested and attention infrared dim small target detection network architecture. By introducing a densely nested interactive network as the backbone network, we integrated a channel attention mechanism to enhance feature representation by adaptively weighting different channels during infrared dim small target feature extraction.
  • Through the architectural design of multi-scale heads and bottom-up feature pyramid feature fusion, we can deepen the focus on target features layer by layer. This structural innovation substantially augments the network’s capacity for detecting subtle features and effectively distinguishing infrared targets against cluttered backgrounds.
  • A novel and effective scale-sensitive and position-sensitive loss function is adopted, which helps the detector to distinguish objects of different scales and positions, thereby further improving the detection performance.

2. Related Work

Top-hat [11] is a classic morphological filtering method that uses various filters to estimate the background of a scene, allowing the target to be distinguished from complex backgrounds. These algorithms exhibit significant dependence on target–background size differentials, consequently constraining their practical application scope. Gao C et al. [5] innovatively expanded the conventional infrared image framework through the development of an Infrared Patch-Image (IPI) model, achieved via localized block construction methodology. By vectorizing and concatenating overlapping image patches sampled using a sliding window, they created a matrix that satisfies the assumptions of low-rank background and sparse targets. PSTNN [12] employs a non-convex method based on the tensor nuclear norm, while RIPT [6] focuses on reweighted infrared patch tensors. Subsequently, other scholars proposed methods based on visual mechanisms. For instance, the research team led by Chen et al. [4] proposed the Local Contrast Measurement (LCM) framework, employing localized contrast analysis to differentiate targets from background elements. The limitations of these algorithms lie in their sensitivity to scene changes, as they struggle with parameter and algorithm adjustments, making them unsuitable for multi-scene applications. Moreover, traditional algorithms face an inherent trade-off between improving sensitivity and reducing false alarm rates, which is difficult to reconcile.
With the rise of deep learning, methods leveraging the massive data and powerful fitting capabilities of CNNs have demonstrated better performance compared to traditional methods. However, general object detection algorithms perform poorly in detecting infrared small and weak targets. The challenges lie in the conflict between the target’s size and the deep feature extraction network, as well as the poor generalization ability of the algorithms. To address this problem, researchers have proposed some improvement and optimization strategies based on the existing target detection methods.
To enable the network to learn infrared small target features across different scales as effectively as possible, Dai Y [13] and others proposed an Asymmetric Context Modulation (ACM) method to enhance features. Zhang T [14] further extracted contextual information of the target in deeper layers of ACM. This architectural module is specifically designed to generate multi-scale feature representations through the implementation of dilated convolutions and adaptive global average pooling operations on high-level feature maps. Generative adversarial networks (GANs) achieve a Nash equilibrium through adversarial games between the generator and the discriminator. In the task of infrared small target detection, Wang H et al. [8] pioneered a novel approach for small target detection through the development of an enhanced generative adversarial network architecture. The researchers conceptualized infrared small target segmentation as an optimization challenge, focusing on achieving equilibrium between missed detection (Md) and false alarm (Fa) rates through balanced parameter optimization. Zhao et al. [15] trained the generator using simulated targets and then utilized the trained generator for target detection. Zhou et al. [16] established the PixelGame framework, reformulating multi-scale infrared target detection as a competitive optimization problem with adversarial balancing of detection errors. Concurrently, Li et al. [9] and collaborators introduced DNANet, implementing dense nested attention mechanisms for improved detection performance. The densely nested interactive architecture of the backbone network facilitates comprehensive contextual information integration through iterative fusion and enhancement processes, optimizing small target feature utilization. He et al. [17] designed the Subpixel Sampling Cuneate Network (SPSCNet) to achieve iterative fusion between high-level and low-level features. UNet-based methods have demonstrated remarkable progress across various segmentation tasks. Wu et al. [10] proposed UIU-Net, which integrates a compact U-Net [18] within a larger U-Net backbone to simultaneously capture global and local contrast information. Similarly, Zhang et al. [19] introduced Sharp U-Net in 2022, leveraging depthwise separable convolutions to optimize lightweight network design, achieving substantial reductions in both parameter count and computational complexity while maintaining competitive segmentation accuracy. While these architectures have shown exceptional performance in biomedical image segmentation, particularly in terms of inference speed and resource efficiency, their applicability to infrared small target detection remains limited due to unique challenges such as target sparsity and complex feature representation.
With the active exploration of various researchers, deep network frameworks specifically designed for infrared small target detection have been continuously proposed, leading to progressively improved detection performance. However, existing methods still face several challenges, such as limited feature representation capabilities, insufficient robustness against complex backgrounds, and constraints imposed by limited network depth that hinder further performance enhancements. These issues restrict the effectiveness and applicability of infrared small target detection in real-world scenarios. To address these limitations, this paper aims to enhance the feature representation capability of each layer within a limited network depth to further improve detection robustness, which is crucial for the performance of infrared small target detection.
Infrared images have a more extreme contradiction between resolution and semantics. On the one hand, due to the long imaging distance, the proportion of small infrared targets in the image usually only occupies a few pixels, and its features are more likely to be covered by the surrounding background features as the number of network layers increases. On the other hand, in many real scenes, small infrared targets are often buried in complex backgrounds with relatively low signal clutter, lacking corresponding texture and shape features. However, deep convolutional neural networks usually use the method of gradually reducing the size of the feature maps to increase the receptive field so as to learn more semantic information. The resolution and semantic level of their feature maps are often a contradiction. Therefore, within constrained network depth configurations, maximizing the representational capacity of features at each network layer becomes critical for enhancing infrared weak and small target detection performance. The above methods still have some shortcomings. Directly using these methods may cause the loss of deep, small targets. Consequently, considerable scope for algorithmic refinement persists in deep learning-based solutions for infrared weak target detection applications.

3. Methodology

This section introduces IDNA-UNet, including its overall structure, core module structure, and the design of the SLS loss function.

3.1. The Overall Structure

IDNA-UNet consists of two parts: feature extraction and feature fusion. The feature extraction module is inspired by the nested structure in the field of image segmentation and the attention mechanism in the field of object detection [9]. The feature fusion module designs a bottom-up feature pyramid fusion module (BFPFM) to integrate low-level features into deeper high-level features.
In the multiple down-sampling stages of infrared small target detection, high-dimensional feature representations, despite containing target semantic data, frequently fail to effectively maintain and accentuate the critical attributes of low-visibility small targets. On the other hand, low-dimensional features can effectively retain target characteristics but lack high-level semantics. This represents a significant bottleneck in infrared small target detection. Therefore, the existence of this inherent contradiction should be carefully considered at the outset of network design, and targeted improvements to the network architecture should be made. As shown in Figure 1, IDNA-UNet is based on the encoder and decoder architecture of U-Net. Our backbone network structure retains the dense nesting structure (DNIM) of DNANet [9], which adds multiple skip connection paths based on U-Net to achieve hierarchical feature fusion, which effectively utilizes the global contextual information of small targets. Compared to DNANet, we introduce the SENet module at the end of each convolutional neural network. In traditional CNNs, pooling layers tend to weaken the target features within each sample, while SENet reallocates the weights based on the importance of each channel of features, effectively enhancing the weights of the channel information. The feature maps x i of different scales are output from the backbone network and x i is input into a feature fusion module with multi-scale heads. Multi-scale feature representations are normalized to uniform dimensionality through systematic upsampling operations. A scale and position sensitive (SLS) loss is applied to ensure that each scale head can generate predictions of the corresponding scale. The predictions of different scales are connected for the final prediction.

3.2. Feature Extraction Network

In order to ensure high-resolution features and fine positioning information, the feature extraction module designs a dense nested interactive module (DNIM) based on the U-Net framework. On the basis of the U-Net architecture, multiple jump connection paths are added between the encoder and the decoder. The sub-encoder receives the detailed positioning information of the encoder through the jump connection path and directly applies the shallow position information to the deep feature map, thereby maintaining the detailed information of the target. In addition, the output of the shallower sub-encoder node will be connected to the deeper sub-encoder node of the same size to utilize the detailed features further.
X i , j represents the current node output, where i represents the index of the downsampling layer and j is the index of the dense block convolution layer along the normal skip connection path. When j = 0 , each node only receives features from dense normal skip connections. The feature map represented by X i , j is calculated as
X i , j = P max F X i 1 , j
where F ( · ) represents multiple cascaded convolutional layers of the same convolutional block and P m a x ( · ) represents maximum pooling with a stride of 2. When j > 0 , each node receives output from three directions, including dense ordinary jump connections and nested interactive jump connections, and the feature map stack represented by X i , j is calculated as
X i , j = F X i , k k = 0 j 1 , P max F X i + 1 , j 1 , u F X i 1 , j
where u ( · ) represents the upsampling layer, [ · , · ] represents a cascade layer, and k = 0 indicates the X i 1 , 0 convolutional block.
In the dense nested feature extraction module network, each X i , j represents a convolutional neural network. The feature extraction module (FEM) architecture implements a novel Residual SENet Block (RSB) node design, supplanting conventional ResBlocks. Each RSB node integrates ResNet18’s residual unit architecture with SENet’s channel attention mechanism, forming a hybrid structure that combines residual learning with adaptive channel-wise feature recalibration. Our design philosophy emphasizes two core objectives: (1) maintaining model lightweightness through parameter-efficient architecture design, and (2) achieving substantial global feature enhancement with minimal computational overhead. To preserve architectural consistency, each RSB node incorporates skip connections immediately following the initial convolutional layer, as illustrated in Figure 2.
Small targets can be easily submerged in the high-level features of the network, but using only low-level features, the semantic information cannot be grasped well, which can easily lead to missed detection and false alarms. The SENet framework employs residual-based learning to establish feature channel importance metrics, dynamically allocating weight values to enhance network focus on discriminative feature maps while suppressing less informative ones. Invalid or ineffective feature maps have small weights. SENet mainly includes two operations: squeeze and excitation.
In Figure 3, squeeze compresses the W × H × C feature map containing global information into a 1 × 1 × C feature vector through global average pooling. This feature vector has a global receptive field and can be expressed as described below.
The first step of SENet is to extract the feature map of the image using the convolutional operation which can be represented as
F t r : X U , X H × W × C , U H × W × C
F t r : u c = v c x = s = 1 c v c s x s
F t r represents a transformation operation that maps the input raw image X to the feature map U, where H and W denote the height and width of the feature map, and C represents the number of channels. v c s represents the parameters of the c convolutional kernel, * denotes convolution, x s represents the s input, V = [ v 1 , v 2 , , v c ] is the learned combination of convolutional kernels, and U = [ u 1 , u 2 , , u c ] is the output.
In the squeeze stage, the size of the input feature U is H × W × C . Through a global average pooling operation, the features at each spatial location are aggregated into a single value, compressing each H × W region into a single value, resulting in a 1 × 1 × C feature vector. It can be expressed as
z = F s q u c = 1 H W i = 1 H j = 1 W u c ( i , j )
z represents the weights generated by the squeeze operation, and F s q ( · ) represents the function of the squeeze operation.
The squeeze operation captures the global features of the feature channels, while the excitation operation captures the correlations between the feature channels, and is implemented using two fully connected layers. The initial fully connected layer implements dimensionality reduction, optimizing parameter efficiency while improving model generalization. Subsequently, the secondary fully connected layer reconstructs the original feature dimensions. The calculation process can be expressed as
s c = F e x ( z , W ) = σ W 2 δ W 1 z
where F e x ( · ) represents the function of the squeeze operation, δ denotes the ReLU function, and σ represents the sigmoid function. W 1 and W 2 represent the fully connected (FC) operations, and S c R 1 × 1 × C denotes the generated attention weights.
X ¯ c = F scale u c , s c = s c u c
where ⊗ represents element multiplication, and X ¯ c is the output feature map. F s c a l e ( u c , s c ) refers to the per-channel multiplication between the weight vector s c and the original feature map u R H × W .

3.3. The Feature Fusion Module of Multi-Scale Heads

After the feature extraction module, we designed a feature fusion module to aggregate multi-scale features. Let x i R H i × W i × C i be the feature map at the i-th scale of the DNIM decoder, where H i × W i is the spatial size, C i is the number of channels, and i 0 , 1 , 2 , 3 , 4 .
The i-th prediction p i R H i × W i × 1 is obtained through the corresponding prediction head, which is implemented by a convolutional layer followed by a sigmoid activation function
p i = S i g m o i d ( C o v ( x i ) )
The prediction results p i from different scale heads are upsampled and input into the BFPFM module for further fusion.

3.4. Bottom-Up Feature Pyramid Fusion Module (BFPFM)

In a common convolutional neural network, as the number of network layers increases, the network’s ability to extract semantic features of the input becomes stronger. However, as the number of downsampling layers increases, the network is prone to lose spatial information on small-scale targets. The feature fusion architecture employs a bottom-up pyramid framework (BFPFM) to effectively encode and preserve small target characteristics by progressively integrating low-level spatial details into high-level semantic features.
BFPFM is shown in Figure 4. The local channel attention mechanism L locally aggregates the channel feature context of each spatial position, and the formula is
L ( X ) = σ B P W Conv 2 δ B PWConv 1 ( X )
where P W C o n v , σ , δ , and B represent a 1 × 1 convolution, the sigmoid function, ReLU, and BN, respectively. The kernel sizes of P W C o n v 1 and P W C o n v 2 are ( C / 4 ) × C × 1 × 1 and C × ( C / 4 ) × 1 × 1 , respectively. The attention weight map L ( X ) R C × H × W has the same shape as the feature map, so it can span channels spatially and highlight subtle details in an element-wise manner. Therefore, the bottom-up feature pyramid fusion module can dynamically perceive the subtle details of weak infrared targets.
Given X as the low-level feature and Y as the high-level feature, the cross-layer fusion feature Z R C × H × W can be obtained through the bottom-up feature pyramid fusion module. The calculation process can be expressed as
Z = X + L X Y
where ⊗ represents element multiplication.

3.5. SLS Loss Function

YOLOv7 uses the bounding box positioning loss function CIoU (Complete-IoU) [20]. CIoU’s design methodology integrates three geometric considerations: the overlapping area between the predicted box and the ground truth box, the center point distance, and the aspect ratio. Soft IoU [21] is also commonly used in semantic segmentation tasks. It is based on the traditional IoU indicator and also performs smoothing to make the prediction result smoother, which affects the calculation of the intersection-over-union ratio. However, for specific application scenarios where small targets usually only contain a few pixels, the evaluation indicators commonly used in these traditional target detection algorithms, such as Iou, SIou, DIou, and Soft IoU, lack sensitivity to the size and position of the target. Targets of different scales and positions may have the same IoU loss or Dice loss, which restricts the detection performance of small targets.
To resolve these limitations, we employ the scale and location sensitive (SLS) loss function, specifically designed to address the scale and positional insensitivity inherent in conventional loss formulations. It consists of scale-sensitive loss and position-sensitive loss and can be expressed as
L S L S = L S + L L
where L S and L L represent scale-sensitive loss and position-sensitive loss; the scale-sensitive loss is implemented by providing weights through IoU loss. Let A p and A g t be the set of predicted and ground truth pixels of the target, respectively.
L s = 1 ω A p A g t A p A g t
s . t . ω = min ( A p , A g t ) + V a r ( A p , A g t ) max ( A p , A g t ) + V a r ( A p , A g t )
where V a r ( · ) is a function that obtains the variance of the provided scalar. It can be seen that the smaller ω is, the larger the gap between | A p | and | A g t | is (assuming that the IoU between | A p | and | A g t | is fixed), that is, the predicted scale and the true scale are very different. At this time, the detector should pay more attention to the target with the larger loss.
The position-sensitive loss is calculated based on the predicted center point and the ground truth center point. Given the sets of predicted pixels A p and ground truth pixels A g t , the center points corresponding to A p and A g t are obtained by averaging the coordinates of all pixels, denoted as c p = ( x p , y p ) and c g t = ( x g t , y g t ) , respectively. Then, the coordinates of these two center points are converted to polar coordinates. For example, for c p , the corresponding distance d p and angle θ p in polar coordinates are
d p = x p 2 + y p 2
θ = arctan ( y p x p )
L L = ( 1 min ( d p , d g t ) max ( d p , d g t ) ) + 4 π 2 ( θ p θ g t ) 2
where d g t and θ g t are the distance and angle of c g t in the polar coordinate system, respectively.

4. Experiment

4.1. Experimental Data and Experimental Settings

Dataset. The datasets used for model evaluation are the publicly available infrared small target datasets, NUDT-SIRST and IRSTD-1k, for training and prediction. We set the training and testing ratio to 3 for the NUDT-SIRST dataset (i.e., 995 images for training and 332 images for testing, where total number of targets ( T a l l ) is 451 and total number of target pixels ( P a l l ) is 14,101). The image size of the NUDT-SIRST dataset is 256 × 256. For the IRSTD-1k dataset, the training and testing ratio is set to 4 (i.e., 800 images for training and 201 images for testing, where total number of targets ( T a l l ) is 297 and total number of target pixels ( P a l l ) is 14,534). Before training, all images are normalized. For the IRSTD-1k dataset, the image size is resized to 256 × 256 resolution before being input into the network.
Hyperparameter Configuration. In this paper, The U-net paradigm with ResNets [22] was chosen as our segmentation backbone, which generates pixel-level segmentation maps. The number of down-sampling layers i is set to 4 due to the fact that shallower U-Net variants are more effective in retaining detailed information in tasks that require high spatial accuracy [23,24,25]. The Adagrad [26] method is used to optimize the network. The network model was trained for a total of 399 epochs, with a batch size of 16 and a learning rate set to 0.05. The experiments in this paper were all conducted on the Ubuntu 18.04 operating system, equipped with an Intel Xeon E5-2678 CPU and an NVIDIA Geforce TITAN RTX 24 GB GPU. The computational environment utilized Python 3.8, and the deep learning models were developed using the PyTorch 5.1.2 framework. Moreover, the latencies of all models are tested on NVIDIA Orin Nano with TensorRT FP16 machines, which have a computational power of 40TOPS. The hyperparameter settings for the model-driven approach are shown in Table 1.
Table 1. Detailed hyper-parameter settings of model-driven methods for comparison.
Table 1. Detailed hyper-parameter settings of model-driven methods for comparison.
MethodsHyperparameter Settings
Top-hat [11]Structure shape: square, local window size: 5 × 5
IPI [5]Patch size: 50 × 50, stride: 10, λ = L / min ( I , J , P ) , L = 3 , threshold factor: ε = 10 7 , k = 10

4.2. Performance Evaluation

Object detection usually uses metrics such as IoU, precision, and recall. These pixel-level evaluation indicators focus more on target shape evaluation. However, since small infrared targets usually lack shape and texture, P d and F a must be combined to evaluate positioning performance. Therefore, we use the intersection over union (IoU) to evaluate the ability to describe the shape of a small target and the detection probability P d and false alarm rate F a to evaluate the ability to locate small targets.
IoU is a pixel-level evaluation that assesses the method’s ability to describe a profile. IoU is calculated by the ratio of the intersection and union areas between the prediction and the label and is defined as follows
I o U = A i n t e r A u n i o n
where A i n t e r and A u n i o n represent the interaction area and the union area.
Detection probability ( P d ) is an object-level evaluation metric. It measures the ratio of the correctly predicted target numbers T c o r r e c t over all target numbers T a l l . P d is defined as follows
P d = T c o r r e c t T a l l
The false alarm rate ( F a ) is another target-level evaluation metric. It is used to measure the ratio of falsely predicted pixels P f a l s e over all target pixels P a l l . F a is defined as follows
F a = P f a l s e P a l l

4.3. Performance Comparison with Previous Methods

To further verify the effectiveness of our proposed IDNA-UNet on the infrared dim target detection dataset, we evaluate it with different types of methods from both quantitative and qualitative perspectives, including traditional methods and deep learning-based methods. Traditional methods are Top-hat and IPI; deep learning-based methods are MDvsFA, ALCNet [7], DNANet, and MSHNet. For a fair comparison, all evaluated deep learning-based models are retrained using the official code to converge on the NUDT-SIRST [27] and IRSTD-1k [28] datasets.
Table 2 lists the evaluation indicators of IoU (%), P d (%), and F a ( 10 6 ) of the six different methods on the NUDT-SIRST and IRSTD-1k datasets. These results demonstrate that the IDNA-UNet model has an excellent effect on infrared dim small target detection. The IDNA-UNet target detection method proposed in this paper has a higher IoU and lower F a than the mainstream method. Compared with the better-performing DNANet, F a can be reduced by 11.11 × 106. Our method’s detection rate ( P d ) has also achieved a higher accuracy. The ROC curve of the Figure 5 also highlights that our model quickly reaches the top-left corner and demonstrates competitive performance in terms of AUC.
Additionally, the table provides detailed information regarding the computational complexity and parameter counts for five state-of-the-art infrared small target detection networks, offering insights into their efficiency and resource requirements. From the perspective of lightweight design, MD&FA, ALCNet, and MSHNet exhibit lower computational complexity, while in terms of segmentation performance, DNANet, MSHNet, and the proposed IDNA-UNet demonstrate higher segmentation accuracy. Considering engineering applications, deployment tests on the NVIDIA Orin Nano embedded platform reveal that ALCNet performs best in real-time capability, whereas IDNA-UNet effectively balances segmentation accuracy and computational complexity, with its inference speed generally meeting real-time requirements.
Traditional algorithms such as Top-hat and IPI are easily disturbed by dense background noise, resulting in large false alarm areas, or even missed detection, with poor performance. In contrast, deep learning algorithms such as MD&FA have much better results, but they are also prone to identifying small false alarm areas at some local high points. They cannot complete high-quality multi-scale small target detection. IDNA-UNet can still accurately detect targets of different sizes in an environment with a low signal-to-noise ratio while effectively avoiding noise interference and reducing the false alarm rate.
We visualized the infrared weak small target results of each method, as shown in Figure 6. As can be seen from rows 1 and 2, when there are fewer targets or the background is relatively simple, all algorithms, including traditional target detection algorithms, have good performance. However, the Top-hat algorithm performs poorly in shape segmentation performance.
As can be seen from rows 3 and 4, when there is background interference of high brightness and suspected small targets around the target, the Top-hat and IPI algorithms cannot adapt to the needs of this scene. The probability of missed detection and false alarms is very high. Both MDvsFA and MSHNet have missed detections. The occurrence of false alarms shows that relying only on small-area target detection will increase the risk of false alarms, which highlights the effectiveness of our proposed method in integrating global and local features.
For row 5, which is more difficult to detect due to the complex background around it and more than one small target in the scene, the traditional detection algorithm and MD&FA have a large number of false alarms. In contrast, the MSHNet and DNANet algorithms have missed detections. IDNA-Net is suitable for complex scenarios and generates outputs with accurate shapes and low false positive rates. This further highlights that our method not only ensures the detection rate but also achieves a good balance between the detection rate and the false alarm rate.

4.4. Ablation Experiments

We apply ablation experiments to demonstrate the effectiveness of the proposed modules and techniques. Experiments without DNIM use a conventional residual network-based U-Net network. The experiment uses P d , F a , and IoU to evaluate the target localization ability and boundary prediction ability of our method.
In Table 3, compared with the conventional U-Net network, the IoU and Fa values of the DNIM backbone network are significantly reduced by 18.84% and 50.24%. Therefore, the feature extraction module has higher detection accuracy and a lower false alarm rate for small target detection.
The BFPFM module and SLS loss function have a certain improvement effect on the infrared target detection effect. If the BFPFM module is removed, the performance of IoU, Pd, and Fa values on the NUDT-SIRST dataset will decrease by 1.5%, 0.73%, and 10.94 × 10 6 . Therefore, a better feature fusion effect can be achieved through adaptive feature enhancement of the multi-scale head and BFPFM module.
The use of the SLS loss function has improved IoU, P d , and F a values. In particular, the SLS loss function has the greatest impact on F a . Suppose the SLS loss function is not used. In that case, Fa performance will drop by 13.33 × 10 6 , indicating that the loss function can guide the optimization of network parameters during training and improve the detection performance of small targets without affecting the network structure.

4.5. Comparison of SLS Loss Function Performance

We also compare the detection performance of mainstream target detection methods using SLS and Soft IoU loss functions. Table 4 shows the results of infrared dim target detection methods trained using different loss functions.
The detection performance and IoU indicators of different detectors have been improved through the SLS loss function, so the introduction of loss can effectively improve the small target detection performance of the IDNA-UNet model. In addition, through the ablation experiments of DNANet and MSHNet models, it can be seen that the improvement effects are different for different model architectures. In general, the SLS loss function has certain effectiveness and generalization in the field of small target detection.

5. Conclusions and Discussion

We propose a new method called IDNA-UNet for infrared weak and small target detection. We input the image into the feature extraction network, which is composed of densely nested interactive modules (DNIMs) based on U-Net. Dense connection paths are added to the U-Net structure to achieve level-by-level feature fusion, which is conducive to retaining the detailed information of the target in the feature extraction. The channel attention module SENet is introduced in the residual unit module of each DNIM to re-weight the feature map of each channel. In the feature fusion network, a bottom-up feature pyramid fusion module is designed to modulate multi-scale features, embed the small-scale information of low-level features into high-level semantic feature information, and further retain the small target features. From the detection results, this algorithm has a relatively superior performance. It can achieve an excellent balance between the detection rate and the false alarm rate while ensuring the detection rate, thereby completing the infrared small target detection task under different complex backgrounds. Although the algorithm in this paper has made some breakthroughs in infrared weak small target detection, it still faces challenges in practical applications, such as high computational burden and poor adaptability to small targets and dynamic backgrounds. These data-driven methods are all customized segmentation networks, which are essentially black-box models, rather than combining deep learning-based small target detection algorithms with traditional domain knowledge. Only a few studies have attempted such an approach [7,29,30]. In the future, further progress can be made by leveraging the efficient noise suppression and background modeling capabilities of traditional methods, along with the ability of deep learning to automatically extract useful features from complex data. A more proactive exploration of combining traditional image processing techniques with deep learning methods, in an almost “white-box” manner, can be undertaken for infrared small target detection tasks. This will help achieve more stable, accurate, and efficient detection when dealing with complex backgrounds, multi-target scenarios, and low signal-to-noise ratio challenges.

Author Contributions

Conceptualization, Y.L. and X.D.; methodology, J.Z., W.Z. and X.D.; software, X.D. and Y.W.; validation, K.C. and Y.W.; formal analysis, F.Y., X.D. and K.C.; investigation, F.Y. and J.Z.; resources, F.Y., W.Z. and Y.L.; data curation, X.D. and K.C.; writing—original draft preparation, X.D; writing—review and editing, K.C. and Y.W.; visualization, X.D. and K.C.; supervision, J.Z. and Y.L.; project administration, J.Z.; funding acquisition, F.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the results of this study are included in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, W.; Cong, M.; Wang, L. Algorithms for optical weak small targets detection and tracking. In Proceedings of the International Conference on Neural Networks and Signal Processing, Nanjing, China, 14–17 December 2003; IEEE: Piscataway, NJ, USA, 2003; Volume 1, pp. 643–647. [Google Scholar]
  2. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. In Proceedings of the Signal and Data Processing of Small Targets, Bellingham, WA, USA, 4 October 1999; SPIE: Bellingham, WA, USA, 1999; Volume 3809, pp. 74–83. [Google Scholar]
  3. Bae, T.W.; Zhang, F.; Kweon, I.S. Edge directional 2D LMS filter for infrared small target detection. Infrared Phys. Technol. 2012, 55, 137–145. [Google Scholar] [CrossRef]
  4. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  5. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared patch-image model for small target detection in a single image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  6. Dai, Y.; Wu, Y. Reweighted infrared patch-tensor model with both nonlocal and local priors for single-frame small target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef]
  7. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional local contrast networks for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  8. Wang, H.; Zhou, L.; Wang, L. Miss detection vs. false alarm: Adversarial learning for small object segmentation in infrared images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; IEEE/CVF: Piscataway, NJ, USA, 2019; pp. 8509–8518. [Google Scholar]
  9. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2022, 32, 1745–1758. [Google Scholar] [CrossRef]
  10. Wu, X.; Hong, D.; Chanussot, J. UIU-Net: U-Net in U-Net for infrared small object detection. IEEE Trans. Image Process. 2022, 32, 364–376. [Google Scholar] [CrossRef]
  11. Rivest, J.F.; Fortin, R. Detection of dim targets in digital infrared imagery by morphological image processing. Opt. Eng. 1996, 35, 1886–1893. [Google Scholar] [CrossRef]
  12. Zhang, L.; Peng, Z. Infrared small target detection based on partial sum of the tensor nuclear norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  13. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric Contextual Modulation for Infrared Small Target Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; IEEE/CVF: Piscataway, NJ, USA, 2021; pp. 949–958. [Google Scholar] [CrossRef]
  14. Zhang, T.; Li, L.; Cao, S.; Pu, T.; Peng, Z. Attention-guided pyramid context networks for detecting infrared small target under complex background. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 4250–4261. [Google Scholar] [CrossRef]
  15. Zhao, B.; Wang, C.; Fu, Q.; Han, Z. A Novel Pattern for Infrared Small Target Detection With Generative Adversarial Network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 9102–9113. [Google Scholar] [CrossRef]
  16. Zhou, H.; Tian, C.; Zhang, Z.; Li, C.; Xie, Y.; Li, Z. PixelGame: Infrared small target segmentation as a Nash equilibrium. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6255–6266. [Google Scholar] [CrossRef]
  17. He, X.; Ling, Q.; Zhang, Y.; Lin, Z.; Zhou, S. Detecting dim small target in infrared images via subpixel sampling cuneate network. IEEE GEoscience Remote Sens. Lett. 2022, 19, 6513005. [Google Scholar] [CrossRef]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. Zunair, H.; Hamza, A.B. Sharp U-Net: Depthwise convolutional network for biomedical image segmentation. Comput. Biol. Med. 2021, 136, 104699. [Google Scholar] [CrossRef]
  20. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  21. Huang, Y.; Tang, Z.; Chen, D.; Su, K.; Chen, C. Batching soft IoU for training semantic segmentation networks. IEEE Signal Process. Lett. 2019, 27, 66–70. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  23. Zhang, Y.; Wang, X.; Li, Y. Shallow U-Net for Efficient Small Object Segmentation in Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2021, 18, 650–654. [Google Scholar]
  24. Smith, A.; Johnson, B. Optimizing U-Net Architectures for High-Resolution Small Object Detection. Int. J. Comput. Vis. 2022, 130, 350–366. [Google Scholar]
  25. Lee, H.; Kim, S.; Park, J. Lightweight U-Net Variants for Real-Time Small Object Detection. Pattern Recognit. Lett. 2020, 138, 45–52. [Google Scholar] [CrossRef]
  26. Duchi, J.C.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  27. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Lin, J.; Su, H.; Jin, W.; Zhang, Y.; et al. A dataset for infrared image dim-small aircraft target detection and tracking under ground/air background. Sci. Data Bank 2019, 5, 4. [Google Scholar]
  28. Zhang, T.; Li, L.; Zhang, Z.; Peng, Z. ISNet: Shape Matters for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9221–9233. [Google Scholar]
  29. Li, L.; Zhang, T.; Pu, T.; Peng, Z. Ristdnet: Robust infrared small target detection network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7738–7750. [Google Scholar]
  30. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. Effective infrared small target detection utilizing a novel local contrast method. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1822–1826. [Google Scholar] [CrossRef]
Figure 1. The architecture of IDNA-UNet consists of the feature extraction module and the feature fusion module.
Figure 1. The architecture of IDNA-UNet consists of the feature extraction module and the feature fusion module.
Sensors 25 00814 g001
Figure 2. The network structure of convolutional neural network with residual connections.
Figure 2. The network structure of convolutional neural network with residual connections.
Sensors 25 00814 g002
Figure 3. The structure of SENet, including squeeze and excitation.
Figure 3. The structure of SENet, including squeeze and excitation.
Sensors 25 00814 g003
Figure 4. The structure of the bottom-up feature pyramid fusion module.
Figure 4. The structure of the bottom-up feature pyramid fusion module.
Sensors 25 00814 g004
Figure 5. ROC curves of different methods on NUDT-SIRST.
Figure 5. ROC curves of different methods on NUDT-SIRST.
Sensors 25 00814 g005
Figure 6. Visual comparison of experimental results of different methods. The targets to be detected are marked with green circles, and missed detections and false detections are marked with red and yellow dotted circles, respectively.
Figure 6. Visual comparison of experimental results of different methods. The targets to be detected are marked with green circles, and missed detections and false detections are marked with red and yellow dotted circles, respectively.
Sensors 25 00814 g006
Table 2. Experimental results on the NUDT-SIRST and IRSTD-1k dataset.
Table 2. Experimental results on the NUDT-SIRST and IRSTD-1k dataset.
ModelNUDT-SIRSTISTD-1k#Params (M)Latency (ms)
IoU↑PdFaIoU↑PdFa
Top-hat22.3470.5495.3712.5668.94188.78--
IPI18.6767.8745.7826.0170.3936.69--
MD&FA52.8482.3450.5552.8886.931.323.139.66
ALCNet78.4596.3435.9948.0293.329.871.56.63
DNANet82.1798.9323.6563.0197.79.7174.6924.05
MSHNet81.0798.2525.8164.5197.812.374.059.06
Ours82.3498.7512.5466.1597.87.14.6817.68
Table 3. Ablation of the DNIM, BFPFM, and L S L S in IDNA-UNet.
Table 3. Ablation of the DNIM, BFPFM, and L S L S in IDNA-UNet.
DNIMBFPFM L SLS IoU↑ P d F a
×63.5097.7062.78
×80.8498.0223.48
×82.2798.0125.87
82.3498.7512.54
Table 4. Performance comparison of common object detection methods (IDNA-UNet, DNANet, MSHNet) using L S L S and Soft IoU loss functions.
Table 4. Performance comparison of common object detection methods (IDNA-UNet, DNANet, MSHNet) using L S L S and Soft IoU loss functions.
Loss FunctionIDNA-UNetDNANetMSHNet
IoU↑ P d F a IoU↑ P d F a IoU↑ P d F a
L i o u 82.2798.0125.8782.1796.9323.6581.0397.8024.66
L S L S 82.3498.7512.5480.7597.3141.1381.0798.2525.81
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Du, X.; Cheng, K.; Zhang, J.; Wang, Y.; Yang, F.; Zhou, W.; Lin, Y. Infrared Small Target Detection Algorithm Based on Improved Dense Nested U-Net Network. Sensors 2025, 25, 814. https://doi.org/10.3390/s25030814

AMA Style

Du X, Cheng K, Zhang J, Wang Y, Yang F, Zhou W, Lin Y. Infrared Small Target Detection Algorithm Based on Improved Dense Nested U-Net Network. Sensors. 2025; 25(3):814. https://doi.org/10.3390/s25030814

Chicago/Turabian Style

Du, Xinyue, Ke Cheng, Jin Zhang, Yuanyu Wang, Fan Yang, Wei Zhou, and Yu Lin. 2025. "Infrared Small Target Detection Algorithm Based on Improved Dense Nested U-Net Network" Sensors 25, no. 3: 814. https://doi.org/10.3390/s25030814

APA Style

Du, X., Cheng, K., Zhang, J., Wang, Y., Yang, F., Zhou, W., & Lin, Y. (2025). Infrared Small Target Detection Algorithm Based on Improved Dense Nested U-Net Network. Sensors, 25(3), 814. https://doi.org/10.3390/s25030814

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop