Next Article in Journal
Binary Hunter–Prey Optimization with Machine Learning—Based Cybersecurity Solution on Internet of Things Environment
Next Article in Special Issue
A Short Video Classification Framework Based on Cross-Modal Fusion
Previous Article in Journal
A Residual-Dense-Based Convolutional Neural Network Architecture for Recognition of Cardiac Health Based on ECG Signals
Previous Article in Special Issue
An Asymmetric Encryption-Based Key Distribution Method for Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution

1
School of Computer and Information Security, Guilin University of Electronic Science and Technology, Guilin 541004, China
2
School of Artificial Intelligence, Guilin University of Electronic Science and Technology, Guilin 541004, China
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(16), 7205; https://doi.org/10.3390/s23167205
Submission received: 6 July 2023 / Revised: 3 August 2023 / Accepted: 4 August 2023 / Published: 16 August 2023
(This article belongs to the Special Issue Smart Mobile and Sensing Applications)

Abstract

:
Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance of modeling cross-modality features. To address these challenges, we propose Dual-branch Progressive learning for infrared and visible image fusion with a complementary self-Attention and Convolution (DPACFuse) network. On the one hand, we propose Cross-Modality Feature Extraction (CMEF) to enhance information interaction and the extraction of common features across modalities. In addition, we introduce a high-frequency gradient convolution operation to extract fine-grained information and suppress high-frequency information loss. On the other hand, to alleviate the CNN issues of insufficient global information extraction and computation overheads of self-attention, we introduce the ACmix, which can fully extract local and global information in the source image with a smaller computational overhead than pure convolution or pure self-attention. Extensive experiments demonstrated that the fused images generated by DPACFuse not only contain rich texture information, but can also effectively highlight salient objects. Additionally, our method achieved approximately 3% improvement over the state-of-the-art methods in MI, Qabf, SF, and AG evaluation indicators. More importantly, our fused images enhanced object detection and semantic segmentation by approximately 10%, compared to using infrared and visible images separately.

1. Introduction

Infrared and visible image fusion (IVF) has gained significant attention and is widely used in various applications [1]. Specifically, the effective fusion of shared and specific features from different modalities enables the generation of high-quality fused images, which, in turn, benefits downstream tasks, such as object detection [2,3,4], medical image processing [5,6], semantic segmentation [7,8,9], and pedestrian detection [10,11]. Although IVF has received much attention in various applications, IVF remains challenging due to the significant differences in appearance between these two image types. In infrared images, thermal target structures can be effectively highlighted, but these images often exhibit low contrast and blur properties. On the other hand, visible images have rich color and texture information, but they are easily affected by factors such as illumination and weather. Therefore, the effective fusion of these two different modalities into high-quality images still faces many technical difficulties.
Recently, there has been a surge in the development of deep learning-based IVF algorithms [12,13,14,15,16]. These methods typically involve feature extraction from the source images, the fusion of the extracted features, and the reconstruction of the fused features to obtain the final fused image [17]. Networks with strong feature extraction capabilities can usually synthesize fused images with better quality.
The main feature extraction method is currently the CNN-based auto-encoder structure, which is mainly divided into a shared encoder structure [18,19,20] and a dual-branch private encoder structure [21,22,23]. However, although the above-mentioned auto-encoder structure has good feature extraction ability, there are also some shortcomings. Firstly, the structure, based on the shared encoder, cannot distinguish the unique information in each mode. Secondly, the structure of the dual-branch private encoder often ignores the common features between modes, such as background and some large-scale features. In addition, the context-free CNN structure can only extract local features in a small service domain, and it is difficult to extract global information to generate higher-quality fusion images [15]. Moreover, many IVF networks may cause high-frequency information loss in image features when performing forward propagation [24]. Therefore, to address the above issues in IVF, this paper proposes a progressive feature extractor Cross-Modality Feature Extraction (CMFE). By introducing the CMFE module, compared with the structure of the shared encoder, our model can effectively extract the shared features between different modalities, while also better distinguishing the unique features between different modalities, to better realize IVF. Compared with the dual-branch private encoder, our structure can enhance the information interaction between different modalities to ensure thorough feature extraction, and our structure can better integrate the unique and shared features of modalities.
Vision Transformer has received extensive attention on many vision tasks. Therefore, many scholars have adopted Transformer-based methods in the feature extraction stage [13,15,25,26]. However, many transformer-based models are often limited by computational resources, the size of the input image, and weak local perception, which limits the timeliness and applicability of the IVF. To synthesize the respective advantages of Transformer and CNN architectures, we introduce ACmix [27], which integrates the flexibility of self-attention and the lightness of convolution. Therefore, we propose a complementary fusion network based on convolution and a multi-head self-attention mechanism to solve the IVF problem, integrating the advantages of CNN in extracting local information and computing portability and the ability of self-attention in context awareness and long-distance modeling. Compared with CNN-based architectures, our architecture has more powerful feature extraction capabilities and can better extract deep features in modalities. Additionally, compared with some Transformer-based architectures, our architecture has better flexibility (e.g., no fixed image size, small computational cost, etc.) and adaptability.
To this end, we propose Dual-branch Progressive learning for IVF with complementary self-Attention and Convolution (DPACFuse), such that the network can take into account the unique features and shared features of modalities, and integrate the respective advantages of CNN and a multi-head self-attention mechanism to better realize the IVF. The main contributions of this work can be summarized as follows:
  • We propose a dual-branch progressive image fusion framework, based on complementary self-attention and convolution, for the IVF, which can take into account both global and local information of the source image to achieve better feature fusion.
  • We propose Cross-Modality Feature Extraction (CMFE) to enhance the information interaction between modalities, while suppressing the loss of high-frequency information.
  • Extensive experimental results on the public image fusion datasets, MSRS, RoadScene, and TNO, demonstrate that our method is superior to the current general fusion framework. Meanwhile, we investigate the facilitation of our fused images for object detection and semantic segmentation.
The remainder of this article is organized as follows. In Section 2, we mainly introduce the deep learning-based IVF methods. In Section 3, we introduce and describe, in detail, our fusion framework and the loss function used. In Section 4, we conduct a large number of experiments to verify the effectiveness of DPACFuse and also to explore the effect of IVF on downstream task promotion. In Section 5, we provide some conclusive proof.

2. Related Work

In this section, we briefly introduce the existing deep learning-based methods, mainly including CNN-based, AE-based, GAN-based, and Transformer-based methods.

2.1. CNN-Based and AE-Based Fusion Methods

In recent years, many CNN-based and AE-based methods have been widely used in the field of image fusion. Among them, the dual-branch structure, based on CNN and AE, greatly improves the fusion performance. For example, Tang et al. [21] proposed a dual-branch private encoder structure, namely SeAFusion, which combined a semantic segmentation network to learn more information, and to achieve better fusion results. Meanwhile, SeAFusion is a pioneer in combining IVF with downstream tasks. Additionally, Tang et al. [28] proposed an illumination-driven IVF network to solve the fusion problem in different lighting scenes. However, the designs of the network structures of the above two methods are too simple to effectively deal with complex situations. Another typical work is that of Res2Fusion, in [29], which describes a network with a dual-branch shared encoder structure, and which introduces Res2net and densely connected structures into the encoder to obtain multi-scale information. In the fusion layer, the fusion layer of double nonlocal attention models is used to realize image fusion. This method fully considers the problem of multi-scale information extraction and global modeling of the model, but it is difficult to deploy in the actual environment due to its high complexity and high computational cost. In addition, the shared encoder structure is also widely used in image fusion. For example, a typical method is DenseFuse [18]. The core concept of this method is to construct a deep neural network with dense connections, comprising an encoder (consisting of convolutional layers and dense layers) and a decoder (used for fusion). However, DenseFuse uses hand-designed fusion rules, so its results are not robust. In order to address the limitation of fusion rules that are designed manually, Li et al. proposed RFN-Nest [19] and NestFuse [20], wherein the former mainly utilized a residual fusion network to solve the problem, while the latter adopted the idea of combining spatial and channel attention mechanisms to solve the problem. Moreover, the IFCNN in [22], the PMGI in [30], and the U2Fusion in [31] proposed unified end-to-end networks to realize different fusion tasks.
Although many fusion methods based on CNN and AE have achieved good results, they usually adopt relatively simple CNN structures and hand-designed fusion rules, which limit the global modeling ability of the model and the ability to extract detailed information. At the same time, many methods lack information interaction in the process of feature extraction, which makes it impossible to fully extract more complementary information. In contrast, our method adopts the idea of cross-modality interaction to achieve image fusion, which helps to eliminate the mismatch and noise between different modalities and brings advantages in terms of improving the robustness of the fused image.

2.2. GAN-Based Fusion Methods

The Generative Adversarial Network (GAN) can learn the distribution characteristics of data and generate samples that conform to a specific distribution, which is also widely used in IVF. FusionGAN [32] was the first algorithm to apply this method to realize IVF. However, since only a single discriminator is used in FusionGAN, it cannot balance the information from the different modalities, leading to the loss of a lot of texture information in the fused images. In order to overcome the shortcomings of a single discriminator, Ma et al. [33] proposed a dual discriminator structure, namely DDcGAN, to achieve information fusion with image fusion. Moreover, Rao et al. [17] proposed AT-GAN, which introduced an intensity attention module and a semantic transition module to remove redundant information in infrared and visible images, respectively. At the same time, the quality assessment module is used to achieve the information balance between different modalities. In addition, infrared images have been greatly developed in various object detection tasks, such as pedestrian detection [11,34] and infrared small target detection [35,36]. However, due to the imaging characteristics of infrared images, the application scenarios of these methods are very limited. Therefore, there are many works [37,38,39] that combine IVF with object detection to overcome the limitations of using only a single modality. For example, Liu et al. [12] designed a GAN-based object perception network, TarDAL, which generated high-quality fused images and excellent detectors by including the IVF network and the object detection network in a bilevel optimization formulation.
Although GAN-based models have been widely used in the field of image fusion, the GAN-based fusion method emphasizes that discriminator learning simulates the distribution of the original image data, which may lead to poor image quality. At the same time, finding a way to balance the information from the different modalities is still a problem that needs to be studied.

2.3. Transformer-Based Fusion Methods

The transformer [40] structure is based on a multi-head self-attention mechanism, designed for sequence modeling and transduction tasks, and is known for its focus on long-term dependencies in data. Transformer has seen great success not only in NLP, but also in various visual tasks [41,42,43]. Many models based on Transformer have also been highly developed in the field of image fusion. For instance, Wang et al. [15] proposed a pure transformer fusion network, called SwinFuse, which used the powerful feature representation capability of the self-attention mechanism to perform image fusion. However, it uses a hand-designed fusion strategy, which does not perform well enough in handling fine-grained information. Additionally, Zhao et al. [44] introduced the Dual-branch Transformer and the structure of DenseNet (DNDT), which could consider more complete image information. In addition, inspired by the work of Swin Transformer [45], Ma et al. [13] proposed Swin Fusion, a network architecture for multimodal fusion. In addition, Rao et al. [46] proposed TGFuse, which embedded the Transformer in a GAN-based fusion network to achieve IVF. Furthermore, Qu et al. [26] proposed TransMEF for multi-exposure image fusion, which combined CNN and Transformer to obtain powerful local modeling and global modeling capabilities. However, this method is less flexible and can only input images of fixed size.
Although many transformer-based models perform well in many fusion tasks, many methods still suffer from poor flexibility and poor ability to model trans-membrane states, such as DNDT [44], TransMEF [26], and CGTF [47]. Furthermore, many transformer-based models are computationally expensive, while our method combines the excellent computational efficiency of CNN and the excellent global modeling ability of self-attention to better realize image fusion.

3. Methodology

3.1. Network Architecture

The network architecture of DPACFuse is illustrated in Figure 1a, and is composed of three main phases: feature extraction, feature fusion, and feature reconstruction. Given a pair of aligned infrared (IR) and visible (VI) images, denoted as I i r R H × W × C i n and I v i R H × W × C i n , respectively, the fused image I f R H × W × C o u t is obtained through these phases.
In the feature extraction phase, we extract the specific features of the respective modalities separately using a dual-branch structure. First, we obtain shallow features { F i r 1 , F v i 1 } from the source image through a 3 × 3 convolutional layer. This can be expressed as:
{ F i r 1 , F v i 1 } = { H s e ( I i r ) , H s e ( I v i ) }
where H s e ( · ) represents a 3 × 3 convolutional layer, whose activation function is Leaky Relu and the stride is 1. The convolutions usually have stable optimization performance and are very good at early visual processing. At the same time, convolution has a strong local perception ability, which can effectively mine local information and map it to high-dimensional space.
Then, the ACmix is embedded in the respective branches of the IR and VI images to extract their respective specific features. At the same time, CMFE is deployed between the two modalities to extract their common features, thereby guiding the network to generate better images. We represent the feature extraction of the process in two stages. The intermediate features { F i r 2 , F v i 2 } obtained in the first stage can be expressed by the following formula:
F i r C F 1 , F v i C F 1 = C M E F F i r 1 , F v i 1 F i r A C 1 , F v i A C 1 = A C m i x F i r 1 , A C m i x F v i 1 F i r 2 , F v i 2 = F i r C F 1 F i r A C 1 , F v i C F 1 F v i A C 1
After obtaining the intermediate features { F i r 2 , F v i 2 } in the first stage, we use them as input for the second stage to obtain the output { F i r 3 , F v i 3 } in a similar manner to the first stage:
F i r C F 2 , F v i C F 2 = C M E F F i r 2 , F v i 2 F i r A C 2 , F v i A C 2 = A C m i x F i r 2 , A C m i x F v i 2 F i r 3 , F v i 3 = F i r C F 2 F i r A C 2 , F v i C F 2 F v i A C 2
Then, the features of these two modalities { F i r , F v i } are obtained by a convolutional layer H d e ( · ) . The process can be expressed by the following formula:
{ F i r , F v i } = { H d e ( F i r 3 ) , H d e ( F v i 3 ) }
Finally, we reconstruct the fused image through the feature fusion and image reconstruction module. Since our feature extraction network has a strong enough extraction ability, we opted for a straightforward approach by employing the cascade fusion strategy to directly fuse the F i r and F v i . The fusion process is represented as follows:
F f = H c ( F i r , F v i )
where F f represents the fused feature, and H c ( · ) represents the cascade on the channel dimension. Finally, we can obtain the output I f through the feature reconstructor H R ( · ) :
I f = H R ( F f )

3.2. Specific Framework of ACmix

Due to the excellent context-aware ability of the multi-head self-attention mechanism and the lightness of convolution, we introduce ACmix. As shown in Figure 1c, it can be divided into: projection reconstruction, extract local features, and extract global features.
First, image features I i R H × W × C i n are obtained and, after the projection reconstruction, local and global features are extracted, to obtain the output F o u t R H × W × C o u t . In the projection reconstruction stage, the feature map I i is passed through three separate 1 × 1 convolutional layers, resulting in the generation of three feature maps I i 1 , I i 2 , I i 3 R H × W × C o u t .
In the extract local feature stage, the steps of this stage are different from the traditional standard convolution, that is, we first perform a linear projection of kernel weights, then translate according to the kernel position, and finally aggregate. Firstly, the three feature maps I i 1 , I i 2 , I i 3 in the projection reconstruction stage, which are divided into N groups in the depth direction, and then reshaped to obtain a feature map with the dimensions of R N × C o u t N × H W . After partitioning the feature maps into groups and reshaping them, the resulting feature maps are concatenated to create a new feature map X R 3 N × C o u t N × H W . This concatenated feature map X is then fed through a lightweight fully connected layer to generate Z R k 2 N × C o u t N × H W . Then, Z is subjected to a reshaping operation and then a shift aggregation operation, which is realized by depthwise convolution. Specifically, Z is divided into N groups, and each group Z l R H × W × C o u t N is used as a basic unit of convolution. Finally, the results of N groups are spliced to obtain the output F c o n v R H × W × C o u t of the convolution path. The entire extract local feature stage can be expressed as:
X = Connect ( R ( I i 1 ) , R ( I i 2 ) , R ( I i 3 ) ) Z = F C k ( X ) F c o n v = l = 1 N C d e v ( Z l )
where C o n n e c t ( · ) , R ( · ) , and F C k ( · ) denote the concatenation, reshape, and the light fully connected layer with a kernel size of k, respectively. The value C d e v ( · ) represents the convolution operation in depthwise convolution and Z l represents the input of group lth. The symbol | | denotes the concatenation of the results obtained from all N groups, and the entire process corresponds to depthwise convolution with kernels of size 3. The processing of the extract local feature stage is the same as the traditional convolution operation.
In the extract global feature stage, the multi-head self-attention mechanism is adopted. Specifically, we divided the three feature maps I i 1 , I i 2 , I i 3 obtained in the projection reconstruction stage into N groups (i.e., N attention mechanism heads) in the depth direction, and obtained the Q R H × W × C o u t N , K R H × W × C o u t N , V R H × W × C o u t N of each head. Then we flattened Q, K and V to obtain Q R H W × C o u t N , K R H W × C o u t N and V R H W × C o u t N , and then used these as the inputs of the attention function:
Atention ( Q , K , V ) = Softmax Q K T d k V h e a d j = A t e n t i o n ( F ( I i W j q ) , F ( I i W j k ) , F ( I i W j v ) ) F a t t = l = 1 N R ( h e a d l )
where W j q , W j k , and W j v R C i n × C o u t N are corresponding input projection weights (for ease of presentation, here, we include the process of the first stage). F ( · ) and R ( · ) denote the flatten and reshape operations, respectively. The symbols d k , h e a d j R H × W × C o u t N and | | denote the dimension of K , the output of the j t h head, and the splicing of N heads, respectively. The symbol F a t t R H × W × C o u t represents the final output of the extract global feature stage, which is obtained by concatenating the outputs of the N self-attention heads.
Finally, the features extracted by the ACmix module can be expressed as the sum of the extract local feature path and the extract global feature path output, where the weights are determined by two learnable parameters α and β :
F o u t = α F c o n v + β F a t t

3.3. Specific Framework of CMEF

To enhance the information interaction between modalities, as well as to suppress the loss of high-frequency information, we propose a CMEF module, the model of which is shown in Figure 1b. We introduce the module in two stages: feature combination and feature recombination.
In the feature combination stage, the input features F 1 R H × W × C i n , F 2 R H × W × C i n are given. We first concatenate F 1 and F 2 to get the fusion feature F c a t R H × W × 2 C i n , and then obtain the common feature F c f R H × W × C o u t through the foreground-aware spatial attention and feature-level attention mask. The specific process is as follows:
F c a t = C o n c a t ( F 1 , F 2 ) F c f = ϕ s ( F c a t ) ϕ p C o n v _ 1 ( F c a t ) F c a t
where C o n c a t ( · ) represents the operation of concatenating the features. The symbol ϕ s ( · ) stands for the foreground-aware spatial attention operation which is achieved by calculating the channel-wise maximum value of fusion features. The symbol ϕ p ( · ) represents the feature-level attention mask achieved by a Multiple Layer Perceptron (MLP), and C o n v _ 1 ( · ) followed by a 2-cls softmax operation. This feature-level attention mask means that ϕ p ( · ) can predict a re-scaling score to combine features from different modalities in such a way that the combined features are independent of the specific features.
In the feature recombination stage, it is well known that useless information has a huge impact on image fusion, which misleads the fusion direction of the model, resulting in distortion of the fused image. At this stage, we hope to obtain more common features and filter the interference of useless information as much as possible. Specifically, we integrate these shared features with the fine-grained information of their respective features through channel rescaling operations:
F o u t i r = S E ( F c f ) ϕ G D ( F 1 ) C o n v ( F 1 ) F o u t v i = S E ( F c f ) ϕ G D ( F 2 ) C o n v ( F 2 )
where S E ( · ) represents the Squeeze-and-Excitation Network [48] (its framework is shown in Figure 1b), which can assign weights to each channel to effectively filter the impact of useless information on the fusion process. The symbol ϕ G D ( · ) refers to the gradient convolution operation, and ∇ stands for the gradient operator. Moreover, ⊕ and C o n v ( · ) denote the operation of element-wise addition and 3 × 3 convolution, respectively. Finally, F o u t i r , F o u t v i R H × W × C o u t , respectively, add to their respective characteristics of the backbone network.

3.4. Loss Function

To minimize information loss and improve fusion performance, this paper employs three distinct loss functions for training the network: texture loss, intensity loss, and SSIM (Structural Similarity Index) loss. These loss functions constrain the network from different perspectives. The loss function used in our network can be represented as follows:
L t o t a l = γ 0 L i n t + γ 1 L t e x t u r e + γ 2 L s s i m
where L i n t , L t e x t u r e , and L s s i m represent the intensity loss, texture loss, and SSIM loss, respectively. The parameters γ 0 , γ 1 , and γ 2 are hyper-parameters to represent the contributions of the three losses to the entire loss, respectively.
The intensity loss emphasizes the preservation of pixel intensity information, and it helps the model better learn the overall brightness information and contrast characteristics. Intensity loss is defined as:
L i n t = 1 H W I f m a x ( I i r , I v i ) 1
where · 1 denotes l 1 norm, and max ( · ) represents the maximum value in an element. By emphasizing the overall brightness and contrast characteristics, it enables the model to better understand and learn these important visual attributes.
The texture loss is a key component in image fusion, as it aims to preserve the intricate and fine-grained texture details during the fusion process. We define texture loss as:
L t e x t u r e = 1 H W I f m a x ( I i r , I v i ) 1
where the symbol ∇ represents the Sobel gradient operator. The absolute value calculation, denoted by | · | , is applied to the gradient values to ensure that only positive magnitudes are considered. The value · 1 represents l 1 norm, and m a x ( · ) selects the maximum value from the corresponding elements in the calculation.
The SSIM loss is employed to facilitate the learning of structural information by the model from the input images, and it also takes into account not only structure and contrast, but also illumination, which can be expressed as follows:
L s s i m = ( 1 S S I M ( I f , I i r ) ) / 2 + ( 1 S S I M ( I f , I v i ) ) / 2

4. Experiments

In this section, we provide specific details of the experimental implementation. We then compare DPACFuse with seven other methods. Finally, we demonstrate the outstanding performance of DPACFuse on downstream tasks.

4.1. Experimental Configurations

Datasets. The IVF experiments used three public datasets to verify our fusion method, which were MSRS [28], RoadScene [31] and TNO [49]. We trained our IVF network on the MSRS dataset, which contained 1083 pairs of registered images with semantic labels of nine typical scenes. In addition, we employed the MSRS test set (361 pairs), RoadScene (30 pairs), and TNO (30 pairs) as test datasets to comprehensively verify the performance of DPACFuse. Among them, the RoadScene dataset contained 221 image pairs that mainly focused on capturing typical traffic scenes, including roads, pedestrians, and vehicles. The TNO dataset consisted of multispectral night and day images depicting various military-related scenes.
Evaluation metrics and comparison methods. We used EN, SD, SF, MI, VIF, AG, Qabf, and FMI_pixel as evaluation metrics. In addition, higher metrics implied that the quality of the fusion image was better. Details on these evaluation metrics can be found in [50]. At the same time, we compared DPACFuse with the state-of-the-art methods, including DenseFuse [18], IFCNN [22], U2Fusion [31], SDNet [16], GANMcC [51], SwinFusion [13], and TarDAL [12].
Experimental setup. Our experiments were conducted on a computer equipped with one NVIDIA GeForce RTX 3090 GPU. The proposed method was implemented using the PyTorch platform. Moreover, all input images were normalized to [0, 1] before training. The following values were used for the hyperparameters of the experiment: the initial parameters α and β for the balanced convolution and self-attention paths were set to 1, respectively, the total number of self-attention heads was N = 4 , and the kernel size was k = 3 for the fully connected layer. In addition, the hyperparameters were γ 0 = 20 , γ 1 = 20 and γ 2 = 1 for balancing each loss function. The network parameters were updated using the Adam optimizer with a momentum term of (0.9, 0.999). The training was performed with a batch size of 2, an initial learning rate of 0.001, and a weight decay of 0.0002.

4.2. Comparative Experiment

4.2.1. Qualitative Results

We selected two groups of images in the MSRS test set for subjective evaluation, wherein each group contained two typical scenes that were day and night.
In the daytime scene with sufficient illumination, the VI image contained abundant texture detail and fully showed the environmental information. Although the ability of IR images to display the environment was limited, they could provide semantic information about the structure of thermal targets. By integrating this complementary information, the fusion image could provide comprehensive scene information, and could effectively enrich the semantic information. As presented in Figure 2 and Figure 3, due to the interference of useless information, the salient targets in DenseFuse, IFCNN, and U2Fusion methods weakened to varying degrees and could not maintain their original intensities. We highlighted salient regions with green boxes to illustrate this problem. Although SDNet and GANMcC could maintain the highlight intensity of infrared targets, their performances in retaining texture information was poor, and we illustrated the problem by zooming in on the areas with red boxes. In addition, compared with SwinFusion and TarDAL, DPACFuse not only retained more detailed information, but also better preserved the edge information, as can be seen from the enlarged floors, as well as the steps.
In the dark scene with insufficient illumination, due to the influence of illumination, VI images could only provide limited environmental information, and objects in them were not easy to identify, while IR images were not sensitive to illumination. Therefore, adaptive realization of the IVF in the case of the dark scene was very important, whilst being very challenging. As presented in Figure 2 and Figure 3, all methods could effectively construct the scene information, but there were great differences between different algorithms. Excepting our DPACFuse, SwinFusion, and TarDAL, the other methods failed to maintain the highlighting of thermal targets in infrared images, which we illustrated with the green boxes. In addition, DPACFuse was better than the other methods, such as SwinFusion, in maintaining details, which we illustrated by magnifying the red area.
Overall, the experimental results highlighted the superior performance of DPACFuse in both daytime and dark scenes. It effectively preserved texture details, edge information, and the saliency of thermal targets, demonstrating its superior ability to perform IVF in a variety of environmental conditions.

4.2.2. Quantitative Result

Figure 4 displays the quantitative results of the eight evaluation indicators on the MSRS test set. DPACFuse demonstrated superior performance in nearly all metrics, showcasing its ability to effectively extract information from the source images and its versatility across various complex scenes. The best EN, MI, and FMI_pixel indicated that our fused image contained the most information, and the highest Qabf and AG indicated that our fused image retained the most edge information. In addition, the highest SF and VIF illustrated the best visual effect was presented by our fused image. Although DPACFuse slightly lagged behind SwinFusion in terms of the SD metric, the difference was not significant, which meant that our fused images had good contrast.

4.3. Generalization Experiment

Generalization ability is also an important metric to evaluate a model. We trained the model on the MSRS dataset and verified the generalization ability of DPACFuse on the RoadScene and TNO datasets.

4.3.1. Results of RoadScene

Qualitative analysis. We selected two scenes, day and night, to assess the fusion results, and the visualized results are shown in Figure 5. Observing the results of the daytime scene, we can see that almost all the algorithms suffered from the interference of useless information, among which DenseFuse, U2Fusion, SDNet, GANMcC, and TarDAL were most affected, losing a lot of texture information. We illustrated this problem by zooming in on the red area. In addition, the intensity of the infrared targets of SDNet and GANMcC also weakened to varying degrees, while SwinFusion experienced a decrease in its overall contrast, due to the influence of illumination, which we illustrate by the green box. Except for our DPACFuse and SwinFusion, the other methods weakened in the overall pixel intensity and could not maintain the original pixel intensity.
In the dark scene, it can be seen that DenseFuse, U2Fusion, SDNet, GANMcC, and TarDAL lost a lot of texture details, such as the outline of background leaves and the zebra crossing on the ground. In addition, the salient targets of DenseFuse, SDNet, and GANMcC were severely disturbed by useless information and could not maintain the original pixel intensity. DPACFuse and SwinFusion were only disturbed by a small amount of useless information.
Quantitative analysis. As shown in Figure 6, DPACFuse achieved the highest scores in all indicators, which meant that the fused image generated by DPACFuse not only maintained a lot of information and texture details, but also had the highest contrast and the best visual quality.
The excellent performance of DPACFuse on the RoadScene dataset fully demonstrated the adaptability of our method to various complex traffic scenes, and also proved that DPACFuse has good generalization ability.

4.3.2. Results of TNO

Qualitative analysis. As depicted in the green boxes in Figure 7, DenseFuse, U2Fusion, and IFCNN weakened the strength of salient targets to different extents, with the first two being the most obvious. In addition, GANMcC blurred the contour of the salient targets. Excepting our method and SwinFusion, the fused images of the other methods were affected by other useless spectral information and could not effectively present the texture information, such as the bushes and fences in the red region. It is worth noting that, although SwinFusion had good fusion performance, the fused images suffered from whitening. On the whole, DPACFuse not only excelled in highlighting salient objects, but also effectively preserved the original texture information from the input images.
Quantitative analysis. The results depicted in Figure 8 illustrate that DPACFuse achieved the highest scores in Qabf, MI, VIF, and FMI_pixel metrics. In addition, DPACFuse was also ahead of all the methods, excepting TarDAL, in two metrics: EN and SD. Taking the above analyses together, DPACFuse exhibited excellent performance on the TNO datasets, which further demonstrated its excellent generalization ability.
In conclusion, a large number of experiments on various datasets showed that our method can preserve a large amount of information from the source image and maintain the highlight degree of the infrared target in various complex situations. We attribute these advantages to the following aspects. On the one hand, the CMEF that we designed effectively extracts fine-grained information from source images and enhances the information interaction between different modalities. On the other hand, our network possesses a powerful feature extraction capability, and the ACmix module can effectively extract local and global information.

4.4. Ablation Study

We performed ablation experiments to validate the efficacy of various modules and utilized EN, Qabf, SF, MI, AG, and FMI_pixel for quantitative assessment. In addition, We selected two images for qualitative analysis, one of which was 00537D from MSRS and the other was 00390 from M3FD.
Quantitative analysis. We conducted quantitative experiments on the MSRS test set and summarized the results in Table 1. The values M1 and M2 represent changing ACmix to pure self-attention and convolution, respectively. The data in the table clearly show that the removal of the ACmix led to a decrease in all the indicators. The metrics, MI, SF, and FMI_pixel, showed the most significant declines, indicating a deterioration in the network’s ability to integrate complementary information between modalities. In addition, M3 and M4 denote the removal of HCGD in CMFE and the complete removal of CMFE, respectively. It can be seen that Qabf and AG experienced a large decrease when only HCGD was removed, which illustrated the effectiveness of HCGD in extracting high-frequency information. However, when the CMFE was removed, almost all the indicators significantly decreased, indicating that the performance of the network degraded a lot when there was no interaction between cross-modalities.
Qualitative analysis. As observed in Figure 9, it is evident that the Attention was highly sensitive to whitening, leading to an overall increase in brightness and a loss of fine texture details. Since HCGD is one of the components of CMFE, removing both of them resulted in a significant loss of edges in the fused image. Moreover, removing the entire CMFE module had an even more significant impact, affecting not only the edges but also compromising the background and other critical information.
In summary, the results in Figure 9 and Table 1 indicate the effectiveness and rationality of our designed modules, as well as of the overall network design.

4.5. Downstream IVF Applications

In this section, we applied fused images to object detection and semantic segmentation, and explored the benefits of IVF for downstream tasks.

4.5.1. Object Detection Performance

We employed the pre-trained YOLOv8 [52] detector to detect different images. We randomly selected a test set consisting of 160 images, with 80 images from the MSRS dataset and the remaining 80 images randomly selected from the M3FD dataset [12]. These 160 images contained a variety of scenes in the city, and we marked the most common objects among them, namely people and cars, as the objects to detect.
We assessed the detection performance of various methods using the mean average precision (mAP) metric. The results, indicating the mAP values at different IoU thresholds, are presented in Table 2. In addition, we calculated different mAPs, as in [53]. The prominent thermal target structure of the IR image helps the detector detect the human body, and the VI image can provide rich vehicle semantic information, so the detector can better realize the detection of vehicles. By fusing the two modalities of IR and VI images, performance in detecting both people and vehicles is enhanced. However, from the results, many algorithms tended to weaken the strength of salient objects, such as SwinFusion and SDNet, so their performance in detecting people was much lower than the recognition of source images. Taken together, our method was the best for person and car detection under almost all IoU thresholds.
At the same time, we provide the detection results for visual display. As shown in Figure 10, in the scene in the 00479D image, due to insufficient illumination in the VI image, the DenseFuse, IFCNN, and SwinFusion methods could not detect the person in the image, and a similar situation also occurred in the scene in 01348N. In the scene in 01348N, although most of the methods successfully detected people and cars in the scene, the confidence levels were very different. In Figure 11, the situation is similar to that in Figure 10. Most methods could not identify people or cars due to the influence of illumination or distance factors. Only our method completely recognized people and cars in the scene and maintained a high level of confidence. This fully shows that the images generated by DPACFuse can provide rich semantic information for the object detector.

4.5.2. Semantic Segmentation Performance

We performed semantic segmentation on the MSRS dataset. Specifically, we utilized source images and different fused images to train semantic segmentation networks [54], respectively. For more details on the semantic segmentation network, please refer to [21]. At the same time, we evaluated model effectiveness using Intersection-over-Union (IoU). The segmentation of each object is shown in Table 3. The results clearly demonstrate that DPACFuse achieved the highest performance in all categories of IoU. This outcome strongly indicates that DPACFuse effectively integrates information from IR and VI images, thereby improving the model’s ability in boundary perception, which leads to more accurate segmentation results.
Furthermore, we also provide the segmentation results for visual presentation. As observed in Figure 12, the IR image exhibited good segmentation performance for persons, but it performed poorly in segmenting other objects, such as color cones and curves. In addition, insufficient illumination negatively affected the segmentation performance of the VI images. From the scenes in the two images in the figure, it is evident that our DPACFuse had an excellent effect on both the segmentation of people and the segmentation of other objects, which shows that the images generated by DPACFuse can better promote semantic segmentation.

5. Conclusions

In this paper, we propose a dual-branch progressive fusion framework, named DPACFuse, to be used for infrared and visible image fusion. Firstly, the Cross-Modality Feature Extraction we designed extracts inter-modality shared features as well as suppresses the loss of high-frequency information. Second, with the help of the ACmix module, our architecture more fully extracts the information in the source images for fusion. Finally, extensive experiments on three publicly available datasets showed that our DPACFuse outperforms all current state-of-the-art methods. In addition, in order to evaluate our approach more comprehensively, we also conducted experiments in two downstream tasks, object detection and semantic segmentation, and the results of the experiments further demonstrated the effectiveness and superiority of our approach.

Author Contributions

Investigation, X.W. and D.H.; Methodology, H.Z.; Visualization, H.Z. and H.W.; Writing—review & editing, H.Z., Z.L. and X.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China (Grant No. 62002082), Guangxi Natural Science Foundation (Grant No. 2020GXNSFBA238014), and the university student innovation training program project (No. 202210595023).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors confirm that the data supporting the findings of this study are available within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Deep learning-based image fusion: A survey. J. Image Graph. 2023, 28, 3–36. [Google Scholar]
  2. Wang, J.; Liu, A.; Yin, Z.; Liu, S.; Tang, S.; Liu, X. Dual Attention Suppression Attack: Generate Adversarial Camouflage in Physical World. arXiv 2021, arXiv:2103.01050. [Google Scholar]
  3. Liu, A.; Liu, X.; Yu, H.; Zhang, C.; Liu, Q.; Tao, D. Training Robust Deep Neural Networks via Adversarial Noise Propagation. IEEE Trans. Image Process. 2021, 30, 5769–5781. [Google Scholar] [CrossRef]
  4. Zeng, Y.; Zhang, D.; Wang, C.; Miao, Z.; Liu, T.; Zhan, X.; Hao, D.; Ma, C. LIFT: Learning 4D LiDAR Image Fusion Transformer for 3D Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 17172–17181. [Google Scholar]
  5. Pan, X.; Cheng, J.; Hou, F.; Lan, R.; Lu, C.; Li, L.; Feng, Z.; Wang, H.; Liang, C.; Liu, Z.; et al. SMILE: Cost-sensitive multi-task learning for nuclear segmentation and classification with imbalanced annotations. Med. Image Anal. 2023, 88, 102867. [Google Scholar] [CrossRef]
  6. Jin, C.; Luo, C.; Yan, M.; Zhao, G.; Zhang, G.; Zhang, S. Weakening the Dominant Role of Text: CMOSI Dataset and Multimodal Semantic Enhancement Network. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–15. [Google Scholar] [CrossRef]
  7. Qin, H.; Ding, Y.; Zhang, M.; Yan, Q.; Liu, A.; Dang, Q.; Liu, Z.; Liu, X. BiBERT: Accurate Fully Binarized BERT. arXiv 2022, arXiv:2203.06390. [Google Scholar]
  8. Qin, H.; Zhang, X.; Gong, R.; Ding, Y.; Xu, Y.; Liu, X. Distribution-sensitive Information Retention for Accurate Binary Neural Network. arXiv 2022, arXiv:2109.12338. [Google Scholar] [CrossRef]
  9. Yan, M.; Lou, X.; Chan, C.A.; Wang, Y.; Jiang, W. A semantic and emotion-based dual latent variable generation model for a dialogue system. Caai Trans. Intell. Technol. 2023, 8, 319–330. [Google Scholar] [CrossRef]
  10. Wang, Z.; Feng, J.; Zhang, Y. Pedestrian detection in infrared image based on depth transfer learning. Multimed. Tools Appl. 2022, 81, 39655–39674. [Google Scholar] [CrossRef]
  11. Zhang, J.; Liu, C.; Wang, B.; Chen, C.; He, J.; Zhou, Y.; Li, J. An infrared pedestrian detection method based on segmentation and domain adaptation learning. Comput. Electr. Eng. 2022, 99, 107781. [Google Scholar] [CrossRef]
  12. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware Dual Adversarial Learning and a Multi-scenario Multi-Modality Benchmark to Fuse Infrared and Visible for Object. arXiv 2022, arXiv:2203.16220. [Google Scholar]
  13. Ma, J.; Tang, L.; Fan, F.; Huang, J.; Mei, X.; Ma, Y. SwinFusion: Cross-domain Long-range Learning for General Image Fusion via Swin Transformer. IEEE/CAA J. Autom. Sin. 2022, 9, 1200–1217. [Google Scholar] [CrossRef]
  14. Tang, L.; Deng, Y.; Ma, Y.; Huang, J.; Ma, J. SuperFusion: A Versatile Image Registration and Fusion Network with Semantic Awareness. IEEE/CAA J. Autom. Sin. 2022, 9, 2121–2137. [Google Scholar] [CrossRef]
  15. Wang, Z.; Chen, Y.; Shao, W.; Li, H.; Zhang, L. SwinFuse: A Residual Swin Transformer Fusion Network for Infrared and Visible Images. arXiv 2022, arXiv:2204.11436. [Google Scholar] [CrossRef]
  16. Zhang, H.; Ma, J. SDNet: A Versatile Squeeze-and-Decomposition Network for Real-Time Image Fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
  17. Rao, Y.; Wu, D.; Han, M.; Wang, T.; Yang, Y.; Lei, T.; Zhou, C.; Bai, H.; Xing, L. AT-GAN: A generative adversarial network with attention and transition for infrared and visible image fusion. Inf. Fusion 2023, 92, 336–349. [Google Scholar] [CrossRef]
  18. Li, H.; Wu, X.J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  19. Li, H.; Wu, X.J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  20. Li, H.; Wu, X.J.; Durrani, T. NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  21. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  23. Ma, J.; Tang, L.; Xu, M.; Zhang, H.; Xiao, G. STDFusionNet: An Infrared and Visible Image Fusion Network Based on Salient Target Detection. IEEE Trans. Instrum. Meas. 2021, 70, 5009513. [Google Scholar] [CrossRef]
  24. Zhao, Z.; Xu, S.; Zhang, J.; Liang, C.; Zhang, C.; Liu, J. Efficient and Model-Based Infrared and Visible Image Fusion via Algorithm Unrolling. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 1186–1196. [Google Scholar] [CrossRef]
  25. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Gool, L.V.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. arXiv 2021, arXiv:2108.10257. [Google Scholar]
  26. Qu, L.; Liu, S.; Wang, M.; Song, Z. TransMEF: A Transformer-Based Multi-Exposure Image Fusion Framework using Self-Supervised Multi-Task Learning. arXiv 2021, arXiv:2112.01030. [Google Scholar] [CrossRef]
  27. Pan, X.; Ge, C.; Lu, R.; Song, S.; Chen, G.; Huang, Z.; Huang, G. On the Integration of Self-Attention and Convolution. arXiv 2022, arXiv:2111.14556. [Google Scholar]
  28. Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 2022, 83–84, 79–92. [Google Scholar] [CrossRef]
  29. Wang, Z.; Wu, Y.; Wang, J.; Xu, J.; Shao, W. Res2Fusion: Infrared and Visible Image Fusion Based on Dense Res2net and Double Nonlocal Attention Models. IEEE Trans. Instrum. Meas. 2022, 71, 5005012. [Google Scholar] [CrossRef]
  30. Zhang, H.; Xu, H.; Xiao, Y.; Guo, X.; Ma, J. Rethinking the Image Fusion: A Fast Unified Image Fusion Network based on Proportional Maintenance of Gradient and Intensity. Proc. AAAI Conf. Artif. Intell. 2020, 34, 12797–12804. [Google Scholar] [CrossRef]
  31. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 502–518. [Google Scholar] [CrossRef]
  32. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  33. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef] [PubMed]
  34. Park, S.; Choi, D.H.; Kim, J.U.; Ro, Y.M. Robust thermal infrared pedestrian detection by associating visible pedestrian knowledge. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 4468–4472. [Google Scholar]
  35. Wu, X.; Hong, D.; Chanussot, J. UIU-Net: U-Net in U-Net for Infrared Small Object Detection. IEEE Trans. Image Process. 2023, 32, 364–376. [Google Scholar] [CrossRef]
  36. Wang, A.; Li, W.; Wu, X.; Huang, Z.; Tao, R. Mpanet: Multi-Patch Attention for Infrared Small Target Object Detection. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 3095–3098. [Google Scholar] [CrossRef]
  37. Sun, Y.; Cao, B.; Zhu, P.; Hu, Q. DetFusion: A Detection-Driven Infrared and Visible Image Fusion Network. In Proceedings of the MM’22: 30th ACM International Conference on Multimedia, New York, NY, USA, 4–7 July 2022; pp. 4003–4011. [Google Scholar] [CrossRef]
  38. Zhao, W.; Xie, S.; Zhao, F.; He, Y.; Lu, H. MetaFusion: Infrared and Visible Image Fusion via Meta-Feature Embedding From Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 13955–13965. [Google Scholar]
  39. Wang, D.; Liu, J.; Liu, R.; Fan, X. An interactively reinforced paradigm for joint infrared-visible image fusion and saliency object detection. Inf. Fusion 2023, 98, 101828. [Google Scholar] [CrossRef]
  40. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  41. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2021, arXiv:2010.04159. [Google Scholar]
  42. Zhou, M.; Yan, K.; Huang, J.; Yang, Z.; Fu, X.; Zhao, F. Mutual Information-Driven Pan-Sharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1798–1808. [Google Scholar]
  43. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction Without Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  44. Zhao, H.; Nie, R. DNDT: Infrared and Visible Image Fusion Via DenseNet and Dual-Transformer. In Proceedings of the 2021 International Conference on Information Technology and Biomedical Engineering (ICITBE), Nanchang, China, 24–26 December 2021; pp. 71–75. [Google Scholar] [CrossRef]
  45. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. arXiv 2021, arXiv:2103.14030. [Google Scholar]
  46. Rao, D.; Wu, X.; Xu, T. TGFuse: An Infrared and Visible Image Fusion Approach Based on Transformer and Generative Adversarial Network. arXiv 2022, arXiv:2201.10147. [Google Scholar] [CrossRef]
  47. Li, J.; Zhu, J.; Li, C.; Chen, X.; Yang, B. CGTF: Convolution-Guided Transformer for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2022, 71, 5012314. [Google Scholar] [CrossRef]
  48. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  49. Toet, A. TNO Image Fusion Dataset. Figshare Data. 2014. Available online: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029 (accessed on 1 January 2023).
  50. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  51. Ma, J.; Zhang, H.; Shao, Z.; Liang, P.; Xu, H. GANMcC: A Generative Adversarial Network with Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2021, 70, 5005014. [Google Scholar] [CrossRef]
  52. Reis, D.; Kupec, J.; Hong, J.; Daoudi, A. Real-Time Flying Object Detection with YOLOv8. arXiv 2023, arXiv:2305.09972. [Google Scholar]
  53. Padilla, R.; Passos, W.L.; Dias, T.L.B.; Netto, S.L.; da Silva, E.A.B. A Comparative Analysis of Object Detection Metrics with a Companion Open-Source Toolkit. Electronics 2021, 10, 279. [Google Scholar] [CrossRef]
  54. Peng, C.; Tian, T.; Chen, C.; Guo, X.; Ma, J. Bilateral attention decoder: A lightweight decoder for real-time semantic segmentation. Neural Netw. 2021, 137, 188–199. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Overview of the IVF model architecture. The feature extractor of the network consists of ACmix, CMEF, and 3 × 3 convolutional layers. Its feature reconstructor consists of 5 convolutional layers. (b) Cross-Modality Feature Extraction (CMFE). This mainly consists of a high-frequency convolution calculation based on Sobel operator (HCGD) and Squeeze-and-Excitation Network (SE). The HCGD adopts the idea of residual connection, and GAP and Linear in the SE schematic represent Global Average Pooling and Linear Function, respectively. (c) Specific framework of ACmix. This can be divided into three parts: Projection reconstruction, Extract local features, and Extract global features.
Figure 1. (a) Overview of the IVF model architecture. The feature extractor of the network consists of ACmix, CMEF, and 3 × 3 convolutional layers. Its feature reconstructor consists of 5 convolutional layers. (b) Cross-Modality Feature Extraction (CMFE). This mainly consists of a high-frequency convolution calculation based on Sobel operator (HCGD) and Squeeze-and-Excitation Network (SE). The HCGD adopts the idea of residual connection, and GAP and Linear in the SE schematic represent Global Average Pooling and Linear Function, respectively. (c) Specific framework of ACmix. This can be divided into three parts: Projection reconstruction, Extract local features, and Extract global features.
Sensors 23 07205 g001
Figure 2. Qualitative analysis of DPACFuse with seven methods on 00634N (top) and 01356N (bottom) images from the MSRS dataset.
Figure 2. Qualitative analysis of DPACFuse with seven methods on 00634N (top) and 01356N (bottom) images from the MSRS dataset.
Sensors 23 07205 g002
Figure 3. Qualitative analysis of DPACFuse with seven methods on 00537D (top) and 01012N (bottom) images from the MSRS dataset.
Figure 3. Qualitative analysis of DPACFuse with seven methods on 00537D (top) and 01012N (bottom) images from the MSRS dataset.
Sensors 23 07205 g003
Figure 4. Quantitative comparison of eight methods on the MSRS test set. The x-axis represents cumulative distribution and the y-axis represents the values of the metric. The point ( x , y ) on the curve represents the measurement value of the x × 100 percent of image pairs not exceeding the value of y. The average value is shown in the legend.
Figure 4. Quantitative comparison of eight methods on the MSRS test set. The x-axis represents cumulative distribution and the y-axis represents the values of the metric. The point ( x , y ) on the curve represents the measurement value of the x × 100 percent of image pairs not exceeding the value of y. The average value is shown in the legend.
Sensors 23 07205 g004
Figure 5. Qualitative analysis of DPACFuse with seven methods on FLIR_06307 (top) and FLIR_03952 (bottom) images from the RoadScene dataset.
Figure 5. Qualitative analysis of DPACFuse with seven methods on FLIR_06307 (top) and FLIR_03952 (bottom) images from the RoadScene dataset.
Sensors 23 07205 g005
Figure 6. Quantitative comparison of eight methods on the RoadScene test set. The x-axis represents image pairs and the y-axis represents the values of the metric. The point ( x , y ) in the image represents the measurement y for the xth pair of images. The average value is shown in the legend.
Figure 6. Quantitative comparison of eight methods on the RoadScene test set. The x-axis represents image pairs and the y-axis represents the values of the metric. The point ( x , y ) in the image represents the measurement y for the xth pair of images. The average value is shown in the legend.
Sensors 23 07205 g006
Figure 7. Qualitative analysis of DPACFuse with seven methods on two representative images from the TNO dataset.
Figure 7. Qualitative analysis of DPACFuse with seven methods on two representative images from the TNO dataset.
Sensors 23 07205 g007
Figure 8. Quantitative comparison of eight methods on the TNO test set. The x-axis represents image pairs and the y-axis represents the values of the metric. The point ( x , y ) in the image represents the measurement y for the xth pair of images. The average value is shown in the legend.
Figure 8. Quantitative comparison of eight methods on the TNO test set. The x-axis represents image pairs and the y-axis represents the values of the metric. The point ( x , y ) in the image represents the measurement y for the xth pair of images. The average value is shown in the legend.
Sensors 23 07205 g008
Figure 9. The results of ablation experiments.
Figure 9. The results of ablation experiments.
Sensors 23 07205 g009
Figure 10. Object detection results on the MSRS dataset. The results are provided for two scenes from 00479D (top) and 01348N (bottom), respectively.
Figure 10. Object detection results on the MSRS dataset. The results are provided for two scenes from 00479D (top) and 01348N (bottom), respectively.
Sensors 23 07205 g010
Figure 11. Object detection results on the M3FD dataset. The results are provided for two scenes from 01136 (top) and 00390 (bottom), respectively.
Figure 11. Object detection results on the M3FD dataset. The results are provided for two scenes from 01136 (top) and 00390 (bottom), respectively.
Sensors 23 07205 g011
Figure 12. Visualization results of semantic segmentation on the MSRS dataset. The two scenes from top to bottom are from: 00055D and 00504N.
Figure 12. Visualization results of semantic segmentation on the MSRS dataset. The two scenes from top to bottom are from: 00055D and 00504N.
Sensors 23 07205 g012
Table 1. Quantitative results of six indices under ablation experiments. In the evaluation results, the best-performing method is highlighted in red. The second-best result is represented in blue.
Table 1. Quantitative results of six indices under ablation experiments. In the evaluation results, the best-performing method is highlighted in red. The second-best result is represented in blue.
M1M2M3M4Our
EN6.61256.61466.64376.60176.6571
Qabf0.69120.68200.66350.65070.7005
SF11.266011.120210.987511.045711.3704
MI4.55404.65254.49374.24394.7669
AG3.61813.58093.33743.54193.7201
FMI_pixel0.92890.92910.92770.92620.9314
Table 2. Object detection performance (mAP) of source images and the fused images of different fusion methods. In the evaluation results, the best performance method is highlighted in red. The second-best result is represented in blue.
Table 2. Object detection performance (mAP) of source images and the fused images of different fusion methods. In the evaluation results, the best performance method is highlighted in red. The second-best result is represented in blue.
[email protected][email protected][email protected]
PersonCarAllPersonCarAllPersonCarAll
IR0.78910.54910.66910.73940.47870.60910.19610.19460.1953
VI0.54780.76600.65690.38730.71080.54910.03590.34620.1911
DenseFuse [18]0.77310.80790.79050.72960.77130.75050.15780.45830.3081
IFCNN [22]0.78620.77680.78150.73160.72460.72810.16200.41180.2869
U2Fusion [31]0.78230.79500.79370.73470.77240.75360.15990.40530.2826
SDNet [16]0.75230.76490.75860.65190.73150.69170.10430.38260.2434
GANMcC [51]0.76570.81320.78950.72500.76580.74540.18340.44150.3129
SwinFusion [13]0.76990.79800.78400.69690.74990.72340.11830.38670.2525
TarDAL [12]0.78970.77460.78220.70830.72460.71650.16460.40700.2858
DPACFuse0.80340.82100.81220.74620.77530.76070.18620.43120.3087
Table 3. mIoU(%) values for segmentation semantics for different images on the MSRS dataset. In the evaluation results, the best performance method is highlighted in red. The second-best result is represented in blue.
Table 3. mIoU(%) values for segmentation semantics for different images on the MSRS dataset. In the evaluation results, the best performance method is highlighted in red. The second-best result is represented in blue.
MethodBackgroundCarPersonBikeCurveCar StopCuardrailColor ToneBumpmIoU
IR97.9685.6971.2765.4652.5254.4827.5954.9260.1163.33
VI97.6983.0754.6766.2751.9754.9459.6950.3566.1864.98
DenseFuse [18]98.3689.4873.4769.8457.7663.5665.0762.3466.0071.76
IFCNN [22]98.3889.5472.2570.1556.8564.0854.4363.3571.9271.22
U2Fusion [31]98.2788.0873.4269.3957.8562.7653.0359.769.7569.75
SDNet [16]98.3589.3974.3169.3157.5661.6349.5360.7671.4670.26
GANMcC [51]98.3488.8573.6869.6756.7565.1757.0661.5071.7271.41
SwinFusion [13]98.2588.0870.7468.6674.3161.7067.3464.0667.8671.67
TarDAL [12]98.389.1172.6768.9657.0062.3452.4360.7961.4169.22
DPACFuse98.690.3874.5871.9465.3974.4484.6666.0377.1878.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, H.; Wu, H.; Wang, X.; He, D.; Liu, Z.; Pan, X. DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution. Sensors 2023, 23, 7205. https://doi.org/10.3390/s23167205

AMA Style

Zhu H, Wu H, Wang X, He D, Liu Z, Pan X. DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution. Sensors. 2023; 23(16):7205. https://doi.org/10.3390/s23167205

Chicago/Turabian Style

Zhu, Huayi, Heshan Wu, Xiaolong Wang, Dongmei He, Zhenbing Liu, and Xipeng Pan. 2023. "DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution" Sensors 23, no. 16: 7205. https://doi.org/10.3390/s23167205

APA Style

Zhu, H., Wu, H., Wang, X., He, D., Liu, Z., & Pan, X. (2023). DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution. Sensors, 23(16), 7205. https://doi.org/10.3390/s23167205

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop