Next Article in Journal
Development of a Utility Model for the Measurement of Global Radiation in Photovoltaic Applications in the Internet of Things (IoT)
Next Article in Special Issue
Narrowband Interference Separation for Synthetic Aperture Radar via Sensing Matrix Optimization-Based Block Sparse Bayesian Learning
Previous Article in Journal
Scalable Single-Phase Multi-Functional Inverter for Integration of Rooftop Solar-PV to Low-Voltage Ideal and Weak Utility Grid
Previous Article in Special Issue
Coherent Integration for Radar High-Speed Maneuvering Target Based on Frequency-Domain Second-Order Phase Difference
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering

1
Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
2
Beijing Key Laboratory of Advanced Information Science and Network Technology, Beijing Jiaotong University, Beijing 100044, China
3
College of Electronic and Information Engineering, Hebei University, Baoding 071002, China
4
Research Institute of TV and Electro-Acoustics, Beijing 100015, China
*
Authors to whom correspondence should be addressed.
Electronics 2019, 8(3), 303; https://doi.org/10.3390/electronics8030303
Submission received: 18 January 2019 / Revised: 1 March 2019 / Accepted: 5 March 2019 / Published: 8 March 2019

Abstract

:
In this paper, a remote sensing image fusion method is presented since sparse representation (SR) has been widely used in image processing, especially for image fusion. Firstly, we used source images to learn the adaptive dictionary, and sparse coefficients were obtained by sparsely coding the source images with the adaptive dictionary. Then, with the help of improved hyperbolic tangent function (tanh) and l 0 max , we fused these sparse coefficients together. The initial fused image can be obtained by the image fusion method based on SR. To take full advantage of the spatial information of the source images, the fused image based on the spatial domain (SF) was obtained at the same time. Lastly, the final fused image could be reconstructed by guided filtering of the fused image based on SR and SF. Experimental results show that the proposed method outperforms some state-of-the-art methods on visual and quantitative evaluations.

1. Introduction

By making full use of the complementary information of the remote sensing images and other source images of the same scene, image fusion can be defined as the processing method for integrating this information together to obtain a fused image, which is more suitable for the human visual system [1]. Through image fusion, we can obtain one composite image, which contains more special features, and can provide more useful information. As a powerful tool for image processing, image fusion covers broad range of areas [2,3], such as computer vision, remote sensing, and so on [4].
Diversiform remote sensing image fusion methods have been proposed in recent years, which can be divided into three categories: Pixel-level fusion, feature-level fusion, and decision-level fusion [5]. Feature-level fusion mainly deals with the features of the source images, while decision-level fusion makes the decision after judging the information of the source images. Compared with the aforementioned levels, pixel-level fusion can serve more useful original information, although it has some shortcomings such as being time consuming. Despite complex computation, most researchers conduct image fusion based on pixel-fusion [6,7], such as the image fusion method based on the spatial domain, and the image fusion method based on the transform domain.
Recently, mainstream methods of image fusion have been based on the multi-scale transforms [8,9], such as image fusion based on object region detection and non-subsampled contourlet transform [10] and image fusion based on the complex shearlet transform with guided filtering [11]. For the image fusion method based on multi-scale transforms, the source images are represented by the fixed orthogonal basis functions, and the fused image can be obtained by fusing the coefficients of different sub-bands together in the transform domain. Although the multi-scale geometric transform can represent most features of the image, which are always complex and diverse, there are some features that cannot be represented sparsely. Thus, it cannot represent all the useful features accurately by limited fixed transforms.
The rapidly developing sparse representation methods can not only more sparsely represent the source images, but also effectively extract the potential information hidden in the source images and produce more accurate fused images, compared with the multi-scale transforms [12,13,14]. Based on these findings, scholars apply sparse representation to image fusion. Mitianoudis [13] and Yang [14] laid the foundation for image fusion based on SR. Yu [15] applied sparse representation with K-singular value decomposition (K-SVD) to medical image fusion, Yang [16] applied sparse representation and multi-scale decomposition to remote sensing image fusion, and Yin [17] applied a novel sparse-representation-based method to multi-focus image fusion.
In the sparse model, the generation of the dictionary and sparse coding is crucial for the image fusion [18]. Although the fixed over-complete dictionary can realize good fusion results, it usually takes a lot of time to obtain the sparse coefficients, resulting in inefficiency. In this paper, adaptive dictionary learning [19,20] is adopted for its simplicity and convenience. Motivated by the multi-strategy fusion rule based sigmoid function in reference [21] and the characteristics of the hyperbolic tangent function, the multifarious rule based on tanh and l 0 max is proposed to fuse the sparse coefficients. Finally, by sparse reconstruction, the fused image based on SR is obtained, which is more suitable for the human visual system and subsequent image processing. However, there is more detailed information in the remote sensing images than other kinds of images. When performing image fusion by the method based on SR, it may lose some discontinuous edge features [22], which leads to the loss of some useful information of fused images. In addition, image fusion based on SR also ignores the spatial information, which can reflect the image structure more directly and accurately. As a result, we can simultaneously fuse the source remote sensing images by the method based on SR and SF, and obtain two different fused images, namely the fused image based on SR and the fused image based on SF. In this paper the two fused images above are processed by a guided filter to obtain the final image since a guided filter has good performance with edge preserving [23]. The main contributions of this paper can be summarized as follows.
(1) The learning of the dictionary is vital for sparse representation, and the adaptive dictionary of each source image can be generated in every step of dictionary learning. The final dictionary can be obtained by gathering together the sub-dictionaries. As a result, this work enriches the dictionary and can make the coefficients more sparsely.
(2) As is well known, the information in each source image is complementary and redundant. When fusing images to obtain the fused image, we need to consider the relationship between different source images. For the redundant information of the source images, the weighted rule would be better; on the other hand, the choose-max rule would result in a fused image with less block effect. Based on the above considerations and the characteristics of hyperbolic tangent function, the fusion rule based on tanh and l 0 max is proposed in this paper.
(3) The image fusion methods based on SR can obtain the fused image by sparsely coding the source images and fusing the sparse coefficients. However, it ignores the correlation of the image information in the spatial domain and loses some important detailed information of the source images. In this paper, we adopt the image fusion method based on SF and filter the fused image based on SR and SF by the guided filter. By making full use of the information in the spatial and the sparse representation domain, the fused image can reserve more information of the source images.
The rest of this paper is organized as follows. The theory of the sparse representation is introduced briefly in Section 2. Adaptive dictionary learning is presented in Section 2.1, and the proposed fusion rule is given in Section 2.2. The flow chart of the remote sensing image fusion method based on SR and guided filtering is drawn in Section 3. In Section 4, some experiments and result analysis are done. Finally, conclusions are made in Section 5.

2. Sparse Representation

SR has been widely used in image processing, as one of the most powerful tools to represent signals especially image signals, such as image de-noising [24], image coding [25], object tracking [26], and image super resolution [27], etc.
In the SR model, the image is sparse and can be represented, or approximately represented, by one linear combination of a few atoms from the dictionary [14,28,29]. Suppose that the source image is I , and the over-complete dictionary is D R M × k , the sparse representation model can be formulated as follows [16,22].
α ^ = arg min α α 0 s . t . I D α 2 2 ε
where α denotes the sparse coefficients of the image and 0 denotes the l 0 n o r m , respectively, which indicate the number of non-zero elements in the corresponding vector. Usually, α 0 L < < M , and L is the maximal sparsity. ε indicates the limiting error.
For the image fusion method based on SR, there are two important steps: dictionary learning and sparse coding. Dictionary learning will be discussed in detail in Section 2.1. When performing sparse coding by orthogonal matching pursuit (OMP) [30] in this paper, Equation (1) can be replaced by Equation (2).
α ^ = arg min α I D α 2 2 + μ α 0
where, μ is the penalty factor.

2.1. Adaptive Dictionary Learning

When fuse the source images by the methods based on SR, dictionary learning is one of important processes. To make full use of the image information, we generate a dictionary based on the source images themselves. And the generation of the adaptive dictionary can be changed into the iteration of the dictionary atoms. By the iteration process, it can realize dictionary learning with the over-complete dictionary based on the source images.
Since dictionary learning is more efficient for small image blocks, if the dictionary updating step is processed by the original source images directly, the sparsity would be seriously influenced. Thus optimal sparse coefficients cannot be obtained [29]. In order to solve this problem, we divided the source images into image blocks, which can replace the dictionary atoms for better dictionary learning. The improved dictionary generation method can not only obtain the optimal sparse representation but also accelerates the efficiency and accuracy of the SR algorithm. However, since we perform dictionary learning on the image block rather than the whole image, the reshaped vector on every atom is not very large and it reduces the computation cost.
K-singular value decomposition (K-SVD) [31] is one of the most used image fusion methods based on SR. Here, we adopt the K-SVD model on the sub-dictionary of the image block by the following iteration process:
D ^ i j M = arg min D ^ i j M , α i j M i , j P i j M D ^ i j M α i j M 2 2 + μ i j M α i j M 0
where i j denotes the position ( i , j ) in the image M and P i j M denotes the image block with the center pixel at the corresponding position ( i , j ) .
Then, we can obtain the adaptive dictionary of the source image M shown in Equation (4).
D ^ M = { D ^ i j M }
At last, we can gather all the dictionaries of different source images by Equation (5), where n denotes the total number of the source images.
D = [ D 1 , D 2 , , D n ]

2.2. Fusion Rule Based on tanh and l 0 max

As we all know, the fusion rules are vital for the final fusion results and for the sparse coefficients. In most cases, we always take the l 1 max rule to obtain the fused block vectors [7], where l 1 means the sum of absolute values of the vector elements. However, when there are noises or some unwanted pixels in the flat area of the source images, the unwanted portion will be included and lead to incorrect fusion [17]. The information in the source images is redundant and complementary for the image fusion shown in Figure 1. Figure 1a,b are one set of medical images, which contain complementary information, while Figure 1c,d are one set of multi-focus images, which contain redundant information. When the relationship of the image information is redundant, the weighted fusion rule is chosen, and the max fusion rule should be chosen for the complementary sparse coefficients [21]. The fused information would be lost and incomplete if the complementary information is multiplicative by the weighted factor. Based on these considerations, we proposed one new sparse coefficient fusion rule based on tanh and l 0 max . We can obtain the fused coefficients by calculating l 0 n o r m and the weighting factor based on tanh.
The hyperbolic tangent function is one of the hyperbolic functions, and derives from hyperbolic sine function and hyperbolic cosine function [32]. It can be calculated as follows:
tanh ( x ) = sinh ( x ) cosh ( x ) = e x e x e x + e x
where the hyperbolic sine function and hyperbolic cosine function can be defined as Equations (7) and (8), respectively.
sinh ( x ) = e x e x 2
cosh ( x ) = e x + e x 2
Figure 2 shows the different hyperbolic functions. From Figure 2a,b we can see that tanh is symmetrical around the origin point. As x increases, the difference between the value of the hyperbolic sine function and the hyperbolic cosine function narrows, and the value of tanh ( x ) changes from −1 to 1. When there is redundant information in different source images and the weighted fusion rule is chosen, it would be better if different degrees of redundancy corresponded to different weights. Based on the aforementioned factors, we improve tanh shown in Figure 2c to obtain the weighted factor for fusing the sparse coefficients, and the corresponding equation is listed as Equation (9).
w i j = 1 2 [ tanh ( a ( s i j 1 ) ) + 1 ]
where s i j denotes the sparse coefficient at the position ( i , j ) and w i j denotes the corresponding weighted factor when adopting the fusion rule based on tanh. a denotes the sensitivity between the sparse coefficient and the weighted factor. According to the experiments on different image groups and values of the parameter a , we found that 3 is the best.
Compared with Figure 2b, the curve has a steeper slope in Figure 2c when s i j is closer to 1, which means that the weighted factor is very sensitive to the sparse coefficients. When s i j is near 0 or too large, the weighted factor w i j is near 0 or 1, which means that the source images have complementary information, where the fusion rule based on l 0 max is adopted.
Finally, we can obtain the fused sparse coefficients α F i j at the position ( i , j ) by Equation (10).
α F i j = { w i j α A i j + ( 1 w i j ) α B i j i f   α A i j & α B i j 0 max ( α A i j , α B i j ) e l s e
where α A i j and α B i j denotes the sparse coefficients in the source image A and B . α A i j & α B i j 0 means that both α A i j and α B i j are not zero. And w i j can be calculated by Equation (9), where s i j = α A i j .

3. The Proposed Image Fusion Method

An interesting remote sensing fusion method based on sparse representation and guided filtering is presented in this paper, and the framework can be seen in Figure 3. It mainly includes three image processing elements: image fusion based on SR, image fusion based on SF, and guided filtering. The adaptive dictionary was learned by the source images themselves, and the fused sparse coefficients was obtained by the dictionary and proposed fusion rule. Then, the fused image based on SR was reconstructed by the obtained adaptive dictionary and fused sparse coefficients. At the same time, we fused the source images obtained by the image fusion method based on SF such as the gradient fusion. As shown in Figure 3, the guided filter was finally adopted to guide the fused images based on SR and SF. Since there was more detailed information in the fused image based on SF, in the last part of the proposed method, we made the fused image based on SF as the guidance image, and the other fused image served as the input image.

4. The Experiments and Result Analysis

To testify the superiority of the proposed method, a series of experiments on the remote sensing and other source images were conducted in this section. We compared our method with some classical image fusion methods, including the multi-scale weighted gradient-based fusion (MWGF) [33], the image fusion with guided filtering (GuF) [34], image fusion based on Laplace transformation (LP) [35], multiresolution DCT decomposition for image fusion (DCT) [36], the image fusion algorithm in the nonsubsampled contourlet transform domain (NSCT) [37], image fusion with the joint sparsity model (SR) [1], and image fusion based on multi-scale and sparse representation(MST-SR) [8]. With adaptive dictionary learning, the size of every image block was 8 × 8 . Experiments conducted on dictionary learning of different source images showed that when the number of iterations was 3, it guaranteed the convergence and stability of the coefficients. In addition, the experiments in this paper were carried out by Matlab code on an Intel Core i5-2450M (Acer, Beijing, China) 2.50 GHz with 6 GB RAM.

4.1. Objective Valuation Indexes

To evaluate the experimental results more objectively, we adopted some objective valuation indexes [37] to evaluate the fused images by different image fusion methods, which included entropy (EN), spatial frequency (SF), QAB/F, and structural similarity (SSIM).
When we want to balance the wealth of information in one image, EN is a wonderful choice. The larger the value of EN in the fused image is, the more information does the image contain, which means better image fusion result. And EN can be summarized as Equation (11).
E N = i = 0 L 1 p i × log 2 p i
where L denotes the total number of pixels included in the image and p i denotes the probability distribution of pixels for each gray level.
SF can detect the total active of the fused image in the spatial domain and it denotes the expression ability of one image for minor detail contrast. The equation of SF is shown as follows:
S F ( i , j ) = ( R F ) 2 + ( C F ) 2
where R F stands for the horizontal frequency while C F stands for the vertical frequency. And they can be calculated by Equations (13) and (14).
R F = 1 M × N x = 1 M y = 2 N [ F ( x , y ) F ( x , y 1 ) ] 2
C F = 1 M × N x = 2 M y = 1 N [ F ( x , y ) F ( x 1 , y ) ] 2
where F denotes the fused image with the size of M × N .
While QAB/F can balance how much the edge information of the source images A and B does the fused image contain by Sobel operator. It can be defined as Equation (15).
Q A B / F = n , m ( Q n , m A F w n , m A + Q n , m B F w n , m B ) n , m ( w n , m A + w n , m B )
where w n , m A = [ g A ( n , m ) ] L ,   w n , m B = [ g B ( n , m ) ] L . Normally, L is one constant and the value is 1. Taking the source image A as an example, edge information retention value Q n , m A F and edge strength information g A ( n , m ) can be calculated by Equations (16) and (17).
Q n , m A F = Γ g Γ α [ 1 + e K g ( G n , m A F σ g ) ] 1 [ 1 + e K α ( A n , m A F σ α ) ] 1
g A ( n , m ) = s A x ( n , m ) 2 + s A y ( n , m ) 2
where Γ g , K g , σ g , Γ α , K α , σ α are constant and they affect the sigmoid function together. ( G n , m A F , A n , m A F ) = [ ( g n , m F g n , m A ) M ,   1 | α A ( n , m ) α F ( n , m ) | π / 2 ] and M = { 1 i f   g A ( n , m ) g F ( n , m ) 1 o t h e r w i s e . α A ( n , m ) = tan 1 [ s A y ( n , m ) s A x ( n , m ) ] and s A x ( n , m ) ,   s A y ( n , m ) denote the convolution results of Sobel model with the center pixel at the position ( n , m ) in the horizontal and vertical directions with the source image A .
SSIM is the structural similarity between the source images and the fused image. And the equation of SSIM is as follows:
S S I M ( A , B , F ) = 1 2 ( S S I M ( A , F ) + S S I M ( B , F ) )
where S S I M ( A , F ) denotes SSIM of the source image A and fused image F , and so is S S I M ( B , F ) . More detail of their calculation is shown in Equations (19) and (20).
S S I M ( A , F ) = ( 2 μ A μ F + C 1 ) · ( 2 σ A F + C 2 ) ( μ A 2 + μ F 2 + C 1 ) ( σ A 2 + σ F 2 + C 2 )
S S I M ( B , F ) = ( 2 μ B μ F + C 1 ) · ( 2 σ B F + C 2 ) ( μ B 2 + μ F 2 + C 1 ) ( σ B 2 + σ F 2 + C 2 )
where μ A , μ B , μ F denote the average of pixels of the image A , B and F , respectively. σ A 2 , σ B 2 , σ F 2 denote the variance and σ A F , σ B F denote the joint variance. For the convenience of calculation, we make C 1 = C 2 = 0 .
The larger all the indexes above are, the better the fused image is. What’s more, when obtaining the adaptive dictionary by the proposed method, there is slight deviation of the final results. We adopt the mean of the evaluation values in three times.

4.2. Large Scale Image Fusion of Optical and Radar Images

Figure 4 shows one SAR image of the harbor around Oslo with a size of 1131 × 942 and the registered optical image on a large scale for the whole scenery [38]. Due to the use of the high-resolution digital elevation model (DEM), the optical image fits onto the signatures of the buildings very well. Figure 4c,d are partially enlarged details of Figure 4a,b at the position of the red rectangle in Figure 4a. Figure 5 and Figure 6 are the corresponding fused images obtained by the methods above, and partially enlarged views.
Since the optical image in Figure 4b is colorful, we processed the image fusion in the RGB dimension separately. Although the visual effect of Figure 6a is better, there was a greater color contrast in Figure 5a, which introduced some incorrect information in the left corner. In Figure 6, the partially enlarged detail images of Figure 5d by DCT and Figure 5f by SR are very blurred which seriously affects the fused images. Compared with Figure 6g, the left corner in Figure 6h contains more information of the remote sensing image in Figure 5c, which indicates that the fused image by our method is better.
Table 1 shows the corresponding index values of the fused images in Figure 5 and the best values are in bold. From Table 1, we can see the image fusion methods based on the spatial domain such as MWGF and GuF have big ability to preserve the spatial frequency, and MWGF has a better value of QAB/F. However, the visual result of MWGF is the worst. QAB/F of the proposed method ranks third among the compared methods, which is worse than the methods based on the spatial domain. This explains why we adopt the image fusion method based on the spatial domain and guide it with the fused image-based SR in this paper. The values of EN, SF, and SSIM of the fused image obtained by the proposed method are better, which indicates that the proposed method has a better ability to fuse the remote sensing image.

4.3. Image Fusion of Remote Sensing Images

To testify the effectiveness and universality of the proposed method, the classical image pairs shared by Durga Prasad Bavirisetti (https://sites.google.com/view/durgaprasadbavirisetti/datasets) are used to test the performance of the fused algorithms. The dataset contains rich remote sensing images and we conduct our experiments on different kinds of image pairs, which contain the forest with greater high-frequency information, rivers with low-frequency information, and so on. To save space, we only show the four groups and the results analysis. The four groups include rich information with different types and are representative in the dataset, shown in Figure 7. Figure 8, Figure 9, Figure 10 and Figure 11 are the fused images obtained by the diverse compared methods of the different source images.
Figure 7a,b are forests and rural areas with fewer buildings, of which the top view is sharper and has richer detailed information. From Figure 8, we can see that the trees in Figure 8a–e is more darker than Figure 8f–g and has less information in the second line of Group 1 in Figure 7, which indicates that the image fusion based on SR is more powerful than the methods based on the spatial domain and transform domain. And there are some artificial textures in the roof of Figure 8f. Above all, the fused image of Figure 8h obtained by the proposed has better visual effect.
Compared with Group 2, there are some suburbs next to the forests in Group 1. And the contrast in Figure 9c,e,h looks better. From the roofs in the fused images shown in Figure 9, the flat area and edges in Figure 9h obtained by the proposed method look more comfortable and are more suitable for we human visual system, which indicates that the proposed method has powerful ability to fuse remote sensing images.
There are some river and coastal area in Group 3. And by comparing the fused images in Figure 10, the center in Figure 10a looks very bad and some areas in Figure 10g are too bright, which have the strong exposure. From these figures, we can see that there is less artificial texture in Figure 10h, which means the fused image obtained by the proposed method have better visual result.
Group 4 is one set of classic multi-sensor image pair, which can be found in most of papers about remote sensing image fusion. By comparing the bottoms of the fused images in Figure 10, we can find that there are some unwanted spots and artificial texture in Figure 10d, and the small round black area is very blurred or even lost in Figure 10a–c,f. Since the rivers display as black areas like wide line or curve in the fused images, it has worst visual effect in Figure 10f, of which the detailed information has been lost. As a result, the fused image in Figure 10h looks more comfortable for our eyes and the proposed method has better ability to fuse remote sensing images.
Similarly, we use the aforementioned objection evaluation indexes to value the fused images in Figure 8, Figure 9, Figure 10 and Figure 11 and the objective values are shown in Table 2, Table 3, Table 4 and Table 5. As shown in Table 2 and Table 3, the algorithm proposed in this paper has obtained the best results for Group 1 and Group 2 in Figure 7. This fully demonstrates that the proposed method has a better ability to perform remote sensing image fusion. Compared with Group 1 and Group 2, there is more low frequency information and less detail and edges in Group 3 and Group 4. However, the proposed method is more suitable for the images with great detail. As a result, the SSIM of the fused image by NSCT is better than others in Table 4, but other values of the proposed method are satisfactory. All these values demonstrate that the proposed method performs better in terms of remote sensing image fusion.

5. Conclusions

Due to the good performance of sparse representation and the rich information in the spatial domain, this paper presents one new remote sensing image fusion method based on sparse representation and guided filtering. It also makes full use of the redundant and complementary information of different source images. Experimental results show that our method is more suitable for the human visual system and has better objective evaluation index values. However, the proposed image fusion method is very powerful for the details such as image edges. Although remote sensing images have rich detailed information, it would be inefficient if there is much more low frequency information than high frequency information. How to overcome this shortcoming will be investigated in future work.

Author Contributions

X.M. wrote the original draft and she performed the experiments with J.F., S.L. wrote the review & editing. S.H. provided Funding acquisition, and he also provided resources with S.X.

Funding

This research was funded by the Natural Science Foundation of China under Grant No. 61572063 and No. 61401308; Natural Science Foundation of Hebei Province under Grant No. F2016201142 and No. F2016201187; Science Research Project of Hebei Province under under Grant No. QN2016085. Opening Foundation of Machine vision Engineering Research Center of Hebei Province under Grant 2018HBMV02, Science Research Project of Hebei Province under Grant QN2016085, Natural Science Foundation of Hebei University under Grant 2014-303, Post-graduate’s Innovation Fund Project of Hebei University under Grant hbu2018ss01.

Acknowledgments

Some of the images adopted in experiments are downloaded from the website of https://sites.google.com/view/durgaprasadbavirisetti/datasets. This work was supported by the High-Performance Computing Center of Hebei University. We also thank Qu XiaoBo, Zhou Zhiqiang, Li Shutao, and Liu Yu for their shared codes of image fusion. We also thank the Editor and Reviewers for the efforts made in processing this submission and we are particularly grateful to the reviewers for their constructive comments and suggestions which help us improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Z.; Yin, H.; Chai, Y.; Li, Y.; Qi, G. A novel multi-modality image fusion based on image decomposition and sparse representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  2. Anandhi, D.; Valli, S. An algorithm for multi-sensors image fusion using maximum a posteriori and nonsubsampled contourlet transform. Comput. Electr. Eng. 2017, 65, 139–152. [Google Scholar] [CrossRef]
  3. Gao, Z.; Zhang, C. Texture clear multi-modal image fusion with joint sparsity model. Optik 2017, 130, 255–265. [Google Scholar] [CrossRef]
  4. Hu, S.; Yang, D.; Liu, S.; Ma, X. Block-matching based mutimodal medical image fusion via PCNN with SML. In Proceedings of the 2016 IEEE 13th International Conference on Signal Processing (ICSP), Chengdu, China, 6–10 November 2016; pp. 13–18. [Google Scholar]
  5. Li, H.; Chai, Y.; Ling, R.; Yin, H. Multifocus image fusion scheme using feature contrast of orientation information measure in lifting stationary wavelet domain. J. Inf. Sci. Eng. 2013, 29, 227–247. [Google Scholar]
  6. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  7. Zhang, J.; Feng, X.; Song, B.; Li, M.; Lu, Y. Multi-focus image fusion using quality assessment of spatial domain and genetic algorithm. In Proceedings of the Conference on Human System Interactions, Krakow, Poland, 25–27 May 2008; pp. 71–75. [Google Scholar] [CrossRef]
  8. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  9. Li, H.; Qiu, H.; Yu, Z.; Li, B. Multifocus image fusion via fixed window technique of multiscale images and non-local means filtering. Signal Process. 2017, 138, 71–85. [Google Scholar] [CrossRef]
  10. Meng, F.; Song, M.; Guo, B.; Shi, R.; Shan, D. Image fusion based on object detection and non-subsampled contourlet transform. Comput. Electr. Eng. 2017, 62, 375–383. [Google Scholar] [CrossRef]
  11. Liu, S.; Shi, M.; Zhu, Z.; Zhao, J. Image fusion based on complex-shearlet domain with guided filtering. Multidimens. Syst. Signal Process. 2017, 28, 207–224. [Google Scholar] [CrossRef]
  12. Donoho, D.L. Compressed Sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  13. Mitianoudis, N.; Stathaki, T. Pixel-based and region-based image fusion schemes using ICA bases. Inf. Fusion 2007, 8, 131–142. [Google Scholar] [CrossRef] [Green Version]
  14. Yang, B.; Li, S. Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 2010, 59, 884–892. [Google Scholar] [CrossRef]
  15. Yu, N.; Qiu, T.; Liu, W. Medical image fusion based on sparse representation with K-SVD. In Proceedings of the World Congress on Medical Physics and Biomedical Engineering, Beijing, China, 26–31 May 2012; pp. 550–553. [Google Scholar] [CrossRef]
  16. Yang, Y.; Wu, L.; Huang, S.; Wan, W.; Que, Y. Remote sensing image fusion based on adaptively weighted joint detail injection. IEEE Access 2018, 6, 6849–6864. [Google Scholar] [CrossRef]
  17. Yin, H.; Li, Y.; Chai, Y.; Liu, Z.; Zhu, Z. A novel sparse-representation-based multi-foucs image fusion approach. Neurocomputing 2016, 216, 216–229. [Google Scholar] [CrossRef]
  18. Zong, J.; Qiu, T. Medical image fusion based on sparse representation of classified image patches. Biomed. Signal Process. Control 2017, 34, 195–205. [Google Scholar] [CrossRef]
  19. Zhu, Z.; Chai, Y.; Yin, H.; Li, Y.; Liu, Z. A novel dictionary learning approach for multi-modality medical image fusion. Neurocomputing 2016, 214, 471–482. [Google Scholar] [CrossRef]
  20. Elad, M.; Yavneh, I. A plurality of sparse representation is better than the sparsest one alone. IEEE Trans. Inf. Theory 2009, 55, 4701–4714. [Google Scholar] [CrossRef]
  21. Luo, X.; Zhang, Z.; Zhang, C.; Wu, X. Multi-focus image fusion using HOSVD and edge intensity. J. Visual Commun. Image Represent. 2017, 45, 46–61. [Google Scholar] [CrossRef]
  22. Ma, X.; Hu, S.; Liu, S. SAR image de-noising based on invariant K-SVD and guided filter. Remote Sens. 2017, 9, 1311. [Google Scholar] [CrossRef]
  23. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  24. Liu, S.; Liu, M.; Li, P.; Zhao, J.; Zhu, Z.; Wang, X. SAR image de-noising via sparse representation in shearlet domain based on continuous cycle spinning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2985–2992. [Google Scholar] [CrossRef]
  25. Zhang, X.; Sun, J.; Ma, S.; Lin, Z.; Zhang, J.; Wang, S.; Gao, W. Globally variance-constrained sparse representation and its application in image set coding. IEEE Trans. Image Process. 2018, 27, 3753–3765. [Google Scholar] [CrossRef] [PubMed]
  26. Sitani, D.; Subramanyam, A.V.; Majumdar, A. Online single and multiple analysis dictionary learning-based approach for visual object tracking. J. Electron. Imag 2019, 28, 013004. [Google Scholar] [CrossRef]
  27. Peng, L.; Yang, J. Single-image super resolution via hashing classification and sparse representation. In Proceedings of the 2017 3rd IEEE International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2017; pp. 1923–1927. [Google Scholar] [CrossRef]
  28. Vanika, S.; Prerna, K.; Angshul, M. Class-wise deep dictionary learning. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1125–1132. [Google Scholar] [CrossRef]
  29. Zhang, Z.; Xu, Y.; Yang, J.; Li, X.; Zhang, D. A survey of sparse representation: Algorithms and applications. IEEE Access 2015, 3, 490–530. [Google Scholar] [CrossRef]
  30. Pati, Y.C.; Rezaiifar, R.; Krishnaprasad, P.S. Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition. In Proceedings of the 27th Annual Asilomar Conference on Signals Systems and Computers, Pacific Grove, CA, USA, 1–3 November 1993; pp. 40–44. [Google Scholar] [CrossRef]
  31. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2007, 15, 3736–3745. [Google Scholar] [CrossRef]
  32. Wang, G. The inverse hyperbolic tangent function and Jacobian sine function. J. Math. Anal. Appl. 2017, 448, 498–505. [Google Scholar] [CrossRef]
  33. Zhou, Z.; Li, S.; Wang, B. Multi-scale weighted gradient-based fusion for multi-focus images. Inf. Fusion 2014, 20, 60–72. [Google Scholar] [CrossRef]
  34. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [CrossRef]
  35. Zhao, P.; Liu, G.; Hu, C.; Huang, H.; He, B. Medical image fusion algorithm based on the Laplace-PCA. In Proceedings of the 2013 Chinese Intelligent Automation Conference, CIAC 2013, Yangzhou, Jiangsu, China, 23–25 August 2013; Volume 256, pp. 787–794. [Google Scholar] [CrossRef]
  36. Shreyamsha Kumar, B.K.; Swamy, M.N.S.; Ahmad, M.O. Multiresolution DCT decomposition for multifocus image fusion. In Proceedings of the 2013 26th IEEE Canadian Conference on Electrical and Computer Engineering, CCECE 2013, Regina, SK, Canada, 5–8 May 2013; pp. 1–4. [Google Scholar] [CrossRef]
  37. Qu, X.; Yan, J.; Xiao, H.; Zhu, Z. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  38. Available online: http://www.dlr.de/hr/en/desktopdefault.aspx/tabid-2434/3770_read-32515 (accessed on 6 March 2019).
Figure 1. Images with different information: (a) CT image; (b) MRI image; (c) left-focus image; (d) right-focus image.
Figure 1. Images with different information: (a) CT image; (b) MRI image; (c) left-focus image; (d) right-focus image.
Electronics 08 00303 g001
Figure 2. Different hyperbolic functions: (a) The hyperbolic functions; (b) tanh; (c) improved tanh.
Figure 2. Different hyperbolic functions: (a) The hyperbolic functions; (b) tanh; (c) improved tanh.
Electronics 08 00303 g002
Figure 3. The framework of the proposed method.
Figure 3. The framework of the proposed method.
Electronics 08 00303 g003
Figure 4. The large scale images: (a) TerraSAR-X staring spotlight image of Oslo; (b) optical image of Oslo; (c) part of (a); (d) part of (b).
Figure 4. The large scale images: (a) TerraSAR-X staring spotlight image of Oslo; (b) optical image of Oslo; (c) part of (a); (d) part of (b).
Electronics 08 00303 g004
Figure 5. The fused images of Figure 4: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 5. The fused images of Figure 4: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g005
Figure 6. Part of Figure 5: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 6. Part of Figure 5: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g006aElectronics 08 00303 g006b
Figure 7. The source remote sensing images: (a) Group 1; (b) Group 2; (c) Group 3; (d) Group 4.
Figure 7. The source remote sensing images: (a) Group 1; (b) Group 2; (c) Group 3; (d) Group 4.
Electronics 08 00303 g007
Figure 8. The fused images of Group 1: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 8. The fused images of Group 1: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g008
Figure 9. The fused images of Group 2: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 9. The fused images of Group 2: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g009
Figure 10. The fused images of Group 3: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 10. The fused images of Group 3: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g010
Figure 11. The fused images of Group 4: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Figure 11. The fused images of Group 4: (a) MWGF; (b) GuF; (c) LP; (d) DCT; (e) NSCT; (f) SR; (g) MST-SR; (h) the proposed.
Electronics 08 00303 g011
Table 1. The evaluation index values of fused images in Figure 5.
Table 1. The evaluation index values of fused images in Figure 5.
MethodsMWGFGuFLPDCTNSCTSRMST-SRThe Proposed
Indexes
EN7.34117.25477.26067.21587.18437.27117.37427.5879
SF31.183430.166232.468230.122331.588430.267432.487833.9360
QAB/F0.63130.61110.57530.37540.54530.57050.57670.5794
SSIM0.57980.59810.59900.54100.59100.58290.59660.6246
Table 2. The evaluation index values of fused images in Figure 8.
Table 2. The evaluation index values of fused images in Figure 8.
MethodsMWGFGuFLPDCTNSCTSRMST-SRThe Proposed
Indexes
EN7.17417.49317.66877.60067.69357.43517.59257.7076
SF54.183953.672855.000655.243554.656753.986854.869756.3409
QAB/F0.70300.71040.71100.61050.71430.70890.70730.7174
SSIM0.79330.79760.80770.77590.81700.79160.80600.8235
Table 3. The evaluation index values of fused images in Figure 9.
Table 3. The evaluation index values of fused images in Figure 9.
MethodsMWGFGuFLPDCTNSCTSRMST-SRThe Proposed
Indexes
EN6.97787.66687.89367.55377.81547.36347.65807.8946
SF53.684453.337953.814953.250753.321653.175754.124754.1590
QAB/F0.66680.66870.63100.47830.61900.63790.63460.6710
SSIM0.62830.63140.63400.54330.61720.59690.64290.6438
Table 4. The evaluation index values of fused images in Figure 10.
Table 4. The evaluation index values of fused images in Figure 10.
MethodsMWGFGuFLPDCTNSCTSRMST-SRThe Proposed
Indexes
EN7.20737.06576.98856.92466.93597.29027.35567.5855
SF15.445316.478018.723919.506818.316418.315218.974819.8665
QAB/F0.54910.56400.55740.36190.54110.47410.56580.5774
SSIM0.66410.69540.68980.47890.69580.61680.67970.6826
Table 5. The evaluation index values of fused images in Figure 11.
Table 5. The evaluation index values of fused images in Figure 11.
Methods MWGFGuFLPDCTNSCTSRMST-SRThe Proposed
Indexes
EN6.09327.16207.34517.30817.33177.11947.06357.6861
SF27.871327.004429.246528.984728.296125.755029.260129.6650
QAB/F0.64730.63180.58570.41430.56530.55860.58860.5968
SSIM0.65280.66470.67360.55480.68710.65900.67330.6930

Share and Cite

MDPI and ACS Style

Ma, X.; Hu, S.; Liu, S.; Fang, J.; Xu, S. Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering. Electronics 2019, 8, 303. https://doi.org/10.3390/electronics8030303

AMA Style

Ma X, Hu S, Liu S, Fang J, Xu S. Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering. Electronics. 2019; 8(3):303. https://doi.org/10.3390/electronics8030303

Chicago/Turabian Style

Ma, Xiaole, Shaohai Hu, Shuaiqi Liu, Jing Fang, and Shuwen Xu. 2019. "Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering" Electronics 8, no. 3: 303. https://doi.org/10.3390/electronics8030303

APA Style

Ma, X., Hu, S., Liu, S., Fang, J., & Xu, S. (2019). Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering. Electronics, 8(3), 303. https://doi.org/10.3390/electronics8030303

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop