Next Article in Journal
Salivary ATR-FTIR Spectroscopy Coupled with Support Vector Machine Classification for Screening of Type 2 Diabetes Mellitus
Previous Article in Journal
Fabrication of a Low-Cost Microfluidic Device for High-Throughput Drug Testing on Static and Dynamic Cancer Spheroid Culture Models
Previous Article in Special Issue
A Non-Conventional Review on Multi-Modality-Based Medical Image Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodality Medical Image Fusion Using Clustered Dictionary Learning in Non-Subsampled Shearlet Transform

1
Department of Computer Science and Engineering, Graphic Era (Deemed to Be University), Dehradun 248002, Uttarakhand, India
2
School of Computer Science Engineering and Technology, Bennett University, Greater Noida 201310, Uttar Pradesh, India
3
Department of Computer Science and Engineering, Engineering College Ajmer, Ajmer 305025, Rajasthan, India
4
School of Computer Science, University of Petroleum and Energy Studies, Dehradun 248007, Uttarakhand, India
5
Department of Applied Data Science, Noroff University College, 4612 Kristiansand, Norway
6
Artificial Intelligence Research Center (AIRC), Ajman University, Ajman 346, United Arab Emirates
7
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 13-5053, Lebanon
8
University of Zilina, Univerzitna 1, 01026 Zilina, Slovakia
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(8), 1395; https://doi.org/10.3390/diagnostics13081395
Submission received: 4 March 2023 / Revised: 23 March 2023 / Accepted: 24 March 2023 / Published: 12 April 2023

Abstract

:
Imaging data fusion is becoming a bottleneck in clinical applications and translational research in medical imaging. This study aims to incorporate a novel multimodality medical image fusion technique into the shearlet domain. The proposed method uses the non-subsampled shearlet transform (NSST) to extract both low- and high-frequency image components. A novel approach is proposed for fusing low-frequency components using a modified sum-modified Laplacian (MSML)-based clustered dictionary learning technique. In the NSST domain, directed contrast can be used to fuse high-frequency coefficients. Using the inverse NSST method, a multimodal medical image is obtained. Compared to state-of-the-art fusion techniques, the proposed method provides superior edge preservation. According to performance metrics, the proposed method is shown to be approximately 10% better than existing methods in terms of standard deviation, mutual information, etc. Additionally, the proposed method produces excellent visual results regarding edge preservation, texture preservation, and more information.

1. Introduction

Computational methods of image processing are often used to achieve image fusion. The primary goal of image fusion is to reduce the volume of data produced with a sharp, comprehensive image that can be useful in clinical and scientific research. Compared to integrating separate modalities, synthesis outcomes between two or more modalities in a single image are more comprehensible, correct, and high quality [1]. Due to multimodality medical image fusion, a new medical image can be reconstructed via fusion algorithms. Multimodal image fusion aims to merge multiple input images into a single coherent whole. This fused image contains more medical information than individual medical images. Single sensor, multisensor, multi-view, multi-model, and multi-focus fusions are only a few types of image fusion. One sensor, part of a unified sensor fusion system, can take many sequential images from different viewpoints and fuse them into a single, unified image. Using many sensors, a system creates several distinct scene video sequences, which are combined to form a single video. Fusion imaging uses many camera angles to take images of the same scene from different vantage points [2]. By piecing together the individual photos, a whole image can be made.
By combining the images from several different medical imaging models, we can obtain a single image that is consistent throughout. When there are several input images of the same scene, each with a varied depth of focus [3,4,5], multi-focus fusion systems are used to combine the information from all of the images into one. Image fusion techniques are used for image processing in a wide variety of application fields, some of which include medical image fusion and sensor networks used for area surveillance, tracking, and environmental monitoring. These are only two examples of the many possible application areas. Under the context of this scenario, the team is required to make substantial use of diagnostic imaging data, which may include results from CT scans, MRIs, and other tests. There are situations when a single diagnostic test might not be sufficient, necessitating additional multimodal medical imaging. Even though there are a variety of different solutions available, this method of processing medical images is still widely utilized [6,7,8]. The combination of two or more medical image components results in an increased number of newly created medical images being made available for use. In this paper, a new fusion approach in the shearlet domain is presented for the purpose of integrating a wide variety of medical image types [9].
The study of image fusion has recently gained popularity due to its utility in diverse advancing fields, such as medicine, remote sensing, and defense. This technique continues to produce vital data for image fusion since it is inexpensive, resilient, and provides high-resolution images. However, obtaining crucial data for image fusion is a common and challenging issue due to the high cost of devices and the amount of blur data.
Image fusion is the process of combining two or more images, each of which may be different from the others or identical, to create a new image that incorporates features from each original image. This new image should keep maximum information from the original images while also minimizing any artifacts that may have been introduced during the fusion process, as is the case in many practical applications [10]. The fundamental objective of fusion is to create a single high-resolution image from a collection of lower-resolution ones. Sharp images are necessary for diagnosing diseases, such as coronary artery disease (CAD), which develops when the heart does not receive adequate oxygen. In addition, neurologists play an important role in the prognosis of brain tumor conditions; hence, image fusion is used to analyze brain scans from different modalities. Each researcher’s motivations might make image fusion an intriguing and novel problem. Satellite imaging, medical imaging, aviation, the detection of concealed weapons, the use of digital cameras for battlefield monitoring and situational awareness, the tracking of targets with surveillance cameras, and the authentication of individuals in the geo-informatics industry are just a few of the many modern applications of image fusion [11]. Reading the literature has numerous benefits, such as dictionary learning, cluster analysis, sum-modified Laplacian (SML), and contrast-based fusion.
In the present study, low-frequency fusion sub-bands with a new SR method use coincidental instances that allow DTCWT and SR simultaneously. Patches in the source image were found using structural similarities, which were then classified and grouped into clusters. All the condensed sub-dictionaries in the cluster are compressed and merged to form an adaptive, clustered, and condensed sub-dictionary (ACCD). The fusion algorithm forms a one-of-a-kind algorithm known as the modified sum-modified Laplacian (MSML), which is based on the LARS algorithm and synthesizes sparse coefficients from the synthetic sparse vectors that are formed using the fusion algorithm. An example of this is the employment of a sub-band fusion approach, which combines the usage of the high-frequency maximal complete ruling, and consistency affirmation working together in an instance. More information about the patient’s health can be gleaned through the fusion of data from many medical imaging modalities. The evidence synthesis in radiographic images is one example of the multimodal approach to a medical diagnosis that recent advances in the field have favored. To better understand the patient’s blood flow and metabolic rate, it is necessary to segment all the medical images depending on their relatively inadequate functional image spatial resolution. The physical structure is assumed to obtain a reasonably high spatial resolution.
With these motivations, a new multimodality medical image fusion method is proposed. The major contributions of the paper include:
  i.
A dictionary learning method based on cluster analysis is introduced in low-frequency sub-band fusion. In this technique, structural image patch attributes are pooled and mathematically connected to increase computation efficiency;
 ii.
For low sub-band fusion, the modified sum-modified Laplacian (MSML) constructs artificially sparse vectors by employing saliency features to calculate low-frequency sub-band local features;
iii.
A directive contrast-based fusion is introduced by calculating the local facts of high-frequency sub-band MSML.
The rest of this paper is organized as: In Section 2, related work is discussed. Section 3 shows the methods that are utilized in the proposed work. In Section 4, the proposed work is discussed. Section 5 shows the result and the discussion. Finally, Section 6 draws conclusions.

2. Related Work

Zhang et al. [12] proposed a multimodality medical image fusion method where multiscale morphology gradient-weighted local energy and a visual saliency map are used to improve the results of existing state-of-the-art methods. The results are good in terms of the statistical methods and visual appearance. However, the contrast of the fused image is not up to the mark for many complex images. Ramlal et al. [13] introduced a method using a hybrid combination of non-subsampled contourlet transform and stationary wavelet transform for medical image fusion. The results are also good regarding visual appearance and performance metrics. However, due to more multiple transforms, the computation cost is increased. To combine medical images from different modalities, Dogra et al. [14] proposed utilizing guided filters and image statistics in the multidirectional shearlet transform domain. Multimodality medical picture fusion was proposed by Ullah et al. [15], who suggested using local features fuzzy sets in conjunction with a novel sum-modified Laplacian in a non-subsampled shearlet transform domain. The non-subsampled shearlet transform and the activity measure were proposed by Huang et al. [16] as a method for optimizing information gain during picture fusion. Shearlet-domain-based fusions produce good results generally, but their lack of contrast in high-texture photos is suboptimal. Multimodality medical image fusion was proposed by Liu et al. [17], using an image decomposition framework, non-subsampled shearlet transformation, and a weighted fusion function. Mehta et al. [18] proposed using a guided filter in the NSCT domain to achieve more comprehensive informatics outcomes in multimodality medical image fusion. Though the guided filter produces respectable outcomes overall, its performance falls short for images with a dense texture in terms of edge retention.
Contrast-based fusion rules are also employed for fusion purposes, and local energy is given to the reactor. Because of this, edges are more reliably found and stabilized when the decomposition approach is employed. In [19], a new layer-based fusion approach considers layer differences by separating the base and detail layers while using saliency characteristics that seek coincidences. Using the characteristics’ ability to highlight important regions of relevance, crisp and smooth fusion outcomes may be achieved with minimal effort. Over the last year, SR-based fusion approaches have suffered a significant drop in popularity in the multisensor image fusion sector. This SR-based fusion technology is only successful if the dictionary is overflowing with components and the best-in-class fusion algorithm is constructed. DCT, DWT, Gabor, and Ridgelet are some of the sparse fusion methods often used [20,21,22]. Another image that is frequently used is that of dictionary learning. Dictionary entries become significantly more difficult when working with images with complicated structures. An intelligent learner merges all the patches from the input images using a heuristic dictionary, proving his or her intelligence. Because the choice of a dictionary is crucial in SR, researchers have referred to the image patch clustering technique. Despite this, a large number of images remain unsolved [23,24,25]. The computation costs of dictionary learning [26,27,28,29,30] and sparse coding are higher than wavelet-based fusion methods.
The use of a suitable fusion rule enables the successful synthesis of a sparse artificial set of coefficients, which is observed in the following cases: A dictionary-based learning approach requires two parameters to be effective for individuals who do not desire to calculate words in advance: how long it takes to build the data and how many steps the learner repeats in the process of creating the data. Because the ideal learning parameter in classical techniques, such as the K-SVD, is decided by the set of rules, it is difficult to manage learning time. Furthermore, sparse coding may incur additional expenses since it may need an evaluation cost less than the input datas’ size. Sparse representation: the number of patches increases in proportion to the size of the input data. On the other hand, SR-based fusion techniques have historically met difficulties because the rules used are rarely relevant in the current temporal context of the experiment. Fusion algorithms, which are commonly used to find visible images in an infrared image algorithm, produce images that are visible in the IR results when applied to the IR results.
The importance of medical imaging, in both medical research and clinical practice with an intent to achieve high image quality, is increasing and demands representation or simulation. In certain situations, the complete spectrum structure of digital image processing can aid in medical diagnosis. Radiologists can diagnose organs or illnesses effectively, with a combination of images of the organs or diseases involved. It is noted that the type and model of the instruments used in medical imaging also restrict their ability to offer such information. The presence of vital organs or living tissues is referred to as “heterogeneity” in medical imaging. The differences in size and shape can occur even when the same modality is used to gather the data due to factors such as the object’s shape, internal structure, or even just the fact that separate images of the same patient were acquired at various times. The boundary between foreground and background cannot be erased in the study of biological anatomy. The outcomes of automatic medical image analysis are dependent on several factors. Photo blending has been proven to enhance image quality drastically. The error- and redundancy-free multimodality medical image fusion technique aims to improve image quality [31,32,33,34,35,36].
Wadhwa et al. suggested a mechanism for predicting the lockdown period to be implemented to successfully contain the spread of COVID-19 in India [37]. Four methods were employed to create an epidemic alarm system, including Random Forest Regression, Decision Tree Regression, Support Vector Regression, and Multiple Linear Regression [38]. Dhaka et al. [39] analyzed the differences between the stationary wavelet transform (SWT) and the discrete wavelet transform (DWT) for different applications and found SWT outperforms DWT. According to a study by Dhaundiyal [40], a novel SWT-based multimodality fusion approach was presented for medical image fusion. In this method, the source images are first decomposed into an approximation layer (coarse layer) and a detail layer using the SWT scheme and then the Fuzzy Local Information C-Means Clustering (FLICM) and local contrast fusion approach are applied to the distinct layers to counteract the blurring effect, maintain sensitivity, and preserve quality evaluation. The suggested approach [41] uses a non-subsampled shearlet transform (NSST) to extract low and high-frequency components from input images. Low-frequency components are fused using a co-occurrence filter (CoF), and a unique process is employed to deconstruct and merge the base layers and detail layers using the local extrema (LE) approach. Sum-modified Laplacian (SML) is used to fuse the high-frequency coefficients in an edge-preserving image fusion approach [41].

3. Preliminaries

This section presents an overview of the methods that are used in the proposed work. Some of the main methods are discussed here in the subsections below.

3.1. Non-Subsampled Shearlet Transform (NSST)

The NSST is not only a practical instrument for multiscale geometric research because of its amazing ability to discover linear singularities but also a correct description of the 2-D sparse method. This is because of its success in detecting linear singularities. The shift-invariance and anisotropic direction selectivity of the discrete wavelet transform are signature features that set it apart from other wavelet transform types. The discrete wavelet transform is an especially useful tool for deciding the location of point-wise singularities. Because it uses a non-subsampled Laplacian pyramid filter, the NSST can perform multiscale directional localization thanks to this implementation choice. It outperforms NSCT in several crucial areas, including productivity, flexibility, and stability against orientation changes, to name a few of these categories. Through NSST, images are decomposed into two major parts, (i) low-frequency components and (ii) high-frequency components. These low- and high-frequency components provide the features of the images, which can be utilized here for multimodality medical image fusion.

3.2. Clustered Dictionary Learning

In this method, the clusters-based dictionary is generated by finding the local features in terms of patches, which can be further used for image fusion. The previously mentioned approach shows that the patch Pk ∈ P lies next to edge cluster Ce if its activity level is more than a threshold. Suppose the patch’s activity level is lower than the lowest threshold, it refers to a smooth cluster 𝒞 𝒮 , except for when it refers to texture cluster 𝒞 𝓉 . This procedure is repeated continuously until all patches in the joint patch set P are arranged. At last, every cluster in the set of clusters 𝒞 𝒮 is data trained in the online dictionary-based learning algorithm, and the compressed sub-dictionaries 𝒟 , 𝒟 𝓉 , and 𝒟 𝒮 are acquired. All the acquired sub-dictionaries are integrated to make a dictionary D. Create the 𝒞 , 𝒞 𝓉 , 𝒞 𝒮 -clustered sub-dictionaries and then create the 𝒟 -clustered combined dictionary where each of the clusters is trained to apply the ODL algorithm to attain the resulting sub-dictionaries 𝒟 , 𝒟 𝓉 ,   and   𝒟 𝒮 ,   respectively , and join each sub-dictionary to make the final dictionary, such as 𝒟 = 𝒟 , 𝒟 𝓉 , 𝒟 𝒮 , where 𝒟 ,   𝒟 𝓉 ,     and   𝒟 𝒮 are the sub-dictionaries of edge cluster, texture cluster, and set of clusters, respectively.

3.3. Visual Saliency Features

To extract saliency characteristics, the largest symmetric surround saliency method is employed (MSS). The following is the method that must be used to implement the largest symmetric surround saliency (MSS) method, as shown in Equation (1).
P(i, j) = ||P1(i, j) − P2(i, j)||
where P(i, j) are the saliency features, £(i, j) is the average pixel values of all CIELAB, and || || is the L2 norm. The average pixel values of all CIELAB are obtained as shown in Equation (2).
P ( i ,   j )   =   1 r x = i m i + m y = j n j + n P x , y
where m = min(i, wi), n = min(j, hj), r = (2i + 1)(2j + 1), w and h are the width and height of the image, respectively.

4. Proposed Methodology

In the proposed methodology, two different modality images are utilized as input images. Initially, NSST is performed over both input images to obtain low- and high-frequency components. Over the low-frequency components of both input images, a gradient operator is applied to obtain horizontal and vertical direction for extracting detailed features. Over these features, saliency features are obtained by utilizing the concept of MSS. Over these saliency features, a modified SML operation is introduced. These features are further clustered by performing dictionary-based learning method. Using modified SML operation, fusion operation is performed on both dictionary-learning-based clusters. On the other side, high-frequency components are processed using directive contrast-based fusion. Finally, inverse NSST is performed over both modified low- and high-frequency components. In proposed work, a dictionary-based learning algorithm is first defined and then the complete fusion substructure of the concept and the technique connected to the fusion algorithm on sub-band images are described. The fundamental objective of this study is to develop a compact, well-organized over-completion dictionary with the optimal structure and high computational efficiency to compete with existing dictionary-based learning approaches. To demonstrate the efficacy of the new approach to dictionary creation, a clustering-based learning mechanism for categorizing input image patches with a geometrically similar structure is used. Therefore, the following features of the input pictures can be maintained and exploited for accurate segmentation. The borders, textures, and smooth areas of an image are the key image components that may alter the overall texture; hence, it is focused on in the present study. In any given image, the details at the image’s edges and textures stand out the most. Edges are perceived differently depending on the smoothness of the component, but they still blend into the background when viewed by a person. The steps of the proposed algorithm are shown in Figure 1.
Step 1 (NSST decomposition): Perform NSST decomposition on input images with parameters c = 1 and d = 8 to obtain low- and high-frequency components on both input multimodal medical images, as shown in Equation (3).
L f 1 N S S T ,   H f 1 N S S T = N S S T A i , j   and   L f 2 N S S T ,   H f 2 N S S T = N S S T B i , j
Step 2 (Low sub-band fusion): The gradient operator is used to obtain horizontal and vertical orientation across the low-frequency components of both input images in order to extract finer information. On top of these components, the idea of maximum saliency (MSS) is used to obtain saliency attributes. A refined SML technique is introduced over these prominent indicators. We then use a dictionary-based learning approach to further categorize these traits. We perform a fusion operation on both clusters based on dictionary learning using a modified SML procedure. Perform the below sub-steps to obtain a low sub-band fused image.
(a)
Find the gradient information GA and GB in horizontal and vertical directions from both input images;
(b)
Estimate modified Laplacian (ML), as shown in Equation (4);
i , j   =   a b s P i , j G r a d i , j + a b s P i , j G r a d V i , j
(c)
Develop MSML by adding the ML as shown in Equation (5);
M S M L 𝓅 𝓀 = i = 1 𝓃 j = 1 𝓃 M L i , j
where 𝓃 × 𝓃 is the size of 𝓅 k ;
(d)
Acquire the 𝒞 , 𝒞 𝓉 , 𝒞 𝒮 clusters using MSML:
(i)
Separate the source images I A   and   I B into 𝓃 × 𝓃 patches, P A   and   P B , respectively;
(ii)
Combine P A   and   P B to make a joint patch set P = P A , P B ;
(iii)
Search the MSML for every 𝓅 𝓀 P ;
(iv)
Fix the thresholds T H 1 , T H 2 by utilizing as shown in Equations (6) and (7):
T H 1 = 0.13     m a x M S M L 𝓅 h
0.07     max M S M L 𝓅 h
(e)
Perform the equation below to make the C ,   C 𝓉 ,   C S clusters. The categorization approach is described, as shown in Equation (8);
C J = C ,   i f   M S M L   𝓅 h T H 1 C 𝓉 ,   i f   T H 1 > M S M L 𝓅 h T H 2 C S ,   i f   T H 2 > M S M L 𝓅 h
(f)
The sum-modified-Laplacian (SML) is a technique that has proven effective in the field of medical picture fusion. When applied to the altered image, fusion rules based on a larger SML always lead to either information loss in the fused spatial domain or image distortion. New filters, the average filter, and the median filter, are available in the latest version of SML, which is utilized for medical picture fusion. MSML is the main computation to evaluate all activity levels of the image patch. It elaborates on the small information, the image constraint. Increasing the value gives more details as it exists. Suppose M S M L   𝒾 ; L A and M S M L 𝒾 ; L B represent the 𝒾 th patch’s modified SML of low-frequency sub-images L A   and   L B , the recommended fusion approach is described, as shown in Equation (9):
a L F 𝒾 = a L A 𝒾 ,   i f   M S M L 𝒾 ; L A M S M L 𝒾 ; L B a L B 𝒾 ,   o t h e r w i s e
where V L F 𝒾 = D a L F 𝒾 + 𝓂 L F 𝒾 and the fusion mean value 𝓂 L F 𝒾 is followed by Equation (10),
𝓂 L F 𝒾 = 𝓂 L A 𝒾 ,   i f   a L F 𝒾 = a L A 𝒾 𝓂 L B 𝒾 ,   o t h e r w i s e
Step 3 (High sub-band fusion): The coefficients show that the sub-images with higher frequencies often have information from the source image. Moreover, because noise is usually caused by high frequencies, it can mess up calculations for fusion, which can lead to wrong sharpness values and hurt the quality of the fusion process. To illustrate these results, a new set of criteria based on the use of directed contrast has been made. According to the step-by-step approach, the following is an explanation of the complete operation.
(a)
Estimate the directive contrast ( D L i , j ) of NSST high-frequency coefficients using low sub-band coefficients as shown in Equations (11) and (12):
D L A i , j = M S M L A i , j A i , j i f   A i , j > 0 M S M L A i , j O t h e r w i s e
Similarly,
D L B i , j = M S M L B i , j B i , j i f   B i , j > 0 M S M L B i , j O t h e r w i s e
(b)
Apply the following fusion rule to the high-frequency coefficients ( H f i , j ) as shown in Equation (13):
H f i , j = H f A i , j i f   D L A > D L B H f B i , j O t h e r w i s e
Step 4: Follow the below for obtaining a fused image using inverse NSST as shown in Equation (14):
R = N S S T 1 L f , H f

5. Experimental Results

Using the software MATLAB Version 9.4 (R2018a: India), the experimental evaluation was completed. The proposed methodology for multimodality medical image fusion was performed.

5.1. Dataset

The analysis was carried out on the entire collection of 210 medical images that were coupled together. The images were obtained from a public access database Atlas (http://www.med.harvard.edu/AANLIB/home.html (accessed on 22 May 2022)) [36]. The multimodal imaging modalities that are frequently used are CT scanning and magnetic resonance imaging (MRI). The complex make-up of human tissue delivers information that is more precise and detailed than ever before. The ability of CT scans to supply highly correct anatomical reconstructions makes them useful not only for diagnosis but also for treatment. When seen at an oblique angle, more of the inner workings of an organ are clear. On the other hand, bone, soft tissue, and lung are more beneficial for studying the skeletal and connective tissue components. One of the methods used to obtain a better understanding of the human body is the use of windows that allow looking through bone, soft tissue, or the lungs, i.e., SPECT. SPECT images are used rather often in the field of CT imaging. These multimodality medical images are further used in our experimental analysis. All the images used for experimental results had a resolution of 512 × 512. If the resolution size of both input image is not the same, preprocessing should be applied to obtain the same resolution of the input images. However, we tested all experiments on the same resolutions of both input images.
Pairs of medical images are available in the public database (http://www.med.harvard.edu/AANLIB/home.html (accessed on 22 May 2022)), and they include modalities including computed tomography (CT) and magnetic resonance imaging (MRI). There are numerous multimodality effects seen in Figure 2, Figure 3, Figure 4 and Figure 5.

5.2. Results and Discussion

The proposed methodology is compared with recently proposed methods, such as those of Zhang et al. [12], Ramlal et al. [13], Dogra et al. [14], Ullah et al. [15], Huang et al. [16], Liu et al. [17], and Mehta et al. [18].
Figure 2a,b are the two input multimodalities CT and MRI. Figure 2c–j are the results of Zhang et al. [12], Ramlal et al. [13], Dogra et al. [14], Ullah et al. [15], Huang et al. [16], Liu et al. [17], Mehta et al. [18], and the proposed method, respectively. In Figure 2, the results are good in terms of edge preservation and providing more informatics clinical details. In this respect, the results of [12] are good, but the textures in homogenous regions are not effectively preserved. Similarly, the results of [13] are also not effectively preserved in terms of contrast and brightness. The results of [14,17] are well preserved in all the details, but in highly textured regions, the results are not excellent. The results of [15,16,18] are good but the high-textured details are not satisfactory. However, in comparison to others, the proposed method gives the best results in terms of sharpness, smoothness, texture preservation, and more informatic clinical details.
Figure 3a,b display both the MR-T2 image and the SPET image. The findings of [12,13,14,15,16,17,18] as well as the proposed approach are presented in Figure 3c–j, respectively. In Figure 3, the outcomes are favorable in terms of edge preservation and added clinical data gleaned through informatics. While [12] achieves outstanding results overall, it does a less than stellar job of preserving textures in areas with high homogeneity. The brightness and contrast of the outputs of [13] are likewise not kept very well. The outcomes of [14,17] are great in low-textured areas and good in high-textured areas, respectively. The results of [15,16,18] are satisfactory, but the high-textured details are lacking. Sharpness, smoothness, texture preservation, and more informatic clinical features are all improved upon using the proposed strategy. On the other hand, in contrast to earlier methods, the one that is proposed yields the best results in terms of sharpness, smoothness, the preservation of texture, and additional informatics clinical features.
The MR-T2 image as well as the SPET image are presented in Figure 4a,b, respectively. The results of Zhang et al. [12], Ramlal et al. [13], Dogra et al. [14], Ullah et al. [15], Huang et al. [16], Liu et al. [17], Mehta et al. [18] are provided in Figure 4c–j, respectively, along with the suggested methodology. Figure 4 demonstrates that the outcomes are positive in terms of edge preservation as well as new clinical data obtained through informatics. The method that has been offered, on the other hand, in contrast to those that have been used in the past, produces the best results in terms of sharpness, smoothness, the preservation of texture, and extra informatics clinical aspects. The results of [12] are good in this instance, but the textures in homogeneous regions are not particularly well preserved. The results from [13] are similarly poorly preserved in terms of contrast and brightness. The results from [14,17] are excellent in that all of the details are preserved, but the results are less than perfect in highly textured regions. Although [15,16,18] produce good results, the high-textured detail is not particularly satisfactory. The proposed method, however, produces the best results when compared to other approaches in terms of sharpness, smoothness, texture preservation, and more clinical informatics information.
Figure 5a,b show zoomed-in regions of input multimodality medical images. Figure 5c–j show the results of Zhang et al. [12], Ramlal et al. [13], Dogra et al. [14], Ullah et al. [15], Huang et al. [16], Liu et al. [17], Mehta et al. [18] as well as the suggested technique, respectively. Figure 5 shows that the outcomes in terms of edge preservation and additional clinical data collected through informatics are both positive. The method presented, on the other hand, produces the best results in terms of sharpness, smoothness, texture preservation, and extra informatics clinical characteristics as compared to previous methods. The findings of [12] are satisfactory in this regard; nonetheless, the textures in the homogenous zones are not kept remarkably. The contrast and brightness of the results of [13] are likewise not adequately preserved in the same way. The outputs of [14,17] do a good job of preserving all of the features, but in areas with a lot of texture, their performance is less than stellar. The results of [15,16,18] are satisfactory; however, the particulars of the high-textured results are not particularly outstanding. On the other hand, in contrast to previous methods, the one that is proposed yields the best results in terms of sharpness, smoothness, the preservation of texture, and additional informatics clinical features.
Visual results were not sufficient for the resulting analysis; hence, the results of the existing methods were tested and evaluated using performance metrics. To check the accuracy of the existing methods, some parameters were used, such as MI AB , F ,   Q AB , F ,   and   BSSIM . The results were tested over 80 pairs of medical images and the average values are shown in Table 1. From Table 1, it can be analyzed that the transform domain approaches give better outcomes. The bold values in Table 1 show the best performance metric values for different image datasets.

6. Conclusions

A diagnostic image analysis based on multimodality is presented in the present study. Advanced human data should be sensitive to better contrast (high), pixel density, edge detail, contrast focus, view dependencies, fusion device edge, and texture detection.
The proposed method gives better results in terms of visual results such as smoothness and sharpness in high-textured images. Other than the visual results, performance metrics are also evaluated where the values of the performance metrics show better results in comparison to existing methods. The study discusses several forms of errors in imaging data. Moreover, it showcases the lack of noise and the improvement in the information presented in the fused image and compares the data obtained for calculation from the original image. The findings suggest that current transform domain methods have better outcomes than using other spatial domain structures. The performance metrics also prove that in addition to visual effects, techniques using transform domain strategies provide enhanced results compared to analogous spatial domain schemes.

Author Contributions

Conceptualization, writing original draft; M.D. and P.S.; data curation, A.M.; formal analysis, A.M.; investigation, P.S., R.S. and D.S.; methodology, project administration, M.D., P.S., L.S. and V.S.; resources, L.S.; software, M.D., R.S., V.S., S.K., A.M. and D.S.; visualization, V.S., P.S., A.M. and S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by the project of Operational Program Integrated Infrastructure: Independent research and development of technological kits based on wearable electronics products, as tools for raising hygienic standards in a society exposed to the virus causing the COVID-19 disease, ITMS2014+ code 313011ASK8.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

NSSTNon-subsampled shearlet transform
MSMLModified sum-modified Laplacian
CADCoronary artery disease
ACCDAdaptive, clustered, and condensed sub-dictionary
SWTStationary wavelet transform
DWTDiscrete wavelet transform
FLICMFuzzy Local Information C-Means Clustering
SMLSum-modified Laplacian
CoFCo-occurrence filter
LELocal extrema
NSCTNon-subsampled contourlet transform
MLModified Laplacian
MRIMagnetic resonance imaging
CTComputed tomography

References

  1. Xu, Z. Medical image fusion using multi-level local extrema. Inf. Fusion 2014, 19, 38–48. [Google Scholar]
  2. Zhang, P.; Yuan, Y.; Fei, C.; Pu, T.; Wang, S. Infrared and visible image fusion using co-occurrence filter. Infrared Phys. Technol. 2018, 93, 223–231. [Google Scholar] [CrossRef]
  3. Li, S.; Yin, H.; Fang, L. Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 2012, 59, 3450–3459. [Google Scholar] [CrossRef] [PubMed]
  4. Sharma, D.; Kudva, V.; Patil, V.; Kudva, A.; Bhat, R.S. A Convolutional Neural Network Based Deep Learning Algorithm for Identification of Oral Precancerous and Cancerous Lesion and Differentiation from Normal Mucosa: A Retrospective Study. Eng. Sci. 2022, 18, 278–287. [Google Scholar] [CrossRef]
  5. Yin, M.; Liu, W.; Zhao, X.; Yin, Y.; Guo, Y. A novel image fusion algorithm based on non subsampled shearlet transform. Optik 2014, 125, 2274–2282. [Google Scholar] [CrossRef]
  6. Ganasala, P.; Kumar, V. Multi-modality medical image fusion based on new features in NSST domain. Biomed. Eng. Lett. 2014, 4, 414–424. [Google Scholar] [CrossRef]
  7. Shah, M.; Naik, N.; Somani, B.K.; Hameed, B.Z. Artificial intelligence (AI) in urology-Current use and future directions: An iTRUE study. Turk. J. Urol. 2020, 46 (Suppl. S1), S27–S39. [Google Scholar] [CrossRef]
  8. Ganasala, P.; Kumar, V. Feature-motivated simplified adaptive PCNN-based medical image fusion algorithm in NSST domain. J. Digit. Imaging 2016, 29, 73–85. [Google Scholar] [CrossRef] [Green Version]
  9. Singh, R.; Srivastava, R.; Prakash, O.; Khare, A. Multi-modal medical image fusion in dual tree complex wavelet transform domain using maximum and average fusion rules. J. Med. Imaging Health Inform. 2012, 2, 168–173. [Google Scholar] [CrossRef]
  10. Qu, X.-B.; Yan, J.-W.; Xiao, H.-Z.; Zhu, Z.-Q. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampledcontourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  11. Patil, V.; Vineetha, R.; Vatsa, S.; Shetty, D.K.; Raju, A.; Naik, N.; Malarout, N. Artificial neural network for gender determination using mandibular morphometric parameters: A comparative retrospective study. Cogent Eng. 2020, 7, 1723783. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Jin, M.; Huang, G. Medical image fusion based on improved multi-scale morphology gradient-weighted local energy and visual saliency map. Biomed. Signal Process. Control 2022, 74, 103535. [Google Scholar] [CrossRef]
  13. Ramlal, S.D.; Sachdeva, J.; Ahuja, C.K.; Khandelwal, N. An improved multi-modal medical image fusion scheme based on hybrid combination of nonsubsampledcontourlet transform and stationary wavelet transform. Int. J. Imaging Syst. Technol. 2019, 29, 146–160. [Google Scholar] [CrossRef]
  14. Dogra, A.; Kumar, S. Multi-modality medical image fusion based on guided filter and image statistics in multidirectional shearlet transform domain. J. Ambient. Intell. Humaniz. Comput. 2022, 1–15. [Google Scholar] [CrossRef]
  15. Ullah, H.; Ullah, B.; Wu, L.; Abdalla, F.Y.; Ren, G.; Zhao, Y. Multi-modality medical images fusion based on local-features fuzzy sets and novel sum-modified-Laplacian in non-subsampled shearlet transform domain. Biomed. Signal Process. Control 2020, 57, 101724. [Google Scholar] [CrossRef]
  16. Huang, D.; Tang, Y.; Wang, Q. An Image Fusion Method of SAR and Multispectral Images Based on Non-Subsampled Shearlet Transform and Activity Measure. Sensors 2022, 22, 7055. [Google Scholar] [CrossRef]
  17. Liu, X.; Mei, W.; Du, H. Multi-modality medical image fusion based on image decomposition framework and nonsubsampledshearlet transform. Biomed. Signal Process. Control 2018, 40, 343–350. [Google Scholar] [CrossRef]
  18. Mehta, N.; Budhiraja, S. Multi-modal Medical Image Fusion using Guided Filter in NSCT Domain. Biomed. Pharmacol. J. 2018, 11, 1937–1946. [Google Scholar] [CrossRef]
  19. Maqsood, S.; Javed, U. Multi-modal medical image fusion based on two-scale image decomposition and sparse representation. Biomed. Signal Process. Control 2020, 57, 101810. [Google Scholar] [CrossRef]
  20. Hu, Q.; Hu, S.; Zhang, F. Multi-modality medical image fusion based on separable dictionary learning and Gabor filtering. Signal Process. Image Commun. 2020, 83, 115758. [Google Scholar] [CrossRef]
  21. Zhu, Z.; Chai, Y.; Yin, H.; Li, Y.; Liu, Z. A novel dictionary learning approach for multi-modality medical image fusion. Neurocomputing 2016, 214, 471–482. [Google Scholar] [CrossRef]
  22. Zhu, Z.; Yin, H.; Chai, Y.; Li, Y.; Qi, G. A novel multi-modality image fusion method based on image decomposition and sparse representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  23. Cao, Y.; Li, S.; Hu, J. Multi-Focus Image Fusion by Nonsubsampledshearlet Transform. In Proceedings of the 2011 Sixth International Conference on Image and Graphics, Hefei, China, 12–15 August 2011; pp. 17–21. [Google Scholar]
  24. Gao, G.; Xu, L.; Feng, D. Multi-focus image fusion based on non-subsampled shearlet transform. IET Image Process. 2013, 7, 633–639. [Google Scholar]
  25. Fu, Z.; Zhao, Y.; Xu, Y.; Xu, L.; Xu, J. Gradient structural similarity based gradient filtering for multi-modal image fusion. Inf. Fusion 2020, 53, 251–268. [Google Scholar] [CrossRef]
  26. Goyal, S.; Singh, V.; Rani, A.; Yadav, N. FPRSGF denoised non-subsampled shearlet transform-based image fusion using sparse representation. Signal Image Video Process. 2020, 14, 719–726. [Google Scholar] [CrossRef]
  27. Benjamin, J.R.; Jayasree, T. An Efficient MRI-PET Medical Image Fusion Using Non-Subsampled Shearlet Transform. In Proceedings of the 2019 IEEE International Conference on Intelligent Techniques in Control, Optimization and Signal Processing (INCOS), Tamilnadu, India, 11–13 April 2019; pp. 1–5. [Google Scholar]
  28. Luo, X.; Zhang, Z.; Zhang, B.; Wu, X. Image fusion with contextual statistical similarity and nonsubsampledshearlet transform. IEEE Sens. J. 2016, 17, 1760–1771. [Google Scholar] [CrossRef]
  29. Zhao, C.; Guo, Y.; Wang, Y. A fast fusion scheme for infrared and visible light images in NSCT domain. Infrared Phys. Technol. 2015, 72, 266–275. [Google Scholar] [CrossRef]
  30. Moonon, A.U.; Hu, J.; Li, S. Remote sensing image fusion method based on nonsubsampledshearlet transform and sparse representation. Sens. Imaging 2015, 16, 23. [Google Scholar] [CrossRef]
  31. Ghimpeţeanu, G.; Batard, T.; Bertalmío, M.; Levine, S. A decomposition framework for image denoising algorithms. IEEE Trans. Image Process. 2015, 25, 388–399. [Google Scholar] [CrossRef] [Green Version]
  32. Hou, R.; Zhou, D.; Nie, R.; Liu, D.; Ruan, X. Brain CT and MRI medical image fusion using convolutional neural networks and a dual-channel spiking cortical model. Med. Biol. Eng. Comput. 2019, 57, 887–900. [Google Scholar] [CrossRef]
  33. Asha, C.S.; Lal, S.; Gurupur, V.P.; Saxena, P.P. Multi-modal medical image fusion with adaptive weighted combination of NSST bands using chaotic grey wolf optimization. IEEE Access 2019, 7, 40782–40796. [Google Scholar] [CrossRef]
  34. Tannaz, A.; Mousa, S.; Sabalan, D.; Masoud, P. Fusion of multi-modal medical images using nonsubsampledshearlet transform and particle swarm optimization. Multidimens. Syst. Signal Process. 2020, 31, 269–287. [Google Scholar] [CrossRef]
  35. Yin, M.; Liu, X.; Liu, Y.; Chen, X. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampledshearlet transform domain. IEEE Trans. Instrum. Meas. 2018, 68, 49–64. [Google Scholar] [CrossRef]
  36. Ouerghi, H.; Mourali, O.; Zagrouba, E. Non-subsampled shearlet transform based MRI and PET brain image fusion using simplified pulse coupled neural network and weight local features in YIQ colour space. IET Image Process. 2018, 12, 1873–1880. [Google Scholar] [CrossRef]
  37. Wadhwa, P.; Tripathi, A.; Singh, P.; Diwakar, M.; Kumar, N. Predicting the time period of extension of lockdown due to increase in rate of COVID-19 cases in india using machine learning. Mater. Today Proc. 2020, 37 Pt 2, 2617–2622. [Google Scholar] [CrossRef]
  38. Dhaka, A.; Singh, P. Comparative Analysis of Epidemic Alert System Using Machine Learning for Dengue and Chikungunya. In Proceedings of the Confluence 2020 10th International Conference on Cloud Computing, Data Science and Engineering, Noida, India, 29–31 January 2020; pp. 798–804. [Google Scholar] [CrossRef]
  39. Diwakar, M.; Tripathi, A.; Joshi, K.; Sharma, A.; Singh, P.; Memoria, M.; Kumar, N. A comparative review: Medical image fusion using SWT and DWT. Mater. Today Proc. 2020, 37 Pt 2, 3411–3416. [Google Scholar] [CrossRef]
  40. Dhaundiyal, R.; Tripathi, A.; Joshi, K.; Diwakar, M.; Singh, P. Clustering based multi-modality medical image fusion. J. Phys. Conf. Ser. 2020, 1478, 012024. [Google Scholar] [CrossRef]
  41. Diwakar, M.; Singh, P.; Shankar, A. Multi-modal medical image fusion framework using co-occurrence filter and local extrema in NSST domain. Biomed. Signal Process. Control 2021, 68, 102788. [Google Scholar] [CrossRef]
Figure 1. Multimodality medical image fusion proposed framework.
Figure 1. Multimodality medical image fusion proposed framework.
Diagnostics 13 01395 g001
Figure 2. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Figure 2. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Diagnostics 13 01395 g002
Figure 3. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Figure 3. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Diagnostics 13 01395 g003
Figure 4. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Figure 4. Results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Diagnostics 13 01395 g004
Figure 5. Zoomed results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Figure 5. Zoomed results of multimodality medical image fusion; (a) input multimodality medical image 1; (b) input multimodality medical image 2; (c) Zhang et al. [12]; (d) Ramlal et al. [13]; (e) Dogra et al. [14]; (f) Ullah et al. [15]; (g) Huang et al. [16]; (h) Liu et al. [17]; (i) Mehta et al. [18]; (j) proposed method.
Diagnostics 13 01395 g005
Table 1. The Comparative analysis in terms of performance metrics.
Table 1. The Comparative analysis in terms of performance metrics.
ParameterDatasetZhang et al. [12]Ramlal et al. [13]Dogra et al. [14]Ullah et al. [15]Huang et al. [16]Liu et al. [17]Mehta et al. [18]Proposed Method
Mutual information (MI)#12.12982.78493.05123.48103.39523.29673.17193.4917
#22.79723.11682.47572.57882.25343.12703.56703.7710
#32.56102.65132.83152.41032.41242.71092.37092.8709
#42.12682.31032.23302.61502.26122.71202.17202.8710
#52.21112.21712.32122.11672.10192.14182.11782.6418
Standard deviation (SD)#166.212281.019175.905384.252682.831081.019882.049885.0563
#258.511872.011172.811875.132576.458777.179878.144878.7798
#355.259671.219571.112471.272371.118771.227271.271072.2350
#458.555572.542271.144673.355074.244475.032072.232075.1180
#567.814166.151569.011571.521972.221771.876172.871673.2276
QAB/F#10.51010.51150.52020.51130.52180.51030.53010.5397
#20.51830.51400.51780.52180.51870.52510.52110.5288
#30.59190.61510.62810.62710.61710.63110.63510.6398
#40.61410.61170.62200.62170.64280.63630.61510.6486
#50.61710.63120.61510.62220.61230.64540.63520.6510
Spatial frequency (SF)#123.121227.751125.571026.818626.71427.011027.550427.9822
#221.111322.783322.614121.611321.542222.412322.723322.8123
#319.092621.181321.011120.181820.071820.001921.001921.3319
#417.355618.231318.111220.112219.455420.001320.131320.2923
#520.096118.832921.414018.343119.912918.239019.012021.4190
Mean#149.324958.234653.854357.120956.023857.512057.512158.5350
#244.143351.724647.444051.335652.127053.821951.340953.9609
#341.345341.123339.175341.012539.124138.124038.213442.7970
#440.468041.364339.123341.343038.325241.112241.141441.8324
#533.128236.887235.992134.450335.545333.465736.445737.0057
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Diwakar, M.; Singh, P.; Singh, R.; Sisodia, D.; Singh, V.; Maurya, A.; Kadry, S.; Sevcik, L. Multimodality Medical Image Fusion Using Clustered Dictionary Learning in Non-Subsampled Shearlet Transform. Diagnostics 2023, 13, 1395. https://doi.org/10.3390/diagnostics13081395

AMA Style

Diwakar M, Singh P, Singh R, Sisodia D, Singh V, Maurya A, Kadry S, Sevcik L. Multimodality Medical Image Fusion Using Clustered Dictionary Learning in Non-Subsampled Shearlet Transform. Diagnostics. 2023; 13(8):1395. https://doi.org/10.3390/diagnostics13081395

Chicago/Turabian Style

Diwakar, Manoj, Prabhishek Singh, Ravinder Singh, Dilip Sisodia, Vijendra Singh, Ankur Maurya, Seifedine Kadry, and Lukas Sevcik. 2023. "Multimodality Medical Image Fusion Using Clustered Dictionary Learning in Non-Subsampled Shearlet Transform" Diagnostics 13, no. 8: 1395. https://doi.org/10.3390/diagnostics13081395

APA Style

Diwakar, M., Singh, P., Singh, R., Sisodia, D., Singh, V., Maurya, A., Kadry, S., & Sevcik, L. (2023). Multimodality Medical Image Fusion Using Clustered Dictionary Learning in Non-Subsampled Shearlet Transform. Diagnostics, 13(8), 1395. https://doi.org/10.3390/diagnostics13081395

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop