Next Article in Journal
A Near-Field Imaging Method Based on the Near-Field Distance for an Aperture Synthesis Radiometer
Previous Article in Journal
The Dunes of Belvedere–San Marco of Aquileia: Integrating High-Resolution Digital Terrain Models and Multispectral Images with Ground-Penetrating Radar Survey to Map the Largest System of Continental Dunes of Northern Italy
Previous Article in Special Issue
Stain Detection Based on Unmanned Aerial Vehicle Hyperspectral Photovoltaic Module
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Blind Hyperspectral Unmixing Framework Based on CUR Decomposition (CUR-HU)

Department of Electrical Engineering and Centre for Intelligent Multidimensional Data Analysis, City University of Hong Kong, Hong Kong
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(5), 766; https://doi.org/10.3390/rs16050766
Submission received: 29 December 2023 / Revised: 9 February 2024 / Accepted: 20 February 2024 / Published: 22 February 2024
(This article belongs to the Special Issue New Methods and Approaches in Airborne Hyperspectral Data Processing)

Abstract

:
Hyperspectral imaging captures detailed spectral data for remote sensing. However, due to the limited spatial resolution of hyperspectral sensors, each pixel of a hyperspectral image (HSI) may contain information from multiple materials. Although the hyperspectral unmixing (HU) process involves estimating endmembers, identifying pure spectral components, and estimating pixel abundances, existing algorithms mostly focus on just one or two tasks. Blind source separation (BSS) based on nonnegative matrix factorization (NMF) algorithms identify endmembers and their abundances at each pixel of HSI simultaneously. Although they perform well, the factorization results are unstable, require high computational costs, and are difficult to interpret from the original HSI. CUR matrix decomposition selects specific columns and rows from a dataset to represent it as a product of three small submatrices, resulting in interpretable low-rank factorization. In this paper, we propose a new blind HU framework based on CUR factorization called CUR-HU that performs the entire HU process by exploiting the low-rank structure of given HSIs. CUR-HU incorporates several techniques to perform the HU process with a performance comparable to state-of-the-art methods but with higher computational efficiency. We adopt a deterministic sampling method to select the most informative pixels and spectrum components in HSIs. We use an incremental QR decomposition method to reduce computation complexity and estimate the number of endmembers. Various experiments on synthetic and real HSIs are conducted to evaluate the performance of CUR-HU. CUR-HU performs comparably to state-of-the-art methods for estimating the number of endmembers and abundance maps, but it outperforms other methods for estimating the endmembers and the computational efficiency. It has a 9.4 to 249.5 times speedup over different methods for different real HSIs.

1. Introduction

Hyperspectral imaging technique has attracted much attention in remote sensing, which collects reflected light of hundreds of contiguous and narrow wavelengths, having rich spectral information about the various materials in the scene. It is a powerful tool in different application fields, such as mineralogy, agriculture, surveillance, space exploration, environmental monitoring, and disease diagnosis [1,2]. There is a tradeoff between the spectral and spatial resolutions, which leads the observed signature of each pixel to be a mixture of several materials [3]. HU techniques are source separation methods that decompose each mixed pixel in HSI into a set of pure spectral signatures called endmembers and their corresponding fractions called abundance maps. More specifically, the HU process is composed of three tasks: (1) estimating the optimal number of endmembers, (2) extracting the endmember signatures, and (3) estimating the abundance maps [4]. Estimating the optimal number of endmembers, virtual dimensionality (VD) [5], is crucial because of the performance of extracting endmembers and estimating the abundance maps depending on it. Most of HU algorithms mainly focus on the last two tasks of the HU process, i.e., estimating the endmembers and abundance maps sequentially or simultaneously  [6,7], by assuming that the number of endmembers is available in advance.
The hyperspectral mixing and unmixing algorithms can be classified into linear and nonlinear models according to the assumptions of the interaction of light with the observed field of view [4,8,9]. Most existing HU algorithms are based on the linear mixing model (LMM), a straightforward and simple method for representing the mixing process. Furthermore, two physical abundance constraints are required while considering linear unmixing: the nonnegative abundance constraint (ANC) and the abundance sum-to-one constraint (ASC) [10]. The endmember extraction and abundance maps estimation of the HU process can be solved separately. Estimating the abundance maps is an ill-posed linear inversion problem, and we can solve it using the least-squares-based methods, such as fully constrained least squares (FCLS) [7]. To reduce the accumulative errors in two-stage-based HU algorithms, the statistical-based HU has been proposed, where the HU is considered a BSS problem, so the endmembers and abundance maps are estimated simultaneously [11].
Nonnegative matrix factorization (NMF) [12] is the popular method for solving BSS problems, which has been utilized in the HU process. NMF decomposes the nonnegative data matrix into a product of two nonnegative matrices: a source matrix and a mixing matrix. Although standard and constrained NMF methods have been proposed for blind HU, they do not guarantee to provide a unique solution, the decomposition results depend on the initialization, and they require high computational costs to obtain the solution because they are iterative methods. In addition, the interpretation and analysis of the resultant matrices from NMF are complex.
Because HSIs consist of relatively few prominent features (i.e., endmembers), we can approximate HSIs using the low-rank approximation method to explore those prominent features and their fraction at each pixel. Many low-rank decomposition methods exist, such as Singular Value Decomposition (SVD), QR factorization, NMF, Interpolative Decomposition (ID), CUR matrix decomposition, and Principal Component Analysis (PCA) [13]. In order to demonstrate that the HSI data are almost low-rank or can be well approximated as low-rank data, we consider a real HSI dataset known as the Samson HSI dataset [14]. It comprises 95 × 95 pixels with 156 spectral bands and has three reference endmembers (i.e., soil, tree, and water). The Samson data cub can be reshaped into an HSI data matrix of size 156 × 9025 , whose rows represent the spectral bands, and its columns represent the pixels. Figure 1a shows the Samson data matrix, which clearly can be well approximated as low-rank data and constructed from a few components. Figure 1b shows the ground truth of normalized endmembers and their abundance maps of the Samson HSI dataset. Moreover, we selected 15 pixels randomly from the HSI data matrix and visualized their reflectance along the spectral bands, as shown in Figure 1c. As can be seen, the 15 signatures can be represented within three groups (i.e., solid, dash, and dash-dot lines), where each member in a group is a linear combination of the group members. In other words, each pixel of the HSI dataset can be represented as linear combinations of the main component of each group. Thus, the HSI is almost a low-rank and well-approximated by a low rank, and the rank here represents the number of endmembers.
Inspired by the CUR matrix factorization methods [15,16] for finding the low-rank approximation, interpreting the resulting matrices from the decomposition, and their performance in various application areas, we propose a novel framework for performing the main three tasks of the HU process based on CUR factorization called CUR-HU. It is a new blind unmixing method incorporating CUR matrix decomposition based on a deterministic sampling method for selecting the columns and rows from the observed mixed HSIs that have the maximum independence and provide a unique solution. CUR-HU automatically estimates the endmembers and their abundance maps without requiring prior knowledge about the scene. Unlike NMF, CUR-HU does not depend on the initialization method, and the resulting matrices are easy to interpret as they contain the same data element as the original given HSIs. To decrease the complexity of computing the SVD of an HSI, which the sampling method requires computing the selected indices of rows and columns, the QR factorization based on an incremental approach is utilized to approximate the SVD and estimate the number of endmembers in a given HSI. Figure 2 shows an overview of the proposed framework (i.e., CUR-HU) for performing the whole HU process. The performance of the proposed method in each stage of the complete HU process is evaluated using synthetic and real standard HSI datasets, with a comparison with several selected state-of-the-art HU baselines. The experiments demonstrate that CUR-HU outperforms other state-of-the-art blind HU algorithms, especially for extracting the endmember spectra and the computational efficiency. CUR-HU has a 9.4 to 249.5 times speedup over different state-of-the-art methods for estimating the endmembers and their abundance maps of different real HSI datasets.
The contributions of this paper are summarized as follows:
  • We propose a new framework based on CUR matrix factorization called CUR-HU for performing the complete HU process. To the best of our knowledge, this is the first time that CUR decomposition has been applied to solve the HU problem.
  • CUR-HU incorporates several techniques to solve the blind HU problem with unmixing performance comparable to state-of-the-art methods but with much higher computational efficiency. It does not require initialization or prior information about the scene.
  • Various experiments on synthetic and real HSIs were conducted, and the results show that CUR-HU performs comparably to state-of-the-art methods for estimating the number of endmembers, extracting the abundance maps, and robustness to spectral variability. It outperforms other methods for estimating the endmembers and the computational efficiency.
  • CUR-HU can be used to initialize the existing HU algorithms, especially iterative methods that are sensitive to the initialization step. Furthermore, it can provide supervised information for methods based on deep learning techniques.
  • CUR-HU has a 9.4 to 249.5 times speedup over different state-of-the-art methods for estimating the endmembers and their abundance maps of different real HSI datasets.
The rest of the paper is organized as follows. Section 2 presents the related work for solving the HU problem. Section 3 briefly introduces the concepts and methods used in our proposed method, including CUR decomposition, selection method, NMF, and the linear mixture model (LMM). Section 4 presents the proposed HU method, which is called CUR-HU. Section 5 provides extensive experiments and performance comparison with the state-of-the-art methods using both simulated and real HSIs. Finally, the conclusion and discussion of our future work are presented in Section 6.

Notation

In this paper, the lower and upper case bold are used for representing vectors and matrices, respectively, and x i j is the element in i-th row and j-th column of matrix X . We follow Matlab style notation, i.e., X ( I , J ) is the submatrix which contains the rows and columns of X indexed by I and J. The SVD of X = W Σ V T in which W R m × m and V R n × n are the left and right singular matrices, and Σ R m × n contains the singular values of X . The Frobenius norm of a matrix is X F and the Moore–Penrose inverse of X is X . The identity matrix of size m × m is represented by I m .

2. Related Work

Various methods have been proposed for estimating the number of endmembers. They can be categorized into two groups [17]: (1) The eigen analysis-based methods, e.g., HFC, NWHFC [5], and HySime  [18]. The HySime method tries to find the optimal signal subspace by minimizing the mean-square error (MSE) between the projected data onto the subspace and the original data, where the number of endmembers is equal to the optimal dimensionality of the subspace. NWHFC and HFC methods aim to find the most significant eigenvalue gap between the eigenvalues of the sample correlation and covariance matrices, which determines the appropriate number of eigenvectors that can represent the HSIs data. (2) The information-theoretic criteria-based algorithms, e.g., the Akaike information criterion (AIC) and the minimum description length (MDL) [19]. Many algorithms have been proposed to solve the HU problem, and they can be divided into three classes according to the computational methods: (1) geometric algorithms, (2) sparse regression-based algorithms [4], and (3) statistical algorithms. Geometrical methods utilize the orientation of HSIs by assuming that the endmembers of HSI are the vertices of a simplex with the minimum or maximum volume enclosing the data set or contained in the convex hull of the data set, respectively. The classical geometrical algorithms include N-FINDR [20], the pixel purity index (PPI) [21], the vertex component analysis (VCA) [6], the simplex growing algorithm (SGA) [22], and simplex identification via split augmented Lagrangian [23]. VCA and N-FINDR are iteratively searching about the simplex with maximum volume, whose vertices are considered endmembers. VCA is an iterative approach that projects data in the orthogonal direction to the subspace spanned by the extracted endmembers in the previous steps, while N-FINDR randomly selects some pixels as the initial endmembers matrix, then changes those pixels by other pixels to find a set of pixels with the maximum volume to extract endmembers. Although the geometrical unmixing methods have worked well for extracting endmembers, they may fail to extract endmembers from the highly mixed HSI data, especially if the pure spectral signatures do not exist. Sparse regression-based methods use large known libraries to formulate the HU problem as a sparse linear regression problem [24,25]. Extracting realistic endmembers directly from the HSIs data is a big challenge without prior knowledge. Because more large spectral libraries have become openly available, many algorithms have been proposed to select the best subset of endmembers signatures from those libraries to represent each pixel by considering the spectral library as a priori knowledge [26]. This approach aims to use only a few spectra in a given spectral library to model each mixed pixel in the HSI. It is called sparse unmixing because the number of endmembers in a given HSI is much smaller than the available spectra in the library, which leads to a sparse solution. The authors in [26] proposed a sparse unmixing using spectral a priori information unmixing method that uses the spectral a priori information in the HSI to address the high mutual coherence problem of the spectral library. However, extracting the endmembers from spectral libraries is usually inconsistent with HSI pixel signatures because of acquisition conditions, affecting the subsequent endmember extraction step, i.e., abundance estimation.
Statistical methods of HU use statistical representation to interpret the mixed pixels. They are powerful techniques for decreasing the accumulative errors in the two-step-based HU methods because they consider the HU as a BSS problem to estimate the endmember signatures and extract the abundance maps simultaneously.
Independent component analysis (ICA) [27,28], Bayesian self-organizing maps [29], independent factor analysis (IFA) [30], and nonnegative matrix factorization (NMF) [12] are some of the common statistical methods utilized for HU. HU based on ICA and IFA methods has been proposed based on two assumptions: the observed spectrum vector should be a linear of the endmember signatures weighted by the correspondent sources (abundance fractions) and the sources should be statistically independent. Thus, the HU itself is not a strict ICA or IFA problem. The authors in [30] claim that the ICA does not play a competitive role in HSI because the statistical independence of the sources, i.e., assumed by ICA and IFA methods, is violated in the HU, compromising the performance of ICA and IFA methods for solving HU. As a result, nonparametric statistical unmixing methods based on the dependent component analysis are proposed in [31,32,33]. It models the abundance maps as mixtures of Dirichlet densities, which enforces the non-negativity and constant sum constraints on the abundance fractions. The mixing matrix is estimated using a generalized expectation–maximization (GEM)-type algorithm. Although this method provides a reasonable approach to avoid the independence assumption of ICA, the Dirichlet densities of the abundances need to be acquired in advance. Moreover, a matrix inversion operation is required in each iteration step, which may lead to performance degradation. Compared with the sparse regression-based methods and geometrical algorithms, the NMF-based HU algorithms are powerful and widely used to extract the endmembers and their abundance maps simultaneously [34].
NMF-based algorithms work well for solving HU, but they are ill-posed geometric algorithms, meaning that the solution may not be unique. Various auxiliary regularizers have been introduced to add to conventional NMF to improve the uniqueness of the HU solution, e.g., minimum volume rank deficient NMF (Min-vol NMF) [35], l 1 / 2 -sparsity constrained NMF ( l 1 / 2 -NMF) [36], manifold regularized sparse NMF [37], spatial group sparsity regularized NMF (SGSNMF) [38], robust collaborative NMF (R-CoNMF) [39], graph regularized NMF [40], and subspace structure regularized NMF (SSRNMF) [41]. The SGSNMF method incorporates a spatial group sparsity regularizer into the NMF-based HU method to explore the sparse structure and the use of spatial information of HSIs. It exploits the idea that the pixels within a local spatial are expected to have the same pattern in the low-rank matrix of the abundance map. For l 1 / 2 -NMF, the l 1 / 2 -norm is used instead of the l 1 -norm to obtain a sparse solution. Because the l 1 / 2 regularization provides high unmixing performance, it is used in different NMF-based HU methods, such as structure-constrained sparse NMF [42] and graph regularized L1/2-NMF [37]. Recently, nonnegative tensor factorizations have been introduced to solve the blind HU problem, e.g., Matrix-Vector Nonnegative Tensor Factorization for Blind HU (MVNTF) [43], which preserves the spatial and spectral information to perform the factorization on the hyperspectral alternative 2D image. Moreover, HU methods based on deep learning (DL) have been proposed to achieve more competitive unmixing performance [44,45,46]. However, there are still several drawbacks, i.e., requiring many training samples and network parameters to achieve satisfactory performance, high computational complexity, and difficulty interpreting the results from the network [47,48]. Most existing HU methods consider only one or two tasks of the HU process, i.e., estimating the end members and their abundance and assuming that the number of endmembers is available, which is uncommon in real scenarios. In this paper, we consider the whole HU process.

3. Preliminaries

This section introduces the primary methods we build on, the Linear Mixture Model (LMM), NMF, and CUR or Skeleton matrix factorization.

3.1. Linear Mixture Model (LMM)

In this paper, we consider the LMM model [3], in which each light ray interacts with only an endmember in the scene before reaching the imaging sensor. The spectrum at each pixel in HSI is modeled as a linear combination of all endmembers. Assume that endmembers exist in the HSI dataset and the observed spectral data vectors are given by
y = x + e
where x R L × 1 and e R L × 1 are the signal and additive noise, respectively, and L is the number of spectral bands. The noise here includes endmember variability, model inadequacies, and sensor noise  [49,50]. The signal vectors lie on an unknown p dimension subspace and are given by
x = k = 1 p m k s k = Ms
where M R L × p is a full rank matrix with a rank equal to p, where p is the number of endmembers in HSI, and p < L . Each column of M contains an endmember signature of HSI, and s R p × 1 is the abundance vector. Let S = s 1 , s 2 , , s N R p × N represent abundance maps matrix. The LMM model can be formulated in a matrix form as
Y = MS + E
where Y = y 1 , y 2 , , y N R L × N is the HSI data matrix, and E R L × N is the additive noise matrix, where N is the number of pixels. The abundance maps matrix should obey positivity and sum-to-one constraints to be physically meaningful.

3.2. Nonnegative Matrix Factorization

The LMM formulation of the HU problem can be considered a BSS problem to extract the endmember spectra and the abundance maps simultaneously. NMF can be used to solve the blind HU problem as it decomposes a given nonnegative data matrix V R + n × m into two other nonnegative matrices W R + n × r (the source matrix) and H R + r × m (the mixing matrix) which satisfy the following approximation.
V WH
An infinite number of H and W pairs satisfy the above factorization. For example, WH = W Γ 1 ( Γ H ) for any invertible Γ R + r × r matrix. The conventional procedure of NMF is to formulate the factorization using an objective function that quantifies the quality of approximation between V and WH , and then, an optimization algorithm is used to minimize or maximize the defined objective function concerning H and W . One of the widely utilized objective functions for NMF is the square of the Frobenius norm between the given data matrix and decomposed matrices as follows,
V WH F 2 = i j V i j ( W H ) i j 2 .
Although the above equation is convex in H and W alone, it is not convex in H and W together [12]. Therefore, finding the global minima of this function w.r.t. H and W analytically is difficult. It is possible to find local minima numerically using optimization methods.
The HU problem can be represented as a conventional NMF form [12,39,51]. Let the HSI matrix Y be V , the endmember matrix M is the source matrix, and the abundance matrix S is the mixing matrix of NMF. Then, the HU problem can be formulated as
arg min M , S Y MS F 2 , s . t . M , S 0
Although various auxiliary regularizers on M and S have been proposed to improve the uniqueness of the solution, the NMF-based HU methods still have some drawbacks, such as the initialization methods, the difficulty of driving the objective function, requiring extensive computational time to solve the optimization problem, the resultant solution is not unique, and the difficulty of interpreting the resultant matrices concerning the original HSIs.

3.3. CUR Matrix Factorization

Low-rank matrix approximations based on the selected columns and rows from a given data matrix naturally provide interpretable results. They have been successfully applied in various application areas. CUR matrix factorization or CUR decomposition is a widely-used technique representing the data matrix as a product of three small submatrices of original data [15,16]. A visual representation of CUR matrix decomposition is shown in Figure 3. CUR of a matrix A R m × n is a low-rank approximation that can be represented as
A C U R
where C R m × c and R R r × n matrices are constructed by extracting c columns and r rows of A , respectively. The matrix U should be formed so that the CUR approximation error A CUR F is very small. CUR matrix decompositions give more interpretable results because they represent data in lower dimensional spaces using actual data points as basis vectors that maintain sparseness or nonnegativity properties, unlike the other matrix factorizations. It originated with a pseudoskeleton and truncated QR approximation [52,53].

4. CUR Factorization for Hyperspectral Unmixing (CUR-HU)

This section presents the proposed framework (i.e., CUR-HU) for performing the main three tasks of the HU process. A general assumption is that a low-dimensional subspace is hidden in the HSI dataset, as demonstrated in Figure 1. The endmember signatures of the given HSI represent the macroscopic objects of the HSI scene, e.g., water, soil, and vegetation. HU algorithms attempt to extract those macroscopic objects from the observed mixed signals. As those endmember signatures are statistically independent, extracting them from the mixed observed signals can be achieved by finding a group of pixel spectra with maximum independence between them. In other words, the extracted endmember signatures should be more mutually independent than the mixed pixel spectra. Therefore, such a methodology would be a progression toward extracting more realistic endmember spectra.
According to the LMM, the spectrum of each pixel of HSI is modeled as a linear combination of all endmembers, so the HSI data matrix has a low-rank structure with a rank almost equal to the number of endmembers. To explore the low-rank pattern of HSIs, we can project the pixels onto low dimensional subspace as follows. Let the HSI matrix be Y = [ y 1 , , y N ] R L × N with a span dimension r. Then the basis of the reduced dimension k < r is a set of basis vectors { ϕ } i = 1 k R L , which provides the best approximation to the original span of Y . The set of basis vectors { ϕ } i = 1 k is constructed to minimize error as follows;
min ϕ i i = 1 k j = 1 N y j i = 1 k y j T ϕ i ϕ i 2 2 , with ϕ i T ϕ t = 1 if i = t , i , t = 1 , , k . 0 if i t ,
The set of the left singular vectors of Y gives the optimal solution to this minimization problem. Let the SVD of the HSI data be Y = W Σ V T , where W R L × r and V R N × r are left and right singular vectors, respectively, and Σ R r × r , with order as σ 1 σ 2 σ r > 0 . So the optimal basis vectors for solving are w i i = 1 k . Although this is an optimal solution, it is limited to approximating the linear or low-order polynomial nonlinearity functions. It is computationally expensive and difficult to physically interpret the resultant singular vectors of a given HSI dataset.
We propose to utilize the CUR matrix decomposition to find the hidden structure of HSI datasets. Therefore, HSI can be represented using CUR matrix decomposition, such that for a given HSI matrix Y R + L × N , as Y C U R , where C R L × p contains the most informative p columns that can represent the columns space of the HSI matrix Y (i.e., a subset of p pixels that can represent the whole N pixels of HSI), R R p × N contains the most informative rows from Y , and U R p × p which makes the difference between data matrix Y and low-rank approximation matrix CUR as small as possible. Analog to Equation (6) of NMF-based HU methods, the HU problem is formulated based on CUR decomposition as follows.
arg min C , U , R Y CUR F 2
The C matrix contains the endmembers of HSI, and the U R matrix represents the abundance matrix whose columns represent the fractional compositions at each of the N pixels. To construct C and R matrices, randomized and deterministic sampling approaches can be used. Randomized approaches are based on sampling columns and rows based on uniform or non-uniform distributions [15,54]. Deterministic approaches consider the matrix approximation as an optimization problem and try to find the best decomposition that minimizes the approximation error [55]. They provide superior accuracy when reconstructing the matrix but have higher computational costs. In this paper, we utilized a recent deterministic selection method, the discrete empirical interpolation method (DEIM), because it gives high approximation accuracy, provides the most informative columns and rows approximating the whole data matrix, and approximates the nonlinear part by considering the continuous domain as a finite set of discrete points in the continuous domain  [55,56]. For more details about using the DEIM method to approximate the nonlinear function, please refer to Appendix A.2.
The DEIM method is completely deterministic and involves few parameters to provide the selected rows and columns indices of the HSI matrix Y based on the knowledge of the SVD. Suppose the truncated SVD Y W r Σ r V r T is computed, then the DEIM algorithm selects the best rows and columns indices q and p by processing the basis vectors W and V , respectively, to construct the R and C matrices. Given an orthonormal matrix W R L × r and a set of distinct indices p R r , then the interpolatory projector for indices p onto the column span of W is
P W P T W 1 P T
where P = I L ( : , p ) R L × r , and the P T W R r × r is invertible. This projector P has a critical characteristic not generally available by orthogonal projectors: for any vector s R L ,
( P s ) ( p ) = P T P s = P T W P T W 1 P T s = s ( p )
The original vector s and its projected vector P s match in the p elements, so this projector is called an interpolatory projector. DEIM processes each vector of the left singular vectors W = [ w 1 , , w r ] . Starting from the first singular vector w 1 , which corresponds to the largest singular value. The first index p 1 is selected as the index of the largest magnitude element in w 1 , as
w 1 p 1 = w 1
Now, p contains only the first slected index p 1 as p p 1 . The interpolatory projector of p 1 onto w 1 is given by
P 1 w 1 P 1 T w 1 1 P 1 T
The second index p 2 is selected as the index of the largest element in w 2 but after removing its interpolatory projection in the w 1 direction.
r 2 = w 2 P 1 w 2 , r 2 p 2 = r 2 .
Because the P 1 w 2 and w 2 are matched in the p 1 position, the r 2 p 1 = 0 . So there are non-repeated indices. Generally, for selecting the index p i , the i 1 indices are available and the interpolatory projection matrix is given by
P i 1 W i 1 P i 1 T W i 1 1 P i 1 T
where W i 1 is the first i 1 left singular vectors, W i 1 = w 1 w i 1 . Then, the index p i is computed as
r i w i P i 1 w i , r i p i = r i
The process of selecting the column indices q is the same above steps to the right singular vectors V . Algorithm 1 shows the procedures to compute the selected row indices p by the DEIM algorithm.
Algorithm 1: Discrete empirical interpolation method algorithm
Remotesensing 16 00766 i001
The interpolation indices p used for constructing the R matrix of CUR decomposition are selected inductively from the basis vectors of U . After processing all singular vectors and obtaining the row indices vector p , the R matrix is constructed as R = Y ( p , : ) . Similarly, obtain q and construct the C matrix as C = Y ( : , q ) .

4.1. Approximating the Singular Vectors

The main limitation of the selection method is requiring the top r right and left singular vectors. For moderate-size HSI, software libraries can be used to compute SVD. Unfortunately, this is inefficient for large-scale HSI. Therefore, iterative approaches can be used to compute the leading r singular vectors of HSI [53,57,58]. We utilize the QR factorization based on an incremental approach (IQR) to approximate the computation of the singular vectors and estimate the number of endmembers in the HSI. The main advantages of IQR are that (1) it is only required to pass once through the HSI matrix Y for approximating the SVD, leading to efficient storage of massive data sets in memory; and (2) it satisfies a deterministic error bound in approximating the SVD of low-rank data. IQR approximates a matrix A R m × n as A Q R , where Q R m × k that contains k orthonormal column vectors and R R k × n is the upper triangular matrix.
Based on eigenanalysis algorithms of estimating the number of endmember signatures of HSI and considering the HSI can be well approximated using a low rank, we can utilize the QR factorization method to estimate the rank of the HSI data matrix, which represents the number of main components that approximate the HSI [5,17,59]. We use the QR factorization based on incremental methods for two tasks: (1) approximating the singular vectors required by the sampling algorithm and (2) estimating the number of endmember signatures of HSI data. By using the IQR, Algorithm A1 in Appendix A, we approximate the HSI matrix Y as Y Q R , where Q R L × p that contains the orthonormal column vectors, R R p × N is the upper triangular matrix, and p is the estimated number of endmembers. IQR processes one pixel of Y at each step to make it orthogonal with the previously orthogonalized pixels. After the orthogonality step, the algorithm verifies if any row of R has a tiny relative norm. If there is such a row, then this row of R and the corresponding column of Q can be deleted because this column of Q has a small contribution to the cumulatively computed Frobenius norm of R .
At the start, we select the first k pixels from the HSI matrix Y which denoted as Y k = y 1 , , y k R L × k , and let k = 2 , as an initial step. Then, Y k = Q k R k , with Q k R L × k and R k R k × k . In step 2 of the algorithm, the norm of rows in the R is computed. According to its value, we can determine which row should be deleted by incrementing the value of k to obtain the next pixel (column) of Y . To make the pixel y k + 1 orthogonal with previously k orthogonalized pixels, the classical Gram–Schmidt operations can be applied as
r = Q k T y k + 1 , f = y k + 1 Q k r
For a robust and stable re-orthogonalization step, the classical Gram–Schmidt is replaced by the following steps (7–8) of Algorithm A1, as proposed in [60]. This modification provides a Q matrix which is numerically orthogonal to working precision, and it can be implemented in an efficient parallel way as it utilizes the Gram–Schmidt process. see [61] for the full analysis.
The output of the above steps is the orthonormal vector q k + 1 and the corresponding r k + 1 . Then, the Q and R matrices are updated, as formulated in the algorithm as
Q = [ Q k , q k + 1 ] ; R = R r k + 1 0 ρ
Next is checking whether there is a truncation step by checking the row of the minimum relative norm in R . Suppose this row has a relative norm less than the tolerance parameter (i.e., t o l , a positive value controlling the accuracy of decomposition). In that case, the corresponding column in Q has little contribution to the factorization. The truncation step is required here by deleting that row from R and replacing it with the last row corresponding to the index k + 1 of R the matrix. Similar to how the corresponding column of Q will delete and put it in its position in the last column of Q matrix. If the minimum relative norm of rows in R does not achieve the truncation condition, then the last column in Q , which is q k + 1 , is accepted as an orthonormal vector, and the process of the algorithm is continuing to obtain the next pixel from Y . According to the linear model (LMM), each pixel in HSI data is a linear combination of endmember signatures in the HSI scene. So the number of column vectors in Q is the minimum number of the subspace dimensions that can approximate the HSI data, representing the number of endmembers in the HSI scene (i.e., p = k ). In general, the estimated number of endmembers of HSI Y using incremental QR factorization within a tolerance error t o l is p = N d n , where d n is the number of times there is satisfaction of the truncation condition (number of deletions that occurred up to the last iteration) of Algorithm A1. The tolerance parameter t o l plays a vital role in estimating the number of endmembers. A large tolerance value will lead to a smaller endmembers number than the actual number but will require less storage and computing time. Conversely, for a smaller t o l , the estimated number of endmembers is equal to or larger than the actual number, and the complexity becomes significant. For more details about the complexity of approximating the singular vectors using the QR factorization, please refer to Appendix A.1.

4.2. CUR-HU

In this section, we present the proposed HU framework (i.e., CUR-HU) for performing the entire unmixing process (i.e., estimating the number of endmembers, estimating the endmembers, and extracting the corresponding abundance maps). As shown in Figure 2, CUR-HU consists of three main stages, preprocessing, estimating the number of endmembers, and CUR decomposition. Algorithm 2 shows the pseudocode of the CUR-HU framework.
In the preprocessing stage, the input HSI cube is reshaped to a matrix (i.e., HSI observed data matrix Y ), and then, the noise of HSI is estimated to obtain the estimated HSI matrix X . Various methods for noise estimation have been proposed in the signal processing area and utilized in hyperspectral imagery, such as the nearest neighbor difference (NND) [62] and multiple regression theory [63]-based methods. The most straightforward technique is the NND method, which assumes that the signal component of the adjacent pixels is almost equal. It is also known as the shift difference method because for a given two adjacent observed vectors y 1 = x 1 + n 1 and y 2 = x 2 + n 2 in a homogenous area of the HSI, the covariance matrix of noise can be estimated using the difference between y 1 and y 2 , which yields y 1 y 2 = x 1 + n 1 x 2 + n 2 n 1 n 2 , where x 1 , x 2 and n 1 , n 2 are signal and noise vectors, respectively. However, noise samples are independent and almost have the same statistics. In this paper, we utilize the multiple regression theory-based method [5,63]. It outperforms the shift difference technique because of the high correlation of the neighboring spectral bands in HSIs. The second stage is performing the QR factorization based on an incremental approach, as presented in Algorithm A1 in Appendix A. This stage is responsible for two tasks: (1) approximating the singular vectors for the sampling algorithm and (2) estimating the number of endmembers as presented in Section 4.1, as shown in lines 4-7 of Algorithm 2. Suppose the number of endmembers is available by any other existing method. In that case, this stage only approximates the top r singular vectors of the HSI data matrix. Then, the CUR-HU can be used to estimate the endmembers and the corresponding abundance maps. Therefore, CUR-HU has much flexibility to cooperate with the existing algorithms at each stage.
In the third stage, the endmembers and their abundance maps are estimated using CUR decomposition. The inputs to this stage are the estimated number of endmembers and singular vectors of HSI data by the second stage and the HSI matrix X . At the start, the sampling algorithm (DEIM) processes the left and right singular vectors to generate the rows and columns indices for constructing the C and R matrices of CUR decomposition, where C contains the most informative columns that can represent all pixels of HSI and R contains the most informative rows (i.e., selected bands) over the whole pixels, as shown in lines 8–11 of Algorithm 2. The C and R matrices are unique, as the deterministic sampling algorithm provides unique indices of selected rows and columns of CUR. After getting the C and R matrices, the U matrix is constructed to make the approximation error as small as possible [64].
arg min U Y CUR F 2
Different methods for constructing the U matrix have been proposed. The simplest method is using the pseudoinverse of the intersection matrix between C and R matrices. Given C R L × p and R R p × N , then U is computed as, U = Y ( I , J ) . Although this is the simplest solution of U , it gives a high approximation error in approximating high-rank or noisy data. The close form for the best middle matrix of CUR decomposition regarding the Frobenius norm error is given by solving the optimization problem in (19), so the optimal U matrix is given by
U = C Y R
So, Y C U R = C C Y R R , where C C and R R are orthogonal projections onto the span of the columns of C and rows of R , respectively. Therefore, using U = C Y R gives the best CUR approximations for given C and R matrices. In other words, the U matrix is formed by weighting the whole elements of Y , which leads to a small approximation error [65].
Analog to the NMF-based methods for HU, the matrix C represents the source matrix of NMF or M matrix of HU LMM, while the U R matrix represents the mixing matrix of NMF or S matrix in Equation (6). Finally, the nonnegative abundance (ANC) and the abundance sum-to-one (ASC) constraints are imposed on the U R matrix. For ANC, we set any negative value of U R to zero, while for ASC, we normalized the output from ANC constraint to U R .
Algorithm 2: The pseudocode of CUR-HU
Input: HSI cube Y R m × n × L , t o l a positive number for controlling the factorization accuracy.
Output: Number of endmembers p, Endmember matrix M R L × p , and Abundance maps S R p × N
Preprocessing:
1Reshape Y into Y R L × N ;//  ( Where N = m × n . (
2Estimate the noise matrix N R L × N .
3Compute the estimated HSI, X = Y N .
Second Stage:
4 [ Q , R , p ] IQR ( X , t o l ) .//  ( Where Q R L × k ( .
( /*  ( Compute the singular vectors ( */
5 [ W ^ , Σ , V ] SVD ( R ) ;//  ( Economy-sized SVD ( .
6Set W Q W ^ ;
Thrid Stage:
/*  ( Get Row and column indices ( */
7 I DEIM ( U ) ; Using Algorithm 1.
8 J DEIM ( V ) ; Using Algorithm 1.
/*  ( Construct C, U, and R matrices ( */
9 C X ( : , J ) ;//  ( Where C R L × p ( .
10 R X ( I , : ) ;//  ( Where R R p × N ( .
11Set U = C Y R ;//  ( Where U R p × p ( .
/*  ( Construct M and S matrices ( */
12Set M C ;    and     S U R ;
13Apply the abundance nonnegativity constraint to S ;
14Apply the abundance sum to one constraint to S ;

5. Experiments

For a comprehensive and quantitative comparison between the different algorithms, the proposed method is evaluated on synthetic and real HSIs datasets. To evaluate the performance of each stage of the CUR-HU framework, different datasets with different experimental settings were conducted. All experiments were implemented with MATLAB R2022a on a desktop computer with an Intel(R) Core(TM) i7-10700 processor @ 2.90 GHz.

5.1. Estimating the Endmember Number

CUR-HU is evaluated for the first task HU process, which is estimating the number of endmembers in the HSI data. In order to evaluate the QR factorization based on an incremental approach for estimating the number of endmember signatures, we compare the performance with state-of-the-art methods, including HySime [18], HFC, NWHFC [5], SPICE [66], FUN [59], and GENE-AH [17] for synthetic and real HSI datasets.

5.1.1. Synthetic HSI Dataset

For HSI synthetic data, we follow the setup of experiments of the HySime method. The endmember signatures are selected randomly from the USGS spectral library (https://www.usgs.gov/labs/spectroscopy-lab/science/spectral-library, accessed on 28 December 2023), and the abundance maps are created according to a Dirichlet distribution [18]. To evaluate our proposed algorithm in a noisy environment, additive Gaussian noise is added to the generated HSI datasets. Let the diagonal elements of the noise correlation matrix follow a Gaussian shape centered at L / 2 , i.e.,
σ i 2 = σ 2 e ( i L / 2 ) 2 2 η 2 j = 1 L e ( j L / 2 ) 2 2 η 2
for i = 1 , , L . Here, the parameter η controls the variance of the Gaussian distribution when η corresponds to white noise, η 0 corresponds to one-band noise, and the parameter σ i 2 controls the total power of noise. The amount of additive noise to HSI is set by the signal-to-noise ratio (SNR), which is defined as
SNR = 10 log 10 E x T x E ε T ε
where x and ε are the signal and noise, respectively. We evaluated the proposed method under changing SNR { 50 , 35 , 25 , 15 } dB and the number of endmembers p  { 3 , 5 , 10 , 15 } . The NWHFC is an extension method of the HFC by adding noise whitening as a preprocessing step. We report the results of both estimated and actual noise correlation matrices with the NWHFC method. The result based on the noise correlation matrix is shown in brackets. The false-alarm probability p f parameter is set as p f 10 3 , 10 4 , 10 5 for both HFC and NWHFC methods. We report the tolerance parameter ( t o l ) that gives the best estimation of the number of the endmembers.
Table 1 shows the estimated dimension of the signal subspace by different methods as a function of endmembers number p and SNR for one-band noise η = 0 . Although the HySime method outperforms the other methods, our proposed method has a comparable performance of estimating the number of endmember signatures with the HySime method, especially for high SNR cases. For lower subspace dimensions, such as p = 3 and p = 5 , the proposed algorithm can estimate the exact number of endmembers under different types of noise and different SNR values. It outperforms NWHFC, SPICE, HFC, GENE-AH, and FUN. The proposed methods perform well for low SNR values (i.e., 25 dB and 15 dB) when setting the tolerance parameter t o l to large values (e.g., 1 × 10 2 or 1 × 10 3 ). Overall, the QR factorization based on an incremental approach can estimate the number of endmembers with a comparable performance to the HySime method.
Table A1 shows the estimated dimension of the signal subspace by different algorithms as a function of endmember number p and SNR for η = 1 / 18 . For the noisy case ( η = 1 / 18 ), the HySime method outperforms all other methods. Our proposed method outperforms the NWHFC, SPICE, HFC, FUN, and GENE-AH methods. Furthermore, it has a performance comparable to the HySime method for estimating the number of endmembers of a given HSI. From Table 1 and Table A1, we can observe the effect of the t o l parameter of IQR factorization. For a large t o l value, the estimated number of endmember signatures is smaller than the actual number, especially for low SNR cases, while setting t o l to a very small value, the estimated number of endmembers exceeds the actual number.

5.1.2. Real HSI Dataset

The proposed algorithm is evaluated on two popular real HSI datasets, the Indian Pines site and Cuprite datasets. The Indian Pines site was collected by AVIRIS [67]. It consists of 145 × 145 pixels with 185 spectral bands. The dataset contains 16 pure materials [68]. The Cuprite dataset originally contained 224 bands. A subset of the dataset was formed after removing the low SNR bands. This subset consists of 188 spectral bands and contains 250 × 190 pixels. The ground truth of the number of endmembers of the Cuprite dataset is not known. The previous studies frequently fixed the number of endmembers between 9 and 14 [38,69]. The tolerance parameter of the proposed method is set to 1 × 10 3 . Table 2 shows the estimated endmembers number on the real HSI datasets by different methods. The proposed method is comparable to the HySime method, but it outperforms the others for estimating the closest number of endmembers to actual numbers.

5.2. Estimating the Endmembers and Abundance Maps

In this section, CUR-HU is evaluated for estimating the endmembers and their abundance by comparing its performance with state-of-the-art NMF-based HU algorithms for both synthetic and real HSI datasets. Two widely utilized evaluation metrics for HU algorithms are used to evaluate the performance of the CUR-HU, i.e., Root-Mean-Square Error (RMSE) and Spectral Angle Distance (SAD), which are computed as
R M S E i = 1 N j = 1 N S i j S ^ i j 2
S A D i = cos 1 m ^ i T m i m ^ i 2 m i 2 .
where R M S E i measures the error between the i t h ground truth abundance map S i and the corresponding extracted abundance map S ^ i , and S A D i measures the spectral angle between the i th extracted endmember m ^ i and the corresponding ground truth m i . Furthermore, the average values of RMSE and SAD are reported. For these experiments on synthetic and real HSI datasets, we follow a procedure similar to  [70,71]. We evaluated the performance of CUR-HU by comparison with the state-of-the-art blind NMF-based HU baselines, l 1 / 2 -NMF [72], R-CoNMF [39], Min-vol NMF [35], KbSNMF-div [70], SSRNMF [41], MVNTF [73], and SGSNMF [38].

5.2.1. Synthetic HSI Datasets

The synthetic HSI dataset contains three endmembers, seawater, clintonite, and sodium bicarbonate, with 480 spectral bands, which were selected from the USGS spectral library. The synthetic image was generated using the hyperspectral imagery synthesis toolbox (HSIST) (https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Imagery_Synthesis_tools_for_MATLAB, accessed on 28 December 2023). HSIST was used to generate the abundance maps with a size of 64 × 64 pixels according to the Spherical Gaussian distribution. Figure 4 shows the ground truth of the normalized endmembers and their abundance maps.
Table 3 shows the performance of CUR-HU in terms of SAD and RMSE for a synthetic HSI dataset. In order to see the effect of ANC and ASC constraints on the abundance maps from the proposed method, we report the RMSE for considering U R (i.e., S ) with ANC and U R with both ASC and ANC constraints as the abundance maps (i.e., S ), which is the final output from Algorithm 2.
As can be seen, using the ANC constraint with abundance maps has no effect because the CUR decomposition keeps the nonnegative property of the HSI data. The abundance matrices of U R with ANC and U R with both ASC and ANC constraints are donated as U R 1 and U R 2 , respectively. Figure 5 shows the estimated abundance maps by the HU-CUR algorithm. The proposed method can extract the exact abundance maps of the simulated dataset. The reason for the high performance of the proposed methods is that the synthetic HSI data consists of pure endmember signatures, which are distinguished from each other. Therefore, the CUR-HU can extract the exact endmembers as the selection method procedures select the most distinguished pixels in the HSI data. Considering the R matrix as the abundance maps without weighting it by the U matrix leads to a high estimation error, as shown in the second row of Figure 5.
In order to see the effect of selecting more rows and columns of CUR of the HSI dataset, Figure 6 shows the ground truth and the estimated endmembers by CUR-HU of the simulated dataset. The first six endmembers extracted by CUR-HU capture the three endmembers, e.g., the extracted endmembers C 3 , C 2 , and C 1 are almost the same as the ground truth endmembers G T 1 , G T 2 and G T 3 , respectively. If the number of endmembers selected by the proposed method increases, highly similar endmembers will exist (e.g., C 5 and C 6 ). Selecting only three columns from the HSI data using CUR-HU is enough to capture exactly the three endmembers of the HSI simulated data.
As can be seen, the proposed method can extract the exact abundance maps and endmember signatures. CUR-HU outperforms other methods for extracting the endmembers and their corresponding abundances of the synthetic HSI dataset.

5.2.2. Real HSI Datasets

In order to evaluate the performance of the CUR-HU method in the real scenario, a set of real HSIs, including the Samson, Cuprite, and Urban datasets from [14], were used.
  • Samson HSI dataset: It consists of 95 × 95 pixels with 156 spectral bands and has three reference endmembers. The ground truth of normalized endmembers and their abundance maps of Samson HSI are shown in Figure 1b.
  • Urban HSI dataset: After pre-processing, it consists of 307 × 307 pixels with 162 spectral bands ranging from 400 to 2500 nm and has a 10 nm spectral resolution. It has different ground truth versions. We use the one that has five endmembers.
  • The Cuprite dataset: After pre-processing, it consists of 188 spectral bands covering wavelengths ranging from 370 to 2480 nm with 250 × 190 pixels, and 12 endmembers are selected as a reference.
Figure 7 shows the extracted endmembers using CUR-HU for the Samson HSI dataset. Extracted endmembers C 3 , C 1 or C 8 , and C 2 or C 6 are almost similar to the ground truth of endmember signatures of water, tree, and soil, respectively.
Figure 8 shows the estimated abundance map of R , U R , U R 1 , and U R 2 matrices for the Samson dataset. As can be seen, the estimated abundance maps after applying both ANC and ASC constraints (i.e., U R 2 , final S from Algorithm 2) are very close to their ground truth.
Table 4 reports the quantitative comparisons regarding SAD and RMSE metrics for the Samson dataset. As can be seen, CUR-HU outperforms the method in terms of SAD for estimating the endmembers. It has the best SAD value for estimating the first endmember (i.e., soil) and the second-best value for estimating the tree and water endmembers. However, it has the best overall average performance. CUR-HU outperforms the MVNTF, l 1 / 2 -NMF, R-CoNMF, and SSRNMF methods for extracting the abundance maps in terms of the RMSE metric.
As mentioned above, the Urban HSI version that we used has five endmembers as the ground truth (i.e., Asphalt, Grass, Tree, Roof, and Dirt). Figure A1 shows the extracted endmembers using CUR-HU for the Urban dataset. As can be seen, CUR-HU can estimate accurate endmembers with almost the same shape and close to the ground truth. Figure 9 shows the estimated abundance map of R, UR, UR1, and UR2 matrices for the Urban dataset.
Table 5 reports the quantitative comparisons regarding SAD and RMSE metrics for the Urban dataset. We report the RMSE value of the estimated abundance maps (i.e., UR2), which is the final output after applying both ANC and ASC constraints. CUR-HU outperforms all other methods for extracting the endmembers of the Uraban HSI dataset, and for estimating the abundance maps, it has the best average RMSE value compared to other methods. Although the quantitative results of the RMSE values do not necessarily reflect a high estimation accuracy of abundance maps, the proposed framework can be utilized with different HU algorithms as an initialization step to provide them the initial values of the number of endmembers, endmember signatures, and abundance maps.
We only report the unmixing performance regarding the SAD values for the Cuprite dataset because the ground truth of Cuprite HSI contains only endmember signatures [14]. Figure 10 shows the extracted endmembers using CUR-HU for the Cuprite dataset. As can be seen, CUR-HU almost estimates the same shape of the ground truth endmembers.
To quantify the estimated endmembers by CUR-HU with the other blind NMF-based HU methods, Table 6 reports the SAD values for each estimated endmember and the average SAD value over the endmembers. The E M 1 to E M 12 represent the Alunite, Andradite, Buddingtonite, Dumortierite, Kaolinite 1, Kaolinite 2, Muscovite, Montmorillonite, Nontronite, Pyrope, Sphene, and Chalcedony endmember signatures of the Cuprite HSI dataset, respectively. As can be seen, the proposed method outperforms the MVNTF, l 1 / 2 -NMF, R-CoNMF, and SGSNMF methods for estimating the endmembers of the Cuprite dataset in terms of SAD value. It has comparable performance to KbSNMF-div and Min-vol NMF methods. Overall, CUR-HU can estimate the endmember and abundance map matrices with comparable accuracy to the state-of-the-art HU method based on the NMF.

5.3. Spectral Variability Point of View

In this section, we evaluate the unmixing performance of CUR-HU under the spectral variability HSI datasets. We compare the performance with the state-of-the-art methods, such as the fully constrained least squares (FCLS), the Perturbed LMM model (PLMM), the Extended Linear Mixing Model (ELMM), the Generalized LMM (GLMM) [74], and DeepGUn [75]. In all experiments, the VCA algorithm was utilized to extract the endmembers of HSIs data and initialize the methods requiring the endmembers reference matrix. The abundance maps from the FCLS algorithm were utilized to initialize all other methods. CUR-HU does not require initialization for the endmembers or abundance maps. To evaluate the performance of unmixing algorithms, we consider the Normalized Root-Mean-Squared Error (NRMSE) for the estimated abundance maps NRMSE A and reconstructed images NRMSE Y . The NRMSE between a true, generic matrix X and its estimate X ^ is computed as
NRMSE X = X X ^ F 2 X F 2

5.3.1. Synthetic HSI Datasets

For a comprehensive and quantitative comparison between the different algorithms, we generated four synthetic datasets, namely S D 0 , S D 2 , S D 3 , and S D 4 , with 50 × 50 pixels for (SD1, SD2, and SD3) and 70 × 70 pixels for S D 0 . These synthetic datasets were created using three and five 224-band endmember signatures for (SD0, SD1, SD2) and (SD3), respectively. The endmember signatures extracted from the USGS library and their abundance maps are represented in the last column of Figure 11. We adopted the generation of those datasets as used in [75,76]. The spectral variability of endmembers was imposed using different strategies for each dataset. For the SD0 dataset, we utilized the variability model based on pixel-wise multiplicative factors generated by random piecewise-linear functions as used in [77], while for SD1, we adopted the variability model used in [74], forming from band-dependent scaling factors that were smoothly varying in the spectral and spatial dimensions. For SD2, we considered a simple strategy used in [75,78] to emulate the real scenario of errors in atmospheric compensation as a function of viewing the geometry given the diffuse and direct light on the scene and the solar path transmittance. For SD3, the endmembers pure pixels of five materials (tree, asphalt, metal, dirt, and roof) were extracted manually from HSI to emulate the realistic spectral variability. Finally, white Gaussian noise was added to all synthetic datasets to set the SNR to 30 dB. The optimal parameters of each method were selected by using the grid search for each dataset. The grid search ranges of each parameter were chosen according to the ranges discussed and tested by authors in the original publications. For both GLMM and ELMM, the parameters were chosen among the following ranges: λ M , λ S [ 0.01 , 0.1 , 1 , 5 , 10 , 50 ] , λ A [ 0.001 , 0.01 , 0.05 ] , and λ Ψ , λ ψ 10 6 , 10 3 , 1 , 10 3 . For the PLMM, the grid search ranges of α , γ , and β were selected as [ 0.01 , 0.1 , 0.35 , 0.7 , 1.4 , 25 ] , [ 10 2 , 0.1 , 1 , 10 , 10 2 ] , and [ 10 9 , 10 5 , 10 4 , 10 3 ] , respectively.
The execution time of each algorithm is reported as the average running time over ten iterations. For a fair comparison, the running time performance of the FCLS method is not included because the FCLS method is used as an initialization step for other methods. We also used the VCA method to initialize and provide the number of endmember signatures for all methods.
Figure 11 shows the qualitative results of estimating the abundance maps for the synthetic datasets. For SD3 and SD4, the estimated abundance maps by CUR-HU (CURMaps) are much closer to the ground truth than the other methods, as shown in Figure 11c,d. Table 7 shows the quantitative results of the estimated abundance map, reconstruction error of the HSI, and algorithm running time of different methods for synthetic datasets. CUR-HU outperforms the other methods for estimating the abundance maps for SD3 and SD4 datasets and has comparable performance with other methods for estimating abundance maps of SD0 and SD1. Furthermore, CUR-HU has comparable accuracy for the reconstruction of the HSI data. According to the running time, our method outperforms all other methods.

5.3.2. Real HSI Data

To evaluate the performance of CUR-HU by comparing it with other methods, we considered three real HSI datasets (i.e., Samson, Houston, and Jasper Ridge datasets). The Samson dataset is the same HSI dataset that was utilized in the previous section. Houston and the Jasper Ridge HSIs were captured by the AVIRIS instrument with 224 spectral bands. After preprocessing, the Houston and the Jasper Ridge HSIs contain 188 and 198 spectral bands, respectively, and each of them has four endmember signatures as a reference [79]. The number of pixels of the Houston and the Jasper Ridge HSIs is 105 × 128 pixels and 100 × 100 pixels, respectively. Figure 12 shows the estimated abundance map of the Samson HSI dataset using different HU methods (i.e., FCLS, PLMM, ELMM, GLMM, and CUR-HU). As can be seen, the estimated abundance map from the proposed method has a clear performance improvement over the other methods, especially for the soil and water abundance maps.
Figure A2a shows the estimated abundance map of the Houston HSI dataset. Although the performance of the proposed method for estimating the oncrete abundance map is weak, it leads to a considerably stronger Met, Roofs and Vegetation abundance maps but a comparable performance for Asphalt abundance map. The estimated abundance maps for the Jasper Ridge HSI dataset are shown in Figure A2b. The performance of the proposed method is comparable with other methods for estimating abundance maps (i.e., Water, Vegetation, and Soil). Although there is confusion between Water and Road abundance maps, the estimated road region is best compared to other maps.
Table 8 shows the quantitative results for different methods on real HSIs. Because the ground truth of the abundance maps is not available, we use the reconstruction error NRMSE Y metric to evaluate the quality. CUR-HU has a comparable performance with other methods. Although the ELMM, GLMM, and PLMM methods can achieve arbitrarily small NRMSE Y , it does not necessarily reflect a high estimation accuracy of abundance maps, as shown in Figure A2a,b and Figure 12. For the execution times, the proposed method outperforms all other algorithms.The speedup of CUR-HU over PLMM, ELMM, GLMM, and DeepGUn methods ranged between 9.4 to 249.5 times estimating the endmembers and their abundance maps of Samson, Houseton, and Jasper Ridge HSI datasets.
Overall, the proposed method yielded a comparable unmixing performance with no parameter tuning and at high computational efficiency. Furthermore, the proposed framework can be utilized in the HU algorithms as an initialization step to provide the number of endmembers, endmember signatures, and abundance maps.

6. Conclusions

In this paper, we propose a new hyperspectral unmixing (HU) framework based on CUR matrix decomposition called CUR-HU that performs the entire HU process, which consists of three stages (i.e., estimating the optimal number of endmembers, estimating the endmembers and their corresponding abundance maps). However, most of the existing HU algorithms mainly focus on only one or two tasks of the HU process. CUR-HU is a fully unsupervised method, consists of three main stages, and incorporates several techniques to perform the HU process with high computational efficiency. For estimating the number of endmembers, we utilized a QR factorization based on an incremental method (IQR), used for approximating the singular vectors of the hyperspectral image (HSI) that are required for the deterministic selecting method, which significantly improves the scalability of the proposed method. Then, for the blind HU that estimates the endmembers and their abundance maps simultaneously, we proposed a method based on CUR matrix decomposition with a deterministic sampling method (i.e., DEIM). A unique advantage of our method is that the CUR decomposition based on IQR decomposition provides the matrix rank information for estimating the number of endmembers, and at the same time, we can compute the low-rank approximation of the original data with meaningful interpretation. Extensive experiments were conducted on various synthetic and real HSIs datasets to demonstrate the effectiveness of each stage of the proposed framework. The experiment results show that CUR-HU performs comparable to state-of-the-art methods for estimating the number of endmembers and abundance maps. However, it outperforms other methods for estimating the endmembers and the computational efficiency. Furthermore, CUR-HU shows robustness to spectral variability and is simple to implement and applicable to parallelizing. The proposed framework can be utilized with the existing HU algorithms to initialize the number of endmembers, signatures, and abundance maps. It can also be used to provide supervised information for HU methods based on deep learning techniques. CUR-HU has a 9.4 to 249.5 times speedup over different state-of-the-art methods for estimating the endmembers and their abundance maps of different real HSI datasets.   

Author Contributions

Conceptualization, M.A.A.A., R.C.C.C. and H.Y.; Methodology, M.A.A.A., R.C.C.C. and H.Y.; Software, M.A.A.A.; Writing—original draft, M.A.A.A.; Writing—review & editing, R.C.C.C. and H.Y.; Supervision, R.C.C.C. and H.Y.; Project administration, R.C.C.C. and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Hong Kong Innovation and Technology Commission (InnoHK Project CIMDA), Hong Kong Research Grants Council (Project 11204821), and City University of Hong Kong (Project 9610460).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. QR Factorization Based on an Incremental Method

In general, at step t of the algorithm, the approximated Q R factorization is given by
Y t Q t R t
where Q t contains the orthonormal columns, R t is the triangular factor at step t, and Y t contains the first t columns of HSI data matrix Y  [55]. So,
Y t Q t R t F t o l · d t · R t F
where . F is the Frobenius norm, and d t is the number of row/column deletions (number of truncation steps) that have been happened up to and including step t. At this point, the dimension of Q t is L × t d t , R t is t d t × t , and d t = min { L , N }  k, where k = rank Q R .
Assume that the difference error matrix at the step t is E t = A t Q t R t , and let
E t F t o l · d t · R t F
By using Gram–Schmidt to orthogonalize the next column t + 1 of Y , the Q R factorization will update as
Y t + 1 = Q t + 1 R t + 1 + E t , 0
In the case of no truncation occurring, then d t + 1 = d t and R t F R t + 1 F . The norm of the error matrix is
E t + 1 F = E t , 0 F t o l · d t · R t F t o l · d t + 1 · R t + 1 F .
In other words, each step has a row and column deleted from or added to R and Q , respectively, according to the truncation condition.
Assume that after the number deletions d t , the matrix R t has dimension m × j (i.e., m = j d j ) and the index of the row of the minimum norm is i. Furthermore, assume that R ^ t + 1 and Q ^ t + 1 are obtained from R t + 1 and Q t + 1 by removing the ith row and column, respectively. The truncation case occurs if r i T = e i T R t + 1 satisfies r i T t o l · R ^ t + 1 F , then deleting row i of R t + 1 and column i of Q t + 1 column i of Q j + 1 and row i of R j + 1 and replacing Q t + 1 and R t + 1 with Q ^ t + 1 and R ^ t + 1 . So,
Q ^ t + 1 R ^ j + 1 = Q t + 1 R t + 1 e i r i T
Hence,
Y t + 1 = Q t + 1 R t + 1 e i r i T + E t , 0 + Q t + 1 e i r i T
Furthermore, R ^ t + 1 contains row m + 1 of R t + 1 , which has a norm larger than the row determined for deletion. At the end of the truncation step, the Q t + 1 and R t + 1 replaced with Q ^ t + 1 and R ^ t + 1 to obtain the approximation within
Y t + 1 t o l · d t + 1 · R t + 1 F
where d t + 1 = d t + 1 .
Algorithm A1: Estimating the endmembers number
Remotesensing 16 00766 i002

Appendix A.1. The Complexity of Computing the Singular Vectors of HSI

The optimal computing of SVD of a large HSI dataset Y of size L by N requires O ( L N 2 + L 2 N + N 3 ) . By approximating the HSI Y using the QR factorization as follows, Y Q R where Q R L × p that contains the orthonormal column vectors, R R p × N is the upper triangular matrix, and p is the estimated number of endmembers. Therefore, computing the SVD based on the QR factorization requires O ( L N p 2 ) , which is more efficient [58].
Table A1. Estimation of the number of endmembers of synthetic HSI under changing SNR and p for ( η = 1 / 18 ).
Table A1. Estimation of the number of endmembers of synthetic HSI under changing SNR and p for ( η = 1 / 18 ).
SNRMethodsNo. of Endmembers p
3 5 10 15
50 dBHySime351015
NWHFC ( P f = 1 × 10 3 )9(3)10(5)12(7)10(11)
NWHFC ( P f = 1 × 10 4 )48(3)33(5)54(10)34(10)
NWHFC ( P f = 1 × 10 5 )43(3)28(5)41(9)27(10)
GENE-AH ( P f = 1 × 10 3 )471116
GENE-AH ( P f = 1 × 10 4 )551115
GENE-AH ( P f = 1 × 10 5 )351020
SPICE351114
FUN ( P f = 1 )351014
Ours ( t o l = 2 × 10 3 )36914
35 dBHySime351014
NWHFC ( P f = 1 × 10 3 )3(3)4(4)7(7)9(9)
NWHFC ( P f = 1 × 10 4 )9(3)9(5)11(7)8(10)
NWHFC ( P f = 1 × 10 5 )3(3)4(4)6(6)8(8)
GENE-AH ( P f = 1 × 10 3 )481823
GENE-AH ( P f = 1 × 10 4 )361521
GENE-AH ( P f = 1 × 10 5 )551918
SPICE45913
FUN ( P f = 1 )35913
Ours ( t o l = 1 × 10 3 )36913
25 dBHySime35913
NWHFC ( P f = 1 × 10 3 )4(3)5(5)11(10)9(11)
NWHFC ( P f = 1 × 10 4 )4(3)5(5)11(9)9(10)
NWHFC ( P f = 1 × 10 5 )4(3)5(5)11(8)8(10)
GENE-AH ( P f = 1 × 10 3 )5131926
GENE-AH ( P f = 1 × 10 4 )871917
GENE-AH ( P f = 1 × 10 5 )451726
SPICE18353017
FUN ( P f = 1 )481214
Ours ( t o l = 5 × 10 3 )371116
Ours ( t o l = 1 × 10 2 )35711
15 dBHySime35913
NWHFC ( P f = 1 × 10 3 )4(3)5(5)11(10)9(11)
NWHFC ( P f = 1 × 10 4 )4(3)5(5)11(9)9(10)
NWHFC ( P f = 1 × 10 5 )4(3)5(5)11(8)8(10)
GENE-AH ( P f = 1 × 10 3 )781729
GENE-AH ( P f = 1 × 10 4 )9101912
GENE-AH ( P f = 1 × 10 5 )592124
SPICE36404040
FUN ( P f = 1 )14231730
Ours ( t o l = 2 × 10 2 )581117

Appendix A.2. DEIM for Approximating the Nonlinear Functions

The discrete empirical interpolation method (DEIM) was proposed in [56] to efficiently approximate the nonlinear function f ( τ ) by projecting it onto a subspace that approximates space of f ( τ ) and is spanned by the standard basis W (left singular vectors). In other words, the DEIM method approximates f ( τ ) by combining projection with interpolation, as it constructs specially selected interpolation indices to specify the interpolation-based projection that provides a nearly optimal subspace approximation to f ( τ ) .
The approximation from projecting f ( τ ) onto the subspace spanned by the basis { w 1 , , w r } is given as f ˜ ( τ ) W z ( τ ) , where z ( τ ) is the corresponding coefficients. To determine z ( τ ) , DEIM constructs r row indices based on the standard basis W , at which the function f ( τ ) is evaluated. Suppose the selecting matrix P defined as P = [ e t 1 , , e t r ] R n × r , where e t j is the t j -th column of the identity matrix I n R n × n . If P T W R r × r is nonsingular, then the coefficient vector z ( τ ) can be computed uniquely from
P T f ( τ ) = ( P T W ) z ( τ )
So,
z ( τ ) = ( P T W ) 1 P T f ( τ )
Therefore, the final approximation of f ( τ ) is
f ˜ ( τ ) W z ( τ ) = W ( P T W ) 1 P T f ( τ )
where P W ( P T W ) 1 P T is called as the DEIM projector. So, given the DEIM projector P , DEIM approximates the function f ( τ ) as f ˜ ( τ ) = P f ( τ ) . To obtain the approximation of the nonlinear function using the DEIM method, we need to find the standard projection basis { W 1 , , W r } and the interpolation indices { t 1 , , t r } that used to form the matrix P . Those interpolation indices used to determine the DEIM projector are selected inductively on the basis { W 1 , , W r } by the DEIM procedures as shown in Algorithm 1. It treats the continuous domain as a finite set of discrete points in this continuous domain. The selection procedure of the DEIM method essentially involves minimizing the approximation error through the selected interpolation index in each iteration.

Appendix B

Figure A2a shows the estimated abundance map of the Houston HSI dataset. Although the performance of the proposed method for estimating the Concrete abundance map is weak, it leads to a considerably stronger Met. Roofs and Vegetation abundance maps but a comparable performance for Asphalt abundance map.
The estimated abundance maps for the Jasper Ridge HSI dataset are shown in Figure A2b. The performance of the proposed method is comparable with other methods for estimating abundance maps (i.e., Water, Vegetation, and Soil). Although there is confusion between Water and Road abundance maps, the estimated road region is best compared to other maps.
Figure A1. Ground truth and extracted endmembers of the Urban HSI: the 1st row, ground truth (GT); and the 2nd row contain the estimated endmembers.
Figure A1. Ground truth and extracted endmembers of the Urban HSI: the 1st row, ground truth (GT); and the 2nd row contain the estimated endmembers.
Remotesensing 16 00766 g0a1
Figure A2. Abundance maps of real HSI datasets for all tested methods.
Figure A2. Abundance maps of real HSI datasets for all tested methods.
Remotesensing 16 00766 g0a2

References

  1. Goetz, A.F. Three decades of hyperspectral remote sensing of the Earth: A personal view. Remote Sens. Environ. 2009, 113, S5–S16. [Google Scholar] [CrossRef]
  2. Borsoi, R.A.; Imbiriba, T.; Bermudez, J.C.M.; Richard, C.; Chanussot, J.; Drumetz, L.; Tourneret, J.Y.; Zare, A.; Jutten, C. Spectral variability in hyperspectral data unmixing: A comprehensive review. IEEE Geosci. Remote Sens. Mag. 2021, 9, 223–270. [Google Scholar] [CrossRef]
  3. Keshava, N.; Mustard, J.F. Spectral unmixing. IEEE Signal Process. Mag. 2002, 19, 44–57. [Google Scholar] [CrossRef]
  4. Bioucas-Dias, J.M.; Plaza, A.; Dobigeon, N.; Parente, M.; Du, Q.; Gader, P.; Chanussot, J. Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 354–379. [Google Scholar] [CrossRef]
  5. Chang, C.I.; Du, Q. Estimation of number of spectrally distinct signal sources in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2004, 42, 608–619. [Google Scholar] [CrossRef]
  6. Nascimento, J.M.P.; Bioucas-Dias, J.M. Vertex component analysis: A fast algorithm to unmix hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 898–910. [Google Scholar] [CrossRef]
  7. Heinz, D.C.; Change, C.-I. Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 529–545. [Google Scholar] [CrossRef]
  8. Ahmed, A.M.; Duran, O.; Zweiri, Y.; Smith, M. Hybrid spectral unmixing: Using artificial neural networks for linear/non-linear switching. Remote Sens. 2017, 9, 775. [Google Scholar] [CrossRef]
  9. Shen, X.; Bao, W. Hyperspectral endmember extraction using spatially weighted simplex strategy. Remote Sens. 2019, 11, 2147. [Google Scholar] [CrossRef]
  10. Heinz, D.; Chang, C.I.; Althouse, M.L. Fully constrained least-squares based linear unmixing [hyperspectral image classification]. In Proceedings of the IEEE 1999 International Geoscience and Remote Sensing Symposium. IGARSS’99 (Cat. No. 99CH36293), Hamburg, Germany, 28 June–2 July 1999; Volume 2, pp. 1401–1403. [Google Scholar]
  11. Zhong, Y.; Wang, X.; Zhao, L.; Feng, R.; Zhang, L.; Xu, Y. Blind spectral unmixing based on sparse component analysis for hyperspectral remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2016, 119, 49–63. [Google Scholar] [CrossRef]
  12. Lee, D.; Seung, H.S. Algorithms for non-negative matrix factorization. Adv. Neural Inf. Process. Syst. 2000, 13. Available online: https://papers.nips.cc/paper_files/paper/2000/hash/f9d1152547c0bde01830b7e8bd60024c-Abstract.html (accessed on 28 December 2023).
  13. Kishore Kumar, N.; Schneider, J. Literature survey on low rank approximation of matrices. Linear Multilinear Algebra 2017, 65, 2212–2244. [Google Scholar] [CrossRef]
  14. Zhu, F. Hyperspectral unmixing: Ground truth labeling, datasets, benchmark performances and survey. arXiv 2017, arXiv:1708.05125. [Google Scholar]
  15. Mahoney, M.W.; Drineas, P. CUR matrix decompositions for improved data analysis. Proc. Natl. Acad. Sci. USA 2009, 106, 697–702. [Google Scholar] [CrossRef]
  16. Chiu, J.; Demanet, L. Sublinear randomized algorithms for skeleton decompositions. SIAM J. Matrix Anal. Appl. 2013, 34, 1361–1383. [Google Scholar] [CrossRef]
  17. Ambikapathi, A.; Chan, T.H.; Chi, C.Y.; Keizer, K. Hyperspectral data geometry-based estimation of number of endmembers using p-norm-based pure pixel identification algorithm. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2753–2769. [Google Scholar] [CrossRef]
  18. Bioucas-Dias, J.M.; Nascimento, J.M.P. Hyperspectral subspace identification. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2435–2445. [Google Scholar] [CrossRef]
  19. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  20. Winter, M.E. N-FINDR: An algorithm for fast autonomous spectral end-member determination in hyperspectral data. In Proceedings of the Imaging Spectrometry V. SPIE, Denver, CO, USA, 19–21 July 1999; Volume 3753, pp. 266–275. [Google Scholar]
  21. Boardman, J.W. Automating spectral unmixing of AVIRIS data using convex geometry concepts. In Proceedings of the JPL, Summaries of the 4th Annual JPL Airborne Geoscience Workshop, Washington, DC, USA, 25–29 October 1993; Volume 1. AVIRIS Workshop. [Google Scholar]
  22. Chang, C.I.; Wu, C.C.; Liu, W.; Ouyang, Y.C. A new growing method for simplex-based endmember extraction algorithm. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2804–2819. [Google Scholar] [CrossRef]
  23. Bioucas-Dias, J.M. A variable splitting augmented Lagrangian approach to linear spectral unmixing. In Proceedings of the 2009 First Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, Lisbon, Portugal, 26–28 August 2009; pp. 1–4. [Google Scholar]
  24. Iordache, M.D.; Bioucas-Dias, J.M.; Plaza, A. Sparse unmixing of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2014–2039. [Google Scholar] [CrossRef]
  25. Xu, X.; Shi, Z.; Pan, B. 0-based sparse hyperspectral unmixing using spectral information and a multi-objectives formulation. ISPRS J. Photogramm. Remote Sens. 2018, 141, 46–58. [Google Scholar] [CrossRef]
  26. Tang, W.; Shi, Z.; Wu, Y.; Zhang, C. Sparse unmixing of hyperspectral data using spectral a priori information. IEEE Trans. Geosci. Remote Sens. 2014, 53, 770–783. [Google Scholar] [CrossRef]
  27. Comon, P.; Jutten, C.; Herault, J. Blind separation of sources, Part II: Problems statement. Signal Process. 1991, 24, 11–20. [Google Scholar] [CrossRef]
  28. Chen, C.H.; Zhang, X. Independent component analysis for remote sensing study. In Proceedings of the Image and Signal Processing for Remote Sensing V. SPIE, Florence, Italy, 22–24 September 1999; Volume 3871, pp. 150–158. [Google Scholar]
  29. Liu, L.; Wang, B.; Zhang, L.; Zhang, J.Q. Decomposition of mixed pixels using Bayesian Self-Organizing Map (BSOM) neural networks. In Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, IEEE, Barcelona, Spain, 23–28 July 2007; pp. 2014–2017. [Google Scholar]
  30. Nascimento, J.M.P.; Bioucas-Dias, J.M. Does independent component analysis play a role in unmixing hyperspectral data? IEEE Trans. Geosci. Remote Sens. 2005, 43, 175–187. [Google Scholar] [CrossRef]
  31. Nascimento, J.M.P.; Bioucas-Dias, J.M. Dependent component analysis: A hyperspectral unmixing algorithm. In Proceedings of the Iberian Conference on Pattern Recognition and Image Analysis, Barcelona, Spain, 23–28 July 2007; Springer: Berlin/Heidelberg, Germany, 2007; pp. 612–619. [Google Scholar]
  32. Barros, A.K. The independence assumption: Dependent component analysis. In Advances in Independent Component Analysis; Springer: Berlin/Heidelberg, Germany, 2000; pp. 63–71. [Google Scholar]
  33. Caiafa, C.F.; Salerno, E.; Proto, A.N.; Fiumi, L. Blind spectral unmixing by local maximization of non-Gaussianity. Signal Process. 2008, 88, 50–68. [Google Scholar] [CrossRef]
  34. Zhu, F.; Wang, Y.; Xiang, S.; Fan, B.; Pan, C. Structured sparse method for hyperspectral unmixing. ISPRS J. Photogramm. Remote Sens. 2014, 88, 101–118. [Google Scholar] [CrossRef]
  35. Leplat, V.; Ang, A.M.; Gillis, N. Minimum-volume rank-deficient nonnegative matrix factorizations. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3402–3406. [Google Scholar]
  36. Qian, Y.; Jia, S.; Zhou, J.; Robles-Kelly, A. L1/2 sparsity constrained nonnegative matrix factorization for hyperspectral unmixing. In Proceedings of the 2010 International Conference on Digital Image Computing: Techniques and Applications, Sydney, Australia, 1–3 December 2010; pp. 447–453. [Google Scholar]
  37. Lu, X.; Wu, H.; Yuan, Y.; Yan, P.; Li, X. Manifold regularized sparse NMF for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2815–2826. [Google Scholar] [CrossRef]
  38. Wang, X.; Zhong, Y.; Zhang, L.; Xu, Y. Spatial group sparsity regularized nonnegative matrix factorization for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6287–6304. [Google Scholar] [CrossRef]
  39. Li, J.; Bioucas-Dias, J.M.; Plaza, A.; Liu, L. Robust collaborative nonnegative matrix factorization for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6076–6090. [Google Scholar] [CrossRef]
  40. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar]
  41. Zhou, L.; Zhang, X.; Wang, J.; Bai, X.; Tong, L.; Zhang, L.; Zhou, J.; Hancock, E. Subspace structure regularized nonnegative matrix factorization for hyperspectral unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4257–4270. [Google Scholar] [CrossRef]
  42. Lu, X.; Wu, H.; Yuan, Y. Double constrained NMF for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2013, 52, 2746–2758. [Google Scholar] [CrossRef]
  43. Yuan, Y.; Feng, Y.; Lu, X. Projection-based NMF for hyperspectral unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2632–2643. [Google Scholar] [CrossRef]
  44. Yu, Y.; Ma, Y.; Mei, X.; Fan, F.; Huang, J.; Li, H. Multi-stage convolutional autoencoder network for hyperspectral unmixing. Int. J. Appl. Earth Obs. Geoinf. 2022, 113, 102981. [Google Scholar] [CrossRef]
  45. Su, Y.; Li, J.; Plaza, A.; Marinoni, A.; Gamba, P.; Chakravortty, S. DAEN: Deep autoencoder networks for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4309–4321. [Google Scholar] [CrossRef]
  46. Pattathal V., A.; Sahoo, M.M.; Porwal, A.; Karnieli, A. Deep-learning-based latent space encoding for spectral unmixing of geological materials. ISPRS J. Photogramm. Remote Sens. 2022, 183, 307–320. [Google Scholar]
  47. Licciardi, G.A.; Del Frate, F. Pixel unmixing in hyperspectral data by means of neural networks. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4163–4172. [Google Scholar] [CrossRef]
  48. Wang, M.; Zhao, M.; Chen, J.; Rahardja, S. Nonlinear unmixing of hyperspectral data via deep autoencoder networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1467–1471. [Google Scholar] [CrossRef]
  49. Manolakis, D.; Marden, D.; Shaw, G.A. Hyperspectral image processing for automatic target detection applications. Linc. Lab. J. 2003, 14, 79–116. [Google Scholar]
  50. Luo, B.; Chanussot, J.; Douté, S.; Zhang, L. Empirical automatic estimation of the number of endmembers in hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2012, 10, 24–28. [Google Scholar]
  51. Févotte, C.; Dobigeon, N. Nonlinear hyperspectral unmixing with robust nonnegative matrix factorization. IEEE Trans. Image Process. 2015, 24, 4810–4819. [Google Scholar] [CrossRef]
  52. Goreinov, S.A.; Tyrtyshnikov, E.E.; Zamarashkin, N.L. A theory of pseudoskeleton approximations. Linear Algebra Its Appl. 1997, 261, 1–21. [Google Scholar] [CrossRef]
  53. Stewart, G.W. Four algorithms for the the efficient computation of truncated pivoted QR approximations to a sparse matrix. Numer. Math. 1999, 83, 313–323. [Google Scholar] [CrossRef]
  54. Drineas, P.; Kannan, R.; Mahoney, M.W. Fast Monte Carlo algorithms for matrices I: Approximating matrix multiplication. SIAM J. Comput. 2006, 36, 132–157. [Google Scholar] [CrossRef]
  55. Sorensen, D.C.; Embree, M. A DEIM induced CUR factorization. SIAM J. Sci. Comput. 2016, 38, A1454–A1482. [Google Scholar] [CrossRef]
  56. Chaturantabut, S.; Sorensen, D.C. Nonlinear model reduction via discrete empirical interpolation. SIAM J. Sci. Comput. 2010, 32, 2737–2764. [Google Scholar] [CrossRef]
  57. Lehoucq, R.B.; Sorensen, D.C.; Yang, C. ARPACK Users’ Guide: Solution of Large-Scale Eigenvalue Problems with Implicitly Restarted Arnoldi Methods; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1998. [Google Scholar]
  58. Baker, C.G.; Gallivan, K.A.; Van Dooren, P. Low-rank incremental methods for computing dominant singular subspaces. Linear Algebra Its Appl. 2012, 436, 2866–2888. [Google Scholar] [CrossRef]
  59. Guerra, R.; Santos, L.; López, S.; Sarmiento, R. A new fast algorithm for linearly unmixing hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6752–6765. [Google Scholar] [CrossRef]
  60. Daniel, J.W.; Gragg, W.B.; Kaufman, L.; Stewart, G.W. Reorthogonalization and stable algorithms for updating the Gram–Schmidt QR factorization. Math. Comput. 1976, 30, 772–795. [Google Scholar] [CrossRef]
  61. Giraud, L.; Langou, J.; Rozložník, M.; Eshof, J.v.d. Rounding error analysis of the classical Gram–Schmidt orthogonalization process. Numer. Math. 2005, 101, 87–100. [Google Scholar] [CrossRef]
  62. Green, A.A.; Berman, M.; Switzer, P.; Craig, M.D. A transformation for ordering multispectral data in terms of image quality with implications for noise removal. IEEE Trans. Geosci. Remote Sens. 1988, 26, 65–74. [Google Scholar] [CrossRef]
  63. Roger, R.; Arnold, J. Reliably estimating the noise in AVIRIS hyperspectral images. Int. J. Remote Sens. 1996, 17, 1951–1962. [Google Scholar] [CrossRef]
  64. Boutsidis, C.; Woodruff, D.P. Optimal CUR matrix decompositions. In Proceedings of the the Forty-Sixth Annual ACM Symposium on Theory of Computing, New York, NY, USA, 31 May–3 June 2014; pp. 353–362. [Google Scholar]
  65. Hamm, K.; Huang, L. Perturbations of CUR decompositions. SIAM J. Matrix Anal. Appl. 2021, 42, 351–375. [Google Scholar] [CrossRef]
  66. Zare, A.; Gader, P. Sparsity promoting iterated constrained endmember detection in hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2007, 4, 446–450. [Google Scholar] [CrossRef]
  67. Vane, G.; Green, R.O.; Chrien, T.G.; Enmark, H.T.; Hansen, E.G.; Porter, W.M. The airborne visible/infrared imaging spectrometer (AVIRIS). Remote Sens. Environ. 1993, 44, 127–143. [Google Scholar] [CrossRef]
  68. Landgrebe, D. Multispectral Data Analysis: A Signal Theory Perspective; Purdue Univiersity: West Lafayette, IN, USA, 1998. [Google Scholar]
  69. Miao, L.; Qi, H. Endmember extraction from highly mixed data using minimum volume constrained nonnegative matrix factorization. IEEE Trans. Geosci. Remote Sens. 2007, 45, 765–777. [Google Scholar] [CrossRef]
  70. Ekanayake, M.; Herath, H. Constrained nonnegative matrix factorization for blind hyperspectral unmixing incorporating endmember independence. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 11853–11869. [Google Scholar] [CrossRef]
  71. Jia, S.; Qian, Y. Spectral and spatial complexity-based hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3867–3879. [Google Scholar]
  72. Qian, Y.; Jia, S.; Zhou, J.; Robles-Kelly, A. Hyperspectral unmixing via l1/2 sparsity-constrained nonnegative matrix factorization. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4282–4297. [Google Scholar] [CrossRef]
  73. Qian, Y.; Xiong, F.; Zeng, S.; Zhou, J.; Tang, Y.Y. Matrix-vector nonnegative tensor factorization for blind unmixing of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2016, 55, 1776–1792. [Google Scholar] [CrossRef]
  74. Imbiriba, T.; Borsoi, R.A.; Bermudez, J.C.M. Generalized linear mixing model accounting for endmember variability. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, BC, Canada, 15–20 April 2018; pp. 1862–1866. [Google Scholar]
  75. Borsoi, R.A.; Imbiriba, T.; Bermudez, J.C.M. Deep generative endmember modeling: An application to unsupervised spectral unmixing. IEEE Trans. Comput. Imaging 2019, 6, 374–384. [Google Scholar] [CrossRef]
  76. Chang, C.I. Hyperspectral Data Processing: Algorithm Design and Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  77. Thouvenin, P.A.; Dobigeon, N.; Tourneret, J.Y. Hyperspectral unmixing with spectral variability using a perturbed linear mixing model. IEEE Trans. Signal Process. 2015, 64, 525–538. [Google Scholar] [CrossRef]
  78. Uezato, T.; Murphy, R.J.; Melkumyan, A.; Chlingaryan, A. A novel endmember bundle extraction and clustering approach for capturing spectral variability within endmember classes. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6712–6731. [Google Scholar] [CrossRef]
  79. Drumetz, L.; Veganzones, M.A.; Henrot, S.; Phlypo, R.; Chanussot, J.; Jutten, C. Blind hyperspectral unmixing using an extended linear mixing model to address spectral variability. IEEE Trans. Image Process. 2016, 25, 3890–3905. [Google Scholar] [CrossRef]
Figure 1. Samson HSI dataset from the view of low-rank representation.
Figure 1. Samson HSI dataset from the view of low-rank representation.
Remotesensing 16 00766 g001
Figure 2. An overview of the proposed framework (i.e., CUR-HU) for solving the blind hyperspectral unmixing problem. CUR-HU considers the main three stages of the whole HU process.
Figure 2. An overview of the proposed framework (i.e., CUR-HU) for solving the blind hyperspectral unmixing problem. CUR-HU considers the main three stages of the whole HU process.
Remotesensing 16 00766 g002
Figure 3. An illustration of CUR matrix decomposition, C and R matrices are subsets of columns and rows of a data matrix A , respectively.
Figure 3. An illustration of CUR matrix decomposition, C and R matrices are subsets of columns and rows of a data matrix A , respectively.
Remotesensing 16 00766 g003
Figure 4. Ground truth of the normalized endmembers and their abundance maps of the synthetic HSI dataset.
Figure 4. Ground truth of the normalized endmembers and their abundance maps of the synthetic HSI dataset.
Remotesensing 16 00766 g004
Figure 5. Ground truth and extracted abundance maps by CUR-HU of simulated HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 , respectively.
Figure 5. Ground truth and extracted abundance maps by CUR-HU of simulated HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 , respectively.
Remotesensing 16 00766 g005
Figure 6. Normalized ground truth and extracted endmembers by CUR-HU of simulated HSI: 1st row, ground truth endmember signatures; the 2nd and 3rd rows contain the first 6 endmembers extracted by CUR-HU.
Figure 6. Normalized ground truth and extracted endmembers by CUR-HU of simulated HSI: 1st row, ground truth endmember signatures; the 2nd and 3rd rows contain the first 6 endmembers extracted by CUR-HU.
Remotesensing 16 00766 g006
Figure 7. Normalized ground truth and extracted endmembers by CUR-HU of Samson HSI: 1st row, ground truth endmember signatures; the 2nd, 3rd, and 4th rows contain the first 9 endmembers extracted by CUR-HU.
Figure 7. Normalized ground truth and extracted endmembers by CUR-HU of Samson HSI: 1st row, ground truth endmember signatures; the 2nd, 3rd, and 4th rows contain the first 9 endmembers extracted by CUR-HU.
Remotesensing 16 00766 g007
Figure 8. Ground truth and extracted abundance maps by CUR-HU of Samson HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 matrices, respectively.
Figure 8. Ground truth and extracted abundance maps by CUR-HU of Samson HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 matrices, respectively.
Remotesensing 16 00766 g008
Figure 9. Ground truth and extracted abundance maps by CUR-HU of the Urban HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 , respectively.
Figure 9. Ground truth and extracted abundance maps by CUR-HU of the Urban HSI: 1st row, ground truth (GT); the 2nd, 3rd, 4th, and 5th rows contain the extracted abundance maps by R , U R , U R 1 , and U R 2 , respectively.
Remotesensing 16 00766 g009
Figure 10. Ground truth endmember spectra and extracted endmember spectra of the “Cuprite dataset”: the blue line is the ground truth, and the red line represents the estimated endmembers by CUR-HU.
Figure 10. Ground truth endmember spectra and extracted endmember spectra of the “Cuprite dataset”: the blue line is the ground truth, and the red line represents the estimated endmembers by CUR-HU.
Remotesensing 16 00766 g010
Figure 11. Abundance maps of synthetic data sets using different HU methods (FCLS, PLMM, ELMM, GLMM, and CUR-HU). The colors represent the abundance values from blue ( s k = 0 ) to red ( s k = 1 ).
Figure 11. Abundance maps of synthetic data sets using different HU methods (FCLS, PLMM, ELMM, GLMM, and CUR-HU). The colors represent the abundance values from blue ( s k = 0 ) to red ( s k = 1 ).
Remotesensing 16 00766 g011
Figure 12. Abundance maps of Samson HSI for all tested methods. The colors represent the abundance values from blue ( s k = 0 ) to red ( s k = 1 ).
Figure 12. Abundance maps of Samson HSI for all tested methods. The colors represent the abundance values from blue ( s k = 0 ) to red ( s k = 1 ).
Remotesensing 16 00766 g012
Table 1. Estimation of the number of endmembers of synthetic HSI under changing SNR and p for ( η = 0 ).
Table 1. Estimation of the number of endmembers of synthetic HSI under changing SNR and p for ( η = 0 ).
SNRMethodsNo. of Endmembers p
351015
50 dBHySime351015
NWHFC ( P f = 1 × 10 3 )3(3)5(5)7(7)10(11)
NWHFC ( P f = 1 × 10 4 )3(3)5(5)7(7)8(8)
NWHFC ( P f = 1 × 10 5 )3(3)4(4)7(6)8(8)
HFC ( P f = 1 × 10 3 )35711
HFC ( P f = 1 × 10 4 )3578
HFC ( P f = 1 × 10 5 )3468
GENE-AH ( P f = 1 × 10 3 )661215
GENE-AH ( P f = 1 × 10 4 )461015
GENE-AH ( P f = 1 × 10 5 )451015
SPICE361015
FUN ( P f = 1 )351015
Ours ( t o l = 2 × 10 3 )351015
35 dBHySime351014
NWHFC ( P f = 1 × 10 3 )3(3)4(4)7(7)9(9)
NWHFC ( P f = 1 × 10 4 )3(3)4(4)7(6)8(8)
NWHFC ( P f = 1 × 10 5 )3(3)4(4)6(6)8(8)
HFC ( P f = 1 × 10 3 )3479
HFC ( P f = 1 × 10 4 )3468
HFC ( P f = 1 × 10 5 )3468
GENE-AH ( P f = 1 × 10 3 )681115
GENE-AH ( P f = 1 × 10 4 )471015
GENE-AH ( P f = 1 × 10 5 )451014
SPICE571113
FUN ( P f = 1 )35913
Ours ( t o l = 1 × 10 3 )35913
25 dBHySime35914
NWHFC ( P f = 1 × 10 3 )3(3)4(5)6(6)9(8)
NWHFC ( P f = 1 × 10 4 )3(3)5(5)6(6)7(7)
NWHFC ( P f = 1 × 10 5 )3(3)4(4)5(5)7(7)
HFC ( P f = 1 × 10 3 )3568
HFC ( P f = 1 × 10 4 )3568
HFC ( P f = 1 × 10 5 )3432
GENE-AH ( P f = 1 × 10 3 )571112
GENE-AH ( P f = 1 × 10 4 )451213
GENE-AH ( P f = 1 × 10 5 )451013
SPICE871512
FUN ( P f = 1 )35912
Ours ( t o l = 5 × 10 3 )461112
15 dBHySime35813
NWHFC ( P f = 1 × 10 3 )4(3)5(5)11(10)9(11)
NWHFC ( P f = 1 × 10 4 )3(3)4(4)3(3)3(2)
NWHFC ( P f = 1 × 10 5 )3(3)4(4)3(3)2(2)
HFC ( P f = 1 × 10 3 )3545
HFC ( P f = 1 × 10 4 )3432
GENE-AH ( P f = 1 × 10 3 )810810
GENE-AH ( P f = 1 × 10 4 )77109
SPICE8374040
FUN ( P f = 1 )38251923
Ours ( t o l = 1 × 10 2 )36711
Table 2. Estimation of the number of endmembers on real HSI datasets.
Table 2. Estimation of the number of endmembers on real HSI datasets.
Methods P f / tol Dataset
Indian PinesCuprite
HySime-1516
HFC 10 3 2420
10 4 2217
10 5 2016
NWHFC 10 3 1811
10 4 1910
10 5 189
FUN11119
SPICE-208
Ours 5 × 10 3 97
1 × 10 3 1712
0.5 × 10 3 2114
Table 3. The unmixing performance comparison in terms of SAD and RMSE for a simulated dataset. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 3. The unmixing performance comparison in terms of SAD and RMSE for a simulated dataset. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsMaterialsAverage
SeawaterClintoniteSodium Bicarbonate
SADRMSESADRMSESADRMSESADRMSE
MVNTF0.39050.40880.13260.15060.05670.13240.19330.2306
l 1 / 2 -NMF0.68550.24880.29900.05730.01680.23940.33370.1818
Min-vol NMF0.88910.10520.28030.08540.01880.07450.39610.0884
R-CoNMF1.74080.17530.71090.14730.06140.14750.83770.1569
SSRNMF0.30170.08170.14100.07740.01600.07680.15290.0786
KbSNMF-div0.27460.31900.11190.12000.13110.29850.17250.2458
SGSNMF0.54160.35790.22880.29060.00900.11230.25980.2538
Ours-ANC04.33 × 10 15 09.15 × 10 17 2.10 × 10 8 6.98 × 10 17 7.02 × 10 9 1.50 × 10 15
Ours-ANC-ASC03.48 × 10 15 01.55 × 10 15 2.10 × 10 8 2.41 × 10 15 7.02 × 10 9 2.48 × 10 15
Table 4. The unmixing performance comparison in terms of SAD and RMSE for a Samson HSI dataset. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 4. The unmixing performance comparison in terms of SAD and RMSE for a Samson HSI dataset. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsMaterialsAverage
SoilTreeWater
SADRMSESADRMSESADRMSESADRMSE
MVNTF0.14880.35170.09440.24540.08870.41620.11060.3378
l 1 / 2 -NMF0.34550.42170.14330.04320.35130.23590.28000.2336
Min-vol NMF0.34630.09670.23150.12450.24290.04320.27360.0881
R-CoNMF0.12530.04310.01050.01180.32190.53210.15260.1957
SSRNMF0.60910.38320.07500.03250.16240.16540.28220.1937
KbSNMF-div0.20780.15740.06470.09110.20140.09670.15800.1151
SGSNMF0.37430.05320.17210.08820.29410.14320.28020.0949
Ours-UR0.04040.21890.02190.27650.11890.12670.06040.2073
Ours-ANC0.04040.21890.02190.27640.11890.09500.06040.1968
Ours-ANC-ASC0.04040.12170.02190.10400.11890.16760.06040.1311
Table 5. The unmixing performance comparison in terms of SAD and RMSE for a real data set, “Urban HSI”. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 5. The unmixing performance comparison in terms of SAD and RMSE for a real data set, “Urban HSI”. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsMaterialsAverage
AsphaltGrassTreeRoofDirt
SADRMSESADRMSESADRMSESADRMSESADRMSESADRMSE
MVNTF0.17210.33540.20800.31640.23100.31830.39410.18580.3689 0.2004 0.27480.2713
l 1 / 2 -NMF0.29660.31970.49930.24100.16030.46070.25180.53500.33790.42930.30920.3917
Min-vol NMF0.1849 0.1376 0.1045 0.1490 0.1798 0.1064 0.19300.2130 0.1521 ̲ 0.5409 0.1628 ̲ 0.2294 ̲
R-CoNMF0.20170.19800.27860.30100.2125 0.1258 ̲ 0.24780.19860.24350.52100.23680.2689
SSRNMF 0.0599 0.32370.17800.24370.17950.26600.42170.28510.15340.35890.19850.2955
KbSNMF-div0.1252 0.1654 ̲ 0.2821 0.2354 ̲ 0.1498 ̲ 0.4240 0.1621 ̲ 0.45840.17420.57890.17870.3724
SGSNMF0.41730.55780.34340.25450.14990.26920.38220.22560.33590.56100.32570.3736
Ours 0.1197 ̲ 0.2886 0.1485 ̲ 0.4110 0.0736 0.2073 0.0981 0.1276 0.0611 0.2794 ̲ 0.1003 0.2628
Table 6. The unmixing performance comparison in terms of SAD for a real data set, “Cuprite HSI”. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 6. The unmixing performance comparison in terms of SAD for a real data set, “Cuprite HSI”. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsMaterialsAverage
EM 1 EM 2 EM 3 EM 4 EM 5 EM 6 EM 7 EM 8 EM 9 EM 10 EM 11 EM 12
MVNTF 0.3938 0.31180.56340.31610.47660.45950.25550.31300.28990.38460.40480.36480.3778
l 1 / 2 -NMF0.91450.56380.84680.7002 0.2366 ̲ 0.58690.4558 0.1127 1.07580.97451.0424 0.2291 0.6449
Min-vol NMF0.4747 0.1161 0.1908 0.1316 0.45350.47030.33580.1651 0.1845 ̲ 0.4885 0.2821 ̲ 0.35110.3037
R-CoNMF0.57450.20540.19360.25150.56830.40790.35340.22660.41060.41810.37700.52280.3758
SSRNMF 0.2521 0.1760 ̲ 0.1885 ̲ 1.0213 0.1375 0.1625 ̲ 0.2240 ̲ 0.1311 ̲ 0.1249 0.0595 0.6160 0.2394 ̲ 0.2777
KbSNMF-div 0.3162 ̲ 0.19770.2731 0.1864 ̲ 0.2842 0.1459 0.2086 0.42220.5450 0.2630 ̲ 0.31740.4818 0.3035 ̲
SGSNMF0.74560.5926 0.1182 0.75990.58240.73930.51790.51600.21020.47530.48750.68650.5360
Ours0.35040.37740.40620.29390.27010.22560.42680.30890.30600.3096 0.2803 0.30710.3222
Table 7. Performance comparison of different HU methods for synthetic HSIs. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 7. Performance comparison of different HU methods for synthetic HSIs. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsData Sets
SD0 SD1 SD2 SD3
NRMSE A NRMSE Y Time  (s) NRMSE A NRMSE Y Time  (s) NRMSE A NRMSE Y Time  (s) NRMSE A NRMSE Y Time  (s)
PLMM [77]0.26040.0007100.60.11970.033618.870.20280.030237.910.50660.0320303.5
ELMM [79]0.25540.032149.060.11100.023122.640.19970.023825.810.43850.010621.97
GLMM [74]0.24800.023541.500.11460.022625.600.18410.022636.770.43710.010826.45
DeepGUn [75]0.05660.044875.200.09690.038436.400.16130.045748.960.25500.140396.23
Ours0.18960.03200.270.30170.03780.170.11730.04500.110.25320.05870.10
Table 8. Performance comparison of different HU methods for real HSIs. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
Table 8. Performance comparison of different HU methods for real HSIs. The best performances are in bold typeface; the second-best performances are underlined; and the third-best performances are italicized.
MethodsData Sets
SamsonHousetonJasper Ridge
NRMSE Y Time  (s) NRMSE Y Time  (s) NRMSE Y Time  (s)
PLMM [77]0.023981.680.0713412.650.0553157.23
ELMM [79]0.011933.190.017172.500.027890.87
GLMM [74]0.001145.860.001675.880.003274.23
DeepGUn [75]0.0862121.880.2355259.600.1094209.64
CUR-HU0.04253.540.05885.130.07730.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abdelgawad, M.A.A.; Cheung, R.C.C.; Yan, H. Efficient Blind Hyperspectral Unmixing Framework Based on CUR Decomposition (CUR-HU). Remote Sens. 2024, 16, 766. https://doi.org/10.3390/rs16050766

AMA Style

Abdelgawad MAA, Cheung RCC, Yan H. Efficient Blind Hyperspectral Unmixing Framework Based on CUR Decomposition (CUR-HU). Remote Sensing. 2024; 16(5):766. https://doi.org/10.3390/rs16050766

Chicago/Turabian Style

Abdelgawad, Muhammad A. A., Ray C. C. Cheung, and Hong Yan. 2024. "Efficient Blind Hyperspectral Unmixing Framework Based on CUR Decomposition (CUR-HU)" Remote Sensing 16, no. 5: 766. https://doi.org/10.3390/rs16050766

APA Style

Abdelgawad, M. A. A., Cheung, R. C. C., & Yan, H. (2024). Efficient Blind Hyperspectral Unmixing Framework Based on CUR Decomposition (CUR-HU). Remote Sensing, 16(5), 766. https://doi.org/10.3390/rs16050766

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop