Next Article in Journal
A Study on Consumers’ Visual Image Evaluation of Wrist Wearables
Previous Article in Journal
Towards an Efficient and Exact Algorithm for Dynamic Dedicated Path Protection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and Isometry

1
School of Electronics and Information Technology, Sun Yat-Sen University, Guangzhou 510006, China
2
Public Experimental Teaching Center, Sun Yat-sen University, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(9), 1117; https://doi.org/10.3390/e23091117
Submission received: 7 July 2021 / Revised: 19 August 2021 / Accepted: 23 August 2021 / Published: 27 August 2021
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Symmetric positive definite (SPD) data have become a hot topic in machine learning. Instead of a linear Euclidean space, SPD data generally lie on a nonlinear Riemannian manifold. To get over the problems caused by the high data dimensionality, dimensionality reduction (DR) is a key subject for SPD data, where bilinear transformation plays a vital role. Because linear operations are not supported in nonlinear spaces such as Riemannian manifolds, directly performing Euclidean DR methods on SPD matrices is inadequate and difficult in complex models and optimization. An SPD data DR method based on Riemannian manifold tangent spaces and global isometry (RMTSISOM-SPDDR) is proposed in this research. The main contributions are listed: (1) Any Riemannian manifold tangent space is a Hilbert space isomorphic to a Euclidean space. Particularly for SPD manifolds, tangent spaces consist of symmetric matrices, which can greatly preserve the form and attributes of original SPD data. For this reason, RMTSISOM-SPDDR transfers the bilinear transformation from manifolds to tangent spaces. (2) By log transformation, original SPD data are mapped to the tangent space at the identity matrix under the affine invariant Riemannian metric (AIRM). In this way, the geodesic distance between original data and the identity matrix is equal to the Euclidean distance between corresponding tangent vector and the origin. (3) The bilinear transformation is further determined by the isometric criterion guaranteeing the geodesic distance on high-dimensional SPD manifold as close as possible to the Euclidean distance in the tangent space of low-dimensional SPD manifold. Then, we use it for the DR of original SPD data. Experiments on five commonly used datasets show that RMTSISOM-SPDDR is superior to five advanced SPD data DR algorithms.

1. Introduction

The past few decades have witnessed the rapid development of machine vision, in which machine learning methods based on different mathematical platforms take various forms of image inputs. The most common form is the vector data flattened by a two-dimensional gray image or multi-channel color image [1,2,3]. This method is rough and simple on data processing and, although it has achieved a certain effect on some models, its problem is destroying the inherent position information and geometry structure of images. Another typical form is to keep the two-dimensional or higher-dimensional tensor data of images. There are lots of learning methods that take this data form as inputs [4,5,6] and the mathematical platform they all use involves tensor algebra. In addition, the convolutional neural network [7,8], as a branch of deep learning, has attained great achievements in processing tensor form for image classification. Though the position and statistics information of original image pixels are well maintained with tensor data as inputs, these methods cannot excavate the inherent geometric relations from the manifolds on which the data are located. On the other hand, taking the convolution neural network as an example, the excessively high computational complexity is also a drawback of the methods using tensor data. The last form is symmetric positive definite (SPD) data. Due to their ability to capture higher-order statistical information and excellent feature expression capacity, SPD data have been widely concerned and applied. Examples of mature applications include joint covariance descriptors for image sets and video frames [9,10,11], region covariance matrices for posture analysis [12,13], pedestrian detection [14,15] and texture discrimination [16,17,18]. Although SPD data provide many advantages and good properties for machine vision tasks, the applications are faced with certain challenges:
  • SPD manifold is not a vector space and does not have the concepts of metric and inner product itself, so the traditional algorithms based on Euclidean space are not feasible here. It is not only unreasonable but also inadequate to apply the principles of Euclidean spaces to analyze SPD manifolds, for example, to directly use the Frobenius inner product to measure the similarity and dissimilarity between two SPD matrices.
  • If the original dimensionality of SPD data is relatively high, there must be lots of redundant information, which not only affects the calculation speed, but may also affect the discrimination effect. In addition, it is well known that high-dimensional spaces can lead to curse of dimensionality, which is embodied in the exponential increase in the geometric volume of the space as the dimensionality increases, making available data extremely sparse. Further, sparsity is a problem for any method that requires statistical significance. Moreover, the high-dimensional feature space is of little significance to the distance metric. Since most classifiers rely on Euclidean distance or Mahalanobis distance, classification performance decreases with the increase in dimensionality.
In order to overcome challenge (1), we can equip the SPD manifold with a Riemannian metric to explain the geometric properties of SPD matrices by the Riemannian geometry. The similarity and dissimilarity between SPD samples can be effectively calculated by the derived corresponding geodesic distance under the assigned Riemannian metric. Some outstanding studies [19,20,21,22] have achieved satisfactory results in this regard. Another treatment method is to embed SPD manifolds to reproducing kernel Hilbert spaces (RKHS) by kernel tricks, then adopt the kernel-based learning methods [21,23,24]. For problem (2), the general solution is dimensionality reduction (DR), or low-dimensional embedding.
This paper proposes a DR algorithm of SPD data based on Riemannian manifold tangent spaces and isometry (RMTSISOM-SPDDR). Different from existing supervised DR methods, our algorithm is unsupervised, which greatly saves the costs due to the fact that the labelling process for supervised learning methods is bound to consume time and effort. In summary, our contributions lie in the followings:
(1)
By using Riemannian manifold geometry, high-dimensional SPD data are mapped to the tangent space at the identity matrix, which makes linear operations possible and preserves the data form and attributes of original SPD data to the maximum extent.
(2)
We embedded the mapped data into a simple low-dimensional tangent space through bilinear transformation, so as to effectively avoid the problems caused by curse of dimensionality such as over-fitting and high calculation costs.
(3)
The bilinear transformation is determined by the proposed isometric criterion to maintain the distance relationship before and after DR. This is an unsupervised process, no manual labeling is required, which greatly saves labor and time costs.
The structure of this paper is as follows: In the Section 2, we introduce relevant preliminary knowledge needed for this paper, including notations, basic concepts of Riemannian manifolds. In the Section 3, some related works are briefly summarized. The Section 4 is our main part; in this section, we explain the proposal and solution of RMTSISOM-SPDDR in detail, with the algorithm flow and complexity analysis. The Section 5 mainly carries on comparative analysis of some advanced algorithms and introduces the connections and differences between them and RMTSISOM-SPDDR. In Section 6, sufficient experiments are conducted to verify the superiority of RMTSISOM-SPDDR. In Section 7, we make a summary of the whole paper.

2. Notations and Preliminaries

2.1. Notations

The basic notations in the paper are presented in Table 1.

2.2. Preliminaries

Riemannian manifolds are based on the concepts of differential manifolds and tangent spaces. For the content of differential manifolds and tangent spaces, please refer to the Appendix A. A generalized differential manifold is a topological space, but there is no distance metric or linear operations between its elements. Tangent spaces of a differential manifold support linear operations, but lack of distance metric between tangent vectors. In order to support machine learning, the differential manifold must be equipped with a Riemannian metric to make it a Riemannian manifold.
The Riemannian metric is a symmetric, positive definite, smooth second order tensor field of a differential manifold. Let Γ 2 ( T p ( M ) ) represent the second order tensor space of the tangent space T p ( M ) , where Γ 2 ( M ) = p M Γ 2 ( T p ( M ) ) . Then the second order tensor field is defined as ω 2 : M Γ 2 ( M ) , for any p M , ω 2 ( p ) Γ 2 ( T p ( M ) ) . If ω 2 satisfies the following properties, for any Φ p , Θ p T p ( M )
(1) symmetric: ω 2 ( p ) ( Φ p , Θ p ) = ω 2 ( p ) ( Θ p , Φ p )
(2) positive definite:
ω 2 ( p ) ( Φ p , Φ p ) 0 and ω 2 ( p ) ( Φ p , Φ p ) = 0     Φ = 0
(3) smooth: ω 2 is smooth on M
Then, ω 2 is called a Riemannian metric on M and ( M , ω 2 ) a Riemannian manifold.
By utilizing the Riemannian metric ω 2 , the length of any smooth curve ρ : ( ε , ε ) M on M can be defined as
L ( ρ ) = ε ε ω 2 ( p ( t ) ) ( ρ ( t ) , ρ ( t ) ) d t
Then, the geodesic distance between any two points on M is further defined, for any p , q M , as
g e o D ( p , q ) = m i n { L ( ρ ) | ρ : ( ε , ε ) M , ρ ( ε ) = p , ρ ( ε ) = q }
Since the symmetric positive definite second order tensor field ω 2 ( p ) on the tangent space T p ( M ) is the inner product of T p ( M ) , once the Riemannian metric is defined, the inner product of the tangent space at each point of the Riemannian manifold is defined. Hence, tangent spaces of Riemannian manifolds are inner product spaces, which are finite-dimensional, namely, Hilbert spaces.

3. Related Works

In this section, we briefly review related works on discriminant analysis and dimensionality reduction (DR) of SPD manifolds. In recent years, the wide applications of multimedia and big data make SPD data occur in more and more occasions and attract attention of an extensive range. As mentioned in the Section 1, the applications of SPD data face some challenges, the most prominent of which are caused by the high dimensionality, for example, the overfitting caused by the case that the sample size is far smaller than the data dimensionality, the increase of computational complexity and the information redundancy caused by high-dimensional data. All these problems seriously affect the effective and reliable applications of SPD data on various occasions. In order to effectually solve above problems, the processing methods of SPD data in almost all relevant studies can be categorized into three kinds. First, the original SPD manifold is embedded into RKHS by an implicit mapping to utilize various kernel learning methods. Second, in order to use the machine learning models built on Euclidean spaces or to perform linear operations on SPD, the original SPD manifold is flattened to a Euclidean space, so as to make use of Euclidean methods. Third, by learning a mapping, the original SPD manifold is directly projected into a lower-dimensional one. These three methods can be said to be the basic methods. Other methods are the derivation or combination of these methods.
There is a lot of research on embedding SPD data into RKHS. This method can not only overcome the defect that original SPD manifolds do not support linear operations, but also be applicable to different occasions by choosing different kernel functions due to the diversity of kernel functions. Wang et al. [24] proposed a discriminant analysis method of multiple manifolds. It first constructs image sets into an SPD manifold and a Grassmann manifold, then embeds the obtained SPD manifold and Grassmann manifold into RKHS by Riemannian kernel and projection kernel, respectively, and, finally, carries out weighted discriminant analysis on them. A kernel-based sparse representation of Lie group and dictionary learning were proposed in [21] and successfully applied in face recognition and texture analysis. Huang et al. [25] mapped the source Euclidean space and Riemannian manifold to a common Euclidean subspace, which is implemented by a corresponding higher-dimensional RKHS. Through this mapping, the problem of learning the metric cross two heterogeneous spaces can be reduced into the problem of learning the Euclidean metric in the target Euclidean space. One of the advantages of using the kernel methods mentioned above is the multiplicity and selectivity of the kernel. However, choosing an impertinent kernel often results in performance degradation. In [26], a kernel learning method is used to overcome the kernel selection problem of classification on the Riemannian manifold. It presents two guidelines for jointly learning the kernel and classifier with the multi-kernel learning method.
Flattening SPD data into Euclidean forms is also a very classical method. Based on a spatio-temporal covariance descriptor, Sanin et al. [27] proposed a new motion and gesture recognition method, as well as a weighted Riemannian locally preserving projection method. The method requires linear operations; the first thing to do in this work is to perform tangential transformation on the spatio-temporal covariance descriptor, namely SPD data. Vemulapalli et al. [28] focused on logarithmic Euclidean Riemannian geometry and proposed a data-driven method for learning the Riemannian metric of SPD matrices. Huang et al. [9] used neural network structure. In this work, the input SPD matrix is converted into a more ideal SPD matrix by the design of the bilinear mapping layer and the nonlinear activation function is applied to the new SPD matrix by the eigenvalue correction layer. Finally, the data of the hidden are mapped to a Euclidian space in the output layer. A common problem with all of these works is that they inevitably destroy the geometry of SPD manifolds. In order to overcome this limitation, [10] proposed a method, LEML, to directly learn the logarithm of SPD matrix. By learning a tangent mapping, LEML can directly convert the matrix logarithm from the tangent space to another tangent space with more distinguishable information. Under the tangential mapping framework, the new metric learning can be reduced to the optimization seeking a Mahalanobis-like matrix, which can be solved by traditional metric learning techniques. Then, a more concise and efficient method, α-CML, was proposed based on the improvement of LEML in [29].
In many cases, SPD data are awfully high-dimensional, so how to effectively represent high-dimensional SPD data in low-dimensional space becomes particularly important. For SPD data, DR is usually performed by seeking a full column rank projection matrix. In [18], a direct DR method from high-dimensional SPD manifolds to low-dimensional ones is proposed. The projection generates a low-dimensional SPD manifold with maximum distinguishing ability through the orthogonal constraint and weighted affiliative similarity metric based on the manifold structure. Under the framework of the neural network, Dong et al. [30] made use of two layers (2D full connection layer and symmetric cleaning layer) to embed original SPD data into a subspace with lower dimension and better differentiation. On the basis of PCA [31], established a common component analysis (CCA) for multiple SPD matrices. This method attempts to find a low-dimensional common subspace but ignores the Riemannian geometry of the SPD manifold. From the perspective of geometric structure, Harandi et al. [32] used bilinear transformation to realize a supervised DR algorithm of SPD data and another unsupervised one. The supervised method gathers samples of the same class and separates samples of different classes at the same time to obtain a low-dimensional embedding representation with strong discrimination. In consideration of the maximum variance unfolding (MVU) algorithm [33], the unsupervised method maps SPD samples to a lower-dimensional SPD manifold in line with the criterion of maximum variance. Under the guidance of the nonparametric estimation theory, ref. [34] proposed a DR method of SPD which makes use of estimated probability distribution function (PDF) to construct an affinity matrix so as to achieve the maximum separability with respect to categories from the perspective of probability.

4. The Proposed Method

In this paper, the proposed dimensionality reduction (DR) method based on tangent spaces and isometry first maps the original SPD data to the tangent space at the identity matrix; then, the bilinear transformation between manifolds is transferred to the tangent spaces. In the tangent space after DR, Euclidean metric is adopted to measure the distance between the data points after DR, while the bilinear transformation is determined by the isometric criterion. Specifically, we aim to keep the distance between data points in the tangent space after DR as close as possible to the corresponding geodesic distance on the original SPD manifold. With this idea in mind, we then derive the model.

4.1. The Isometric Transformation between SPD Manifolds and Their Tangent Spaces

SPD data are a kind of non-Euclidian data which can form a differential manifold under a certain structure. In particular, the tangent space of an SPD manifold is a symmetric matrix space, which can be regarded as the minimum linear extension set of the SPD manifold, that is, the minimum linear space containing the SPD matrices. In principle, machine learning algorithms developed in the symmetric matrix space are also suitable for its SPD subset.
As mentioned earlier, there is no distance metric on a differential manifold or its tangent spaces. A Riemannian metric must be endowed so as to obtain a Riemannian manifold. At present, there are several commonly used Riemannian metrics. The affine invariant Riemannian metric (AIRM) is adopted in this paper; for any X S y m + + D , the inner product of the tangent space T X ( S y m + + D ) at point X is defined as follows:
Φ , Θ X = X 1 / 2 Φ X 1 / 2 , X 1 / 2 Θ X 1 / 2 F , Φ , Θ T X ( S y m + + D )
Particularly, the identity matrix is an SPD and the affine invariant inner product of its tangent space T I D ( S y m + + D ) is as follows:
Φ , Θ I D = I D 1 / 2 Φ I D 1 / 2 , I D 1 / 2 Θ I D 1 / 2 F = Φ , Θ F , Φ , Θ T X ( S y m + + D )
The Euclidean distance between tangent vectors derived from the inner product is
Φ Θ I D 2 = Φ Θ , Φ Θ F , Φ , Θ T X ( S y m + + D )
The geodesic distance on the SPD manifold defined by the AIRM is
g e o D ( X , Y ) = l o g ( Y 1 / 2 X Y 1 / 2 ) F , X , Y S y m + + D
Over here, l o g : S y m + + D S y m D . For any X S y m + + D , by eigen-decomposition, we can have X = U d i a g ( λ 1 , , λ D ) U T . Hence,
l o g ( X ) = U d i a g ( l o g ( λ 1 ) , , l o g ( λ D ) ) U T
It is obvious that l o g ( X ) has the same size as X . Besides, we note that the injection l o g ( ) is monotonic, which does not change the order of eigenvalues. Hence, it realizes the transformation with the minimum change of properties compared with original data.
If we intend to transfer learning tasks from the original SPD manifold to its tangent space, we first need to transform original data from the manifold to one of its tangent spaces. This paper selects the tangent space at identity T I D ( S y m + + D ) and l o g transformation. For any X S y m + + D ,
g e o D ( X , I D ) = l o g ( I D 1 / 2 X I D 1 / 2 ) F = l o g ( X ) F
where l o g ( X ) F is the Euclidean distance between l o g ( X ) and the origin point. Hence, l o g is actually an isometric transformation; the geodesic distance between SPD data and the identity matrix is equal to the Euclidean distance between its tangent vector and the origin of the tangent space.

4.2. The Bilinear Transformation between Tangent Spaces

This paper studies the DR of SPD data. At present, the bilinear transformation of SPD matrices is the most commonly used method in DR algorithms of SPD data:
f W : S y m + + D S y m + + d
For X S y m + + D ,
f W ( X ) = W T X W S y m + + d
where W R D × d , W T W = I d , D d .
Obviously, f W is a transformation between two SPD manifolds. Since the SPD manifold is not a Euclidean space, it is difficult to directly solve f W , no matter the modeling or optimization.
Due to the fact that the tangent space of an SPD manifold contains the SPD manifold itself, this paper proposes a DR model of bilinear transformation based on tangent spaces at identities of SPD manifolds:
g W : T I D ( S y m + + D ) T I d ( S y m + + d )
For any Θ T I D ( S y m + + D ) ,
g W ( Θ ) = W T Θ W T I d ( S y m + + d )
Over here, W R D × d , W T W = I d , D d .
Remark 1:
g W is a transformation between tangent spaces of SPD manifolds, that is, a transformation between two Euclidean spaces. It is obviously much easier to build DR algorithms on Euclidean spaces.
Remark 2:
If we just consider DR between tangent spaces, the constraint W T W = I d is unnecessary. Any matrix W R D × d can make g W ( Θ ) implement DR of the symmetric matrix Θ . However, our objective is to reduce data dimensionality for SPD matrices. By adding the constraint W T W = I d , it can be guaranteed that g W | S y m + + D = f W on the SPD manifold.
The advantage of the DR method based on tangent spaces of SPD manifolds proposed in this paper is that the DR between SPD manifolds is realized by the DR between Euclidean tangent spaces. A brief schematic diagram of the DR process is shown in Figure 1.

4.3. Dimensionality Reduction of SPD Data Based on Tangent Spaces of SPD Manifolds and Isometry

4.3.1. Isometry Modeling

The DR scheme of SPD data based on tangent spaces of SPD manifolds and bilinear transformation described in Section 4.2 is only a framework. Moreover, it is necessary to determine the bilinear transformation matrix W according to a certain criterion. Different criteria generate different DR algorithms. In this paper, an isometric criterion is proposed to keep the distance between data points in the tangent space after DR close to the geodesic distance on the original SPD manifold as much as possible.
For a given set of data on an SPD manifold,
X i S y m + + D ,   i = 1 , , N
Since S y m + + D is a Riemannian manifold, the geodesic distance between two SPD matrices X i and X j (under AIRM) is
g i j = g e o D ( X i , X j ) = l o g ( X j 1 / 2 X i X j 1 / 2 ) F ,   1 i , j N
After l o g transformation, the set of data is transformed to another set of data in T I D ( S y m + + D ) :
l o g ( X i ) T I D ( S y m + + D ) ,   i = 1 , , N
Afterwards, by the bilinear transformation between T I D ( S y m + + D ) and T I d ( S y m + + d ) , we have
W T l o g ( X i ) W T I d ( S y m + + d ) ,   i = 1 , , N
Because the tangent space T I d ( S y m + + d ) is a Euclidean space, the Euclidean distance between any two tangent vectors W T l o g ( X i ) W and W T l o g ( X j ) W is
d i j = W T l o g ( X i ) W W T l o g ( X j ) W F = W T ( l o g ( X i ) l o g ( X j ) ) W F = W T Φ i j W F
where Φ i j = l o g ( X i ) l o g ( X j ) , 1 i , j N .
The so-called isometric criterion is that the geodesic distance g i j is equal to the Euclidean distance d i j .
i = 1 N j = 1 N ( g i j d i j ) 2 W T W = I d m i n

4.3.2. Objective Function

This subsection derives the objective function of the model.
For convenience, we construct two distance matrices, as follows:
G = [ g 11 g 1 N g N 1 g N N ] , D = [ d 11 d 1 N d N 1 d N N ]
According to the isometric criterion, W is selected to minimize the difference between the geodesic distance on the manifold and the distance in the tangent space at identity. Hence, the following model is obtained:
W ^ = a r g   m i n W G D F 2         s . t .   W T W = I d
For g i j G and d i j D , let ε i j = ( g i j d i j ) 2 = g i j 2 + d i j 2 2 g i j d i j ; we expect that g i j and d i j are as close as possible and, replacing g i j with d i j , we have ε i j g i j 2 d i j 2
Let
G 2 = [ g 11 2 g 1 N 2 g N 1 2 g N N 2 ] ,   D 2 = [ d 11 2 d 1 N 2 d N 1 2 d N N 2 ]
Then,
W ^ = a r g   m i n W G D F 2
can be represented in another form:
W ^ = a r g   m i n W G 2 D 2 F 2
Let the centering matrix be Γ ,
Γ = ( I D 1 D ( e e T ) )
where e = [ 1 , 1 , , 1 ] T D , I D is a D × D identity matrix.
Use (19) to center the objective function and obtain
W ^ = a r g   m i n W 1 2 Γ ( G 2 D 2 ) Γ F 2 = a r g   m i n W 1 2 Γ G 2 Γ ( 1 2 Γ D 2 Γ ) F 2 = a r g   m i n W ρ G ρ D F 2
Over here, ρ G = 1 2 Γ G 2 Γ , ρ D = 1 2 Γ D 2 Γ .
It can be proved that ρ G and ρ D are both inner matrices [35], d i j 2 = ( ρ D ) i i + ( ρ D ) j j 2 ( ρ D ) i j .
According to the properties of linear algebra and matrix trace, Equation (20) can be further rewritten as
W ^ = a r g   m i n W [ 2 t r ( ρ G ρ D T ) + ρ G F 2 + ρ D F 2 ]
In consideration of
d i j = Θ i Θ j F = W T ( l o g ( X i ) l o g ( X j ) ) W F X i X j F
hence,
ρ D F 2 ρ S F 2
where
S 2 = [ s 11 2 s 1 N 2 s N 1 2 s N N 2 ] , s i j = X i X j F , ρ S = 1 2 Γ S 2 Γ
It can be seen that the objective function has an upper boundary:
W ^ = a r g   m i n W [ 2 t r ( ρ G ρ D T ) + ρ G F 2 + ρ G F 2 ] a r g   m i n W [ 2 t r ( ρ G ρ D T ) + ρ G F 2 + ρ S F 2 ]
Since G 2 and S 2 are unrelated to W , we can further simplify the objective function as
W ^ = a r g   m i n W 2 t r ( ρ G ρ D T ) = a r g   m i n W 2 i = 1 N j = 1 N ( ρ D ) i j ( ρ G ) i j
In consideration of
2 ( ρ D ) i j = Θ i F 2 + Θ j F 2 Θ i Θ j F 2
substitute (27) into (26),
W ^ = a r g   m i n W 2 i = 1 N j = 1 N Θ i F 2 ( ρ G ) i j + i = 1 N j = 1 N Θ i Θ j F 2 ( ρ G ) i j = a r g   m i n W 2 i = 1 N Θ i F 2 j = 1 N ( ρ G ) i j + i = 1 N j = 1 N Θ i Θ j F 2 ( ρ G ) i j
On account of
j = 1 N ( ρ G ) i j = 0
we have
W ^ = a r g   m i n W i = 1 N j = 1 N Θ i Θ j F 2 ( ρ G ) i j = a r g m i n i = 1 N j = 1 N W T l o g ( X i ) W W T l o g ( X j ) W F 2 ( ρ G ) i j = a r g m i n i = 1 N j = 1 N t r ( W T ( l o g ( X i ) l o g ( X j ) ) W W T ( l o g ( X i ) l o g ( X j ) ) T W ) ( ρ G ) i j
So far, we have derived the objective function of the model and what we need to do next is optimizing the function.

4.4. Solution to RMTSISOM-SPDDR

In this subsection, we solve the derived objective function. Obviously, this is a convex optimization problem, but this objective function does not correspond to a closed-form solution, so we are supposed to find its local optimal solution W ^ through an iterative method.
Let the tth iteration yield W ^ t , then the result of the t+1th iteration can be expressed as
W ^ t + 1 = a r g m i n i = 1 N j = 1 N t r ( W T ( l o g ( X i ) l o g ( X j ) ) W ^ t W ^ t t T ( l o g ( X i ) l o g ( X j ) ) T W ) ( ρ G ) i j = a r g m a x t r ( W T L W ) s . t .   W T W = I d
Over here,
L = i = 1 N j = 1 N ( ρ G ) i j ( l o g ( X i ) l o g ( X j ) ) W ^ t W ^ t T ( l o g ( X i ) l o g ( X j ) ) T
The eigenvalue decomposition of L is needed and W ^ t + 1 is obtained by taking the eigenvectors corresponding to the maximum d eigenvalues. The entire algorithm flow of RMTSISOM-SPDDR is shown in Algorithm 1.
Algorithm 1. Procedures of RMTSISOM-SPDDR.
Input:N training samples X = { X 1 , X 2 , , X N } ,   X i S y m + + D , target dimension d and the maximum iteration times maxIter.
Output: bilinear transformation matrix W D × d .
 1: Map SPD data into the corresponding tangent space at identity using Eqution (7);
 2: Construct Riemannian metric matrix using Equation (16);
 3: Construct the inner product matrix ρ G using Equation (20);
 4: Initialize W ^ 0 = I D ( : , 1 : d ) , namely, the first d columns of I D D × D ;
 5: while t = 1: maxIter
       Calculate L using Equation (32);
           Do eigen-decomposition of L to obtain the bilinear transformation matrix W ^ t ;
      end while
 6: return W ^ t .

4.5. Complexity Analysis

In this subsection, we analyze the computational complexity of RMTSISOM-SPDDR. The complexity of the algorithm comes from five aspects—the mapping from the SPD manifold to its tangent space at identity, the construction of the Riemannian metric matrix, the calculation of the inner matrix ρ G , the construction of L and the eigen-decomposition of L . First of all, the mapping from the SPD manifold to the tangent space at identity requires eigen-decomposition of each data point and taking the logarithm of its eigenvalues. This process can be realized by the built-in function eig in MATLAB and the time complexity of the implementation is O ( N D 3 l o g ( D ) ) . Secondly, the construction of the Riemannian metric matrix requires calculating the geodesic distances of data pairs. Note, it is a symmetric matrix which needs to calculate N ( N + 1 ) / 2 times. Hence, the corresponding time complexity is O ( N ( N + 1 ) / 2 ) = O ( 1 2 N 2 + 1 2 N ) = O ( N 2 ) . The calculation of ρ G involves, first, squaring each element of the Riemannian metric matrix, then multiplying it left and right by the centering matrix, so the time complexity of this procedure is O ( N 2 + N 4 ) = O ( N 4 ) . Besides, the construction of L requires four matrix multiplications and N 2 matrix additions, the time complexity is O ( N 2 ( d 2 D 4 ) ) . Finally, the complexity of the eigen-decomposition for L is O ( D 3 l o g ( D ) ) .
Since these five processes are cascaded, the final time complexity of RMTSISOM-SPDDR is O ( N D 3 l o g ( D ) + N 2 + N 4 + N 2 ( d 2 D 4 ) + D 3 l o g ( D ) ) .

5. Comparison with Other State-of-the-Art Algorithms

In this section, we introduce five state-of-the-art DR algorithms of SPD data, including PDDRSPD [34], PSSSD [36], DARG-Graph [37], SSPDDR [32] and LG-SPDDR [38], in summary. Here, we also point out the differences and connections between RMTSISOM-SPDDR and these algorithms.
These DR algorithms for comparison all belong to supervised learning and affinity, in which data labels are needed. In contrast, RMTSISOM-SPDDR is unsupervised and on the basis of global isometry. Except for LG-SPDDR, the DR of all other comparison algorithms is based on direct bilinear transformation between SPD matrices, while the DR of RMTSISOM-SPDDR is based on the bilinear transformation between tangent vectors of SPD matrices. LG-SPDDR utilizes exp transformation for the dimensional-reduced tangent vectors to map them back to a low-dimensional manifold, which means its framework is different from ours.

5.1. SSPDDR

In SSPDDR [32], a supervised DR method of SPD data is proposed, where the metric δ ( X i , X j ) is given to measure the similarity among SPD data. Moreover, the bilinear transformation is employed to the DR process:
X S y m + + D W T X W S y m + + d
Over here, W R D × d , W T W = I d and D d . We can easily know that, if X is an SPD matrix, W T X W is an SPD matrix as well.
In SSPDDR, the transformation matrix W can be determined as follows:
j = 1 N i = 1 N a ( X i , X j ) δ ( W T X i W , W T X j W ) W T W = I d m i n
Over here, a ( X i , X j ) = a w ( X i , X j ) a b ( X i , X j ) and
a w ( X i , X j ) = { 1 X i   and   X j   a r e   a d j a c e n t   a n d   h a v e   t h e   s a m e   l a b e l 0 o t h e r s
a b ( X i , X j ) = { 1 X i   and   X j   a r e   a d j a c e n t b u t   h a v e   different   l a b e l s 0 o t h e r s
SSPDDR indicates that if X i and X j are adjacent and intraclass, their compactness is maximized in the process of DR, or else their separability is minimized. If X i and X j are not adjacent, their relationship will be neglected in DR.

5.2. PDDRSPD

In PDDRSPD [34], labeled SPD data { ( X i , c i ) | i = 1 , , N } are given, in which c i is the label of X i ,. c i { 1 , , L } .
For each class, PDDRSPD firstly calculates its mean vector and covariance matrix. Subsequently, based on nonparametric estimation, it estimates the Gaussian density functions { f c ( X ) | c = 1 , , L } .
The framework of PDDRSPD is similar to [32], but the affinity coefficients is quite different:
j = 1 N i = 1 N a ( X i , X j ) δ ( W T X i W , W T X j W ) W T W = I d m i n
Over here a ( X i , X j ) = a w ( X i , X j ) a b ( X i , X j ) and
a w ( X i , X j ) = { ( λ w 1 ) | f c ( X i ) f c ( X j ) | X N w ( X i ) | f c ( X i ) f c ( X ) | X j N w ( X i ) 1 i = j 0 o t h e r s
where c represents the common label of X i and X j , N w ( X i ) represents the set of λ w samples belonging to the same class as X i and farthest from X i .
a b ( X i , X j ) = { λ b f c j ( X j ) λ s X N b ( X j ) f c j ( X ) c j C b ( X i ) 0 o t h e r s
Over here, C b ( X i ) = { c i 1 , , c i λ b } and
f c i 1 ( X i ) f c i λ b ( X i ) λ b f c i L ( X i )
N b ( X j ) is the set of λ s SPD matrices belonging to the same class as X j and corresponding to largest Gaussian density values.

5.3. LG-SPDDR

In LG-SPDDR [38], a DR method for SPD data is proposed as
X S y m + + D Y W = e x p ( W T l o g ( X ) W ) S y m + + d
where W R D × d , D d .
To be specific, LG-SPDDR is implemented by three procedures:
  • The given SPD data X are firstly mapped from the original SPD manifold S y m + + D to the corresponding tangent space T I D ( S y m + + D ) = S y m D and turn into l o g ( X ) S y m D ;
  • Subsequently, l o g ( X ) is dimensional-reduced by the bilinear transformation, transformed into another tangent space T I d ( S y m + + d ) = S y m d and turns into W T l o g ( X ) W S y m d ;
  • Finally, through the exp-transformation, W T l o g ( X ) W is remapped back into a new low-dimensional SPD manifold S y m + + d and turns into e x p ( W T l o g ( X ) W ) S y m + + d .
LG-SPDDR determines the transformation matrix W as follows:
j = 1 N i = 1 N a ( X i , X j ) δ ( Y i W , Y j W ) W R D × d m i n
Over here, the definition of affinity coefficients a ( X i , X j ) is similar to [32].
Remark 3:
In spite of the claim that LG-SPDDR is based on Lie groups and Lie algebra, it seems to be irrelevant to the multiplication of Lie groups and the Lie bracket of Lie algebra.

5.4. PSSSD

PSSSD [36] is a supervised metric learning method. Both the bilinear transformation and label information are taken into account in the learning process. Its core idea is a framework for learnable metric, in which:
(1) A set of divergences measuring the similarity between two SPD matrices have been given:
D ( α m , β m ) ( X i X j ) = 1 α β l o g ( | α m ( X i X j 1 ) β m + β m ( X i X j 1 ) α m α m + β m | )
α m β m 0 ,   α m + β m 0
Over here, α m and β m are all learnable, m = 1 , , M .
(2) It exploits multiple transformation matrices to map original SPD manifold to multiple submanifolds:
W m R D × d ,   W m T W m = I d ,   m = 1 , , M
(3) To measure the difference between two SDP matrices X i and X j , the learnable metric is defined as
D Θ ( X i X j ) = d i j T Φ d i j ,   d i j = [ D ( α 1 , β 1 ) ( W 1 T X i W 1 W 1 T X j W 1 ) D ( α M , β M ) ( W M T X i W M W M T X j W M ) ] R M
Over here, Φ R M × M is a learnable SPD matrix. Hence, entire learnable parameters can be represented by
Θ = { Φ , W 1 , , W M , α 1 , , α M , β 1 , , β M }
The metric D Θ ( ) is finally specified as
1 | S | ( i , j ) S y i j m a x ( D Θ ( X i , X j ) ζ S , 0 ) 2 + 1 | | ( i , j ) ( 1 y i j ) m a x ( ζ D Θ ( X i , X j ) , 0 ) 2 Θ m i n
where S represents the set of data pairs having the same label and represents the set of data pairs having different labels; y i j = { 1 I f   X i   a n d   X j   a r e   s i m i l a r 0 o t h e r w i s e , ζ S and ζ are similarity thresholds.
The objective of LG-SPDDR indicates that, if X i and X j are similar to each other and belong to the same class, the metric D Θ ( X i , X j ) is minimized. In contrast, if X i and X j are similar but belong to the different classes, D Θ ( X i , X j ) is maximized.
Strictly speaking, PSSSD is essentially not a DR algorithm. To compare it with RMTSISOM-SPDDR, the matrix W = [ W 1 W M ] R D × M d is used for DR.

5.5. DARG-Graph

Let g = N ( μ , Σ ) be a Gaussian distribution, μ R D be its mean vector and Σ R D × D be its covariance matrix. In DARG-Graph, a number of metrics δ ( g i , g j ) to measure the difference between two Gaussian distributions g i and g i are given.
In DARG-Graph [37], the DR of a Gaussian distribution is proposed as follows:
g = N ( μ , Σ ) g W = N ( W T μ , W T Σ W )
where W R D × d and W T W = I d , D d . We can easily know that g W is a Gaussian distribution as well.
The transformation matrix W is specified as follows:
j = 1 N i = 1 N a ( g i , g j ) δ ( g i W , g j W ) W T W = I d m i n
Over here, a ( X i , X j ) = a w ( X i , X j ) a b ( X i , X j ) and
a w ( g i , g j ) = { e δ ( g i , g j ) σ 2 δ ( g i , g j ) < ε     a n d   h a v e   t h e   s a m e   l a b e l 0 o t h e r s
a b ( g i , g j ) = { e δ ( g i , g j ) σ 2 δ ( g i , g j ) < ε     b u t   h a v e   d i f f e r e n t   l a b e l s 0 o t h e r s
Note that RMTDISOM-SPDDR is a DR algorithm for SPD data, not for Gaussian distributions. When comparing our algorithm with DARG-Graph, we let vector μ = 0 . Σ in N ( μ , Σ ) is an SPD matrix. Hence, the DR of Gaussian distributions is converted to the DR of SPD data.

6. Experiments

In this section, we compare RMTSISOM-SPDDR with five state-of-the-art algorithms by evaluating the performance of algorithms on five commonly used benchmark datasets. This section will introduce the datasets and the setting of experimental parameters, present the experimental results and further discuss the effectiveness of our algorithm.

6.1. Datasets Description and Experiment Settings

Dataset overview: Since we conduct experiments on five datasets and the processing of different datasets is not the same, it is not appropriate to list all descriptions here. We explain the specific description of each dataset one by one in the following corresponding subsections.
Parameter setting: The parameters that need to be adjusted for RMTSISOM-SPDDR include the target dimension d of the SPD data after DR and the maximum iteration number maxIter of iterative optimization. For d , the specific setting is related to the dimensionality of original SPD manifolds on which different datasets are located. It can be observed from the different experimental charts how we set it. For maxIter, we uniformly set it as 50, in this paper, for three reasons. Firstly, in our experimental results, under different target dimensions on all datasets, RMTSISOM-SPDDR converged before 50 iterations consistently. Hence, too many iterations are uneconomical for the calculation cost. Secondly, it is insignificant to improve performance by more iterations, which may result in overfitting, making worse performance. Thirdly, although the optimization process iterated for 50 times, the optimal result may occur before the 50th iteration. Our program recorded the classification result of each iteration and we selected the optimal result as the final experimental result. In addition, the classifier used in the experiments was the kNN classifier. For other algorithms, parameter settings are all shown in Table 2.

6.2. The Experiment on the Motion Dataset

Our first experiment was conducted on the Motion dataset. In this subsection, we successively introduce Motion and the experimental development and present results of RMTSISOM-SPDDR and other algorithms on this dataset. Subsections for other datasets keep the same narrative structure and we do not repeat this description.

6.2.1. The Description of Motion

Motion is derived from the human motion sequences in the HDM05 database [39], which was produced by a system based on optical mark technology. Experimenters worn a suit with 40–50 retro-reflective markers, which were tracked by 6–12 calibrated high-resolution cameras, resulting in clear and detailed motion capture data. The database consists of 14 kinds of motions, such as walking, running, jumping, kicking, etc. Partial samples of Motion are shown in Figure 2. Motion includes 306 image sets, each of which consists of 120 images. By the method in [24], they were made into 93 × 93 SPD matrices. We adopted the same partition method as [32] and obtained SPD matrices for training and testing, respectively, with the ratio of 2:3.

6.2.2. The Experimental Results on Motion

On Motion, we carried out classification experiments for the proposed algorithm (RMTSISOM-SPDDR) and other five comparison algorithms, respectively, and tested the convergence performance of RMTSISOM-SPDDR. Subsequent experiments on other datasets also followed the same development and we do not repeat this description in the other subsections.
From the classification results on Motion shown in Figure 3, we can see, under each target dimension selected, the classification accuracy of RMTSISOM-SPDDR is generally more than 70% and far ahead of all comparison algorithms. Especially under the target dimensions of 48, 43, 38 or 33, the accuracy of RMTSISOM-SPDDR is 6 percentage points higher than that of the comparison algorithm ranked second (LG-SPDDR). On the whole, the fluctuation of the classification accuracies of RMTSISOM-SPDDR under different target dimensions is relatively small compared with other algorithms.
To evaluate the convergence performance of RMTSISOM-SPDDR under different target dimensions on Motion, we obtained Figure 4. The loss function decreased the most in the second iteration, then became stable. Among these four curves, the convergence was the fastest when d = 43 or d = 53 and it basically converged at the fifth iteration. When d = 33 or d = 63 , the loss function continued to decline in 50 iterations, but the rangeability was very small after 25 iterations. It was unnecessary to continue iterations after 50 iterations. On the one hand, the decline of the loss function in continuous iterations was insignificant, compared with the calculation cost paid. On the other hand, excessive pursuit of the decline of the loss function may lead to overfitting. When the embedding dimension was equal to 33, the loss value fluctuated greatly, but the overall trend still declined. RMTSISOM-SPDDR still shows satisfactory convergence performance on Motion. It is worth noting that the lower the target dimension, the smaller the absolute value of the loss function, which means that we need less computation.

6.3. The Experiment on the Traffic Dataset

6.3.1. The Description of Traffic

The Traffic video dataset is derived from the UCSD traffic dataset [40], which contains 254 freeway video sequences recorded by a fixed camera from a freeway in Seattle. These sequences can be roughly divided into three traffic conditions (heavy, medium, light; see Figure 5). The three traffic conditions correspond to 44, 45 and 165 sequences, respectively. Each frame was converted to a grayscale image whose size was adjusted to 125 × 125 pixels. We used HoG features [41] to represent frames of a traffic video. In this way, one feature vector corresponded to one frame. In order to acquire the SPD data needed, the feature vectors were normalized and corresponding covariances were calculated. In the experiments on Traffic, we selected 190 sequences for training and the remaining 64 sequences as the test set.

6.3.2. The Experimental Results on Traffic

Figure 6 presents the results of the classification experiments on Traffic. The classification accuracies of RMTSISOM-SPDDR are not all the way ahead of other comparison algorithms. This is mainly reflected in the three dimensions of 45, 40 and 30, in which the accuracies of DARG are higher than ours. The possible reason might be that the scene is relatively complex, while DARG firstly adopts a Gaussian mixed model to extract data features, so as to obtain more discriminant information in high dimensions. However, in cases of low dimensions, the accuracies of DARG drop sharply, while ours still remain at a high level. For SPD data, we have already discussed the problems caused by their high dimensionality, so the DR of SPD data is the objective we are committed to. In such a case, RMTSISOM-SPDDR shows superiority in the data classification under low target dimensions, which not only meets the requirements of reducing data dimensionality, but also maintains high classification performance.
From Figure 7, the convergence performance of selected target dimensions is as follows: From the horizontal comparison, we find that the convergence curves under these four dimensions tended to be stable after the 10th iteration. It can be observed that RMTSISOM-SPDDR also converged fast on Traffic. Similarly, we find, from the horizontal comparison, that the loss function still decreased under different dimensions, which is reasonable.

6.4. The Experiment on the ICT-TV Database

6.4.1. The Description of ICT-TV

The ICT-TV Database [42] consists of two video datasets of TV series, namely, “Big Bang Theory” (BBT) and “Prison Break” (PB), which are shown in Figure 8 and Figure 9. ICT-TV consists of facial video shots taken in these two TV series, namely, 4667 shots in BBT and 9435 shots in PB. The pictures in the shots are stored with the size of 150 × 150 pixels. [30] were followed to preprocess these datasets. We adjusted all detected facial images to 48 × 60 pixels, then adopted histogram equalization, so as to reduce the effect caused by light. All facial frames were flattened into vector representations, on which PCA was performed. Hence, a set of 100-dimensional vectors were obtained which were applied to generate symmetric positive definite kernel matrices. For classification, we utilized the facial videos of 5 main characters in BBT and 11 main characters in PB, respectively.

6.4.2. The Experimental Results on ICT-TV

Because the size of the datasets (BBT and PB) belonging to the ICT-TV database was excessively large, we segmented them to conduct experiments. We divided BBT (or PB) evenly into 10 sub-datasets. For each sub-dataset, training set and test set were divided from these samples with the ratio of 1:1. Finally, 10 experiments were conducted on these sub-datasets and took average results as final classification accuracies.
As can be seen from Figure 10, all algorithms showed high classification accuracies on BBT, while the result of RMTSISOM-SPDDR is higher than that of comparison algorithms in most dimensions. With the decrease in the target dimension, the performance of algorithms decreases. This may indicate that the lower the dimension is, the more information of the original data is lost, leading to a decline in classification.
The results of classification experiments on PB are shown in Figure 11. We can see that RMTSISOM-SPDDR is the most outstanding. This is mainly reflected in the fact that, under most target dimensions, our algorithm achieves the highest classification accuracy, while in the case of d = 20 , the accuracy of SSPDDR is the highest, followed by ours.
Convergence diagrams of RMTSISOM-SPDDR on BBT are shown in Figure 12. They indicate that our algorithm basically reflects preferable convergence performance. For each dimension, the curve was highly similar, generally leveling off after 15–20 iterations and there was almost no fluctuation.
Convergence diagrams of RMTSISOM-SPDDR on PB are shown in Figure 13. It is similar to the cases on BBT. The longitudinal comparison indicates that fifteen iterations were enough to bring the loss function down to an approximately minimum level and the fluctuation since then was very little.

6.5. The Experiment on the ETH-80 Dataset

6.5.1. The Description of ETH-80

ETH-80 is a benchmark image set for recognition [43]. It covers eight categories (cars, cups, horses, dogs, cows, tomatoes, peas, apples). One category is comprised of ten objects, each of which consists of 41 pictures taken from different angles of an object, adjusted to 43 × 43 pixels. As shown in Figure 14, objects of the same category are different in terms of physical attributes, such as visual angle, color, texture distribution and shape. For this dataset, we made it into SPD data according to [24]. Specifically, let M i be the mean image of the i t h image set, M i = 1 n i j = 1 n i C i j , where n i is the number of images in the i t h image set, C i j is the j t h image in the i t h image set. Then, the SPD data corresponding to the i t h image set can be calculated by X i * = 1 n i 1 j = 1 n i ( C i j M i ) ( C i j M i ) T . Further, we added a correction term to X i * so as to guarantee its positive definiteness, that is, X i = X i * + t r ( X i * ) λ I , λ = 10 5 . Each image set corresponds to an SPD matrix. In the experiment, we evenly divided image sets to training sets and test sets with the ratio of 1:1.

6.5.2. The Experimental Results on ETH-80

It can be seen, form Figure 15, that, under each dimension selected, the classification accuracy of RMTSISOM-SPDDR is generally higher than that of the comparison algorithms and remains at a high level (above 80%). When RMTSISOM-SPDDR reduces the original dimension to different target dimensions, the fluctuation of classification accuracy is relatively small and the dimension can be reduced as much as possible without losing performance. Comparatively speaking, the results of DARG-graph and SSPDDR fluctuate greatly, which means there is a trade-off between classification accuracy and target dimension.
Figure 16 presents results of convergence experiments on ETH-80. The longitudinal comparison shows that the value of loss function gradually decreases with the increase in iteration times. Surprisingly, on ETH80, the second iteration leads to the largest drop and subsequent fluctuations are so small to be ignored. This may be because the initial projection matrix for this dataset was already near the local optimal solution. Moreover, the depression of the optimal solution is relatively deep, so that the results of subsequent iterations do not deviate from the optimal solution. From the horizontal comparison, we find that the loss function still tends to decrease under different target dimensions. Hence, our algorithm consistently shows excellent performance.

7. Discussions and Conclusions

(1)
As a non-Euclidean data form, SPD data have in-depth applications in machine learning. Compared with vector representation, SPD data can more effectively extract higher-order statistical information. Generally, in order to avoid the problems of high computational complexity and sparse distribution in high-dimensional SPD manifolds, we hope to reduce the dimensionality of SPD data while maintaining useful and representative information. However, the nonlinear manifold structure results in difficulties in learning tasks where linear operations are needed.
(2)
SPD data equipped with a Riemannian metric generally constitute a Riemannian manifold. For Riemannian manifolds, we have known that all tangent spaces are complete finite-dimensional Hilbert spaces, namely, Euclidean spaces. Hence, this paper transfers learning tasks from original SPD manifolds to their tangent spaces.
(3)
The tangent space of an SPD manifold is a symmetric matrix space. From the perspective of data form and attributes, it can be proved to be the minimum linear extension of the original manifold. Inspired by this, we map SPD data into the tangent space at identity by the isometric transformation (log).
(4)
A framework is proposed for SPD data dimensionality reduction (DR). RMTSISOM-SPDDR realizes the procedure through the bilinear transformation between tangent spaces at identity matrices. These tangent spaces are Hilbert spaces and the required bilinear transformation can be determined to a specific criterion and then taken as the DR transformation for the original SPD manifold.
(5)
This paper specifies the bilinear transformation by the global isometric criterion. The so-called isometric criterion means that the geodesic distance on the original high-dimensional SPD manifold is equal to the corresponding Euclidean distance in the tangent space of the low-dimensional SPD manifold. This preserves the distance relationship between data points well.
(6)
In comparison to many existing state-of-the-art algorithms, such as PDDRSPD [34], PSSSD [36], DARG-Graph [37], SSPDDR [32] and LG-SPDDR [38], there are three differences between them and our proposed method (RMTSISOM-SPDDR). First, most of these algorithms are based on the bilinear transformation between two manifolds, while the proposed RMTSISOM-SPDDR is based on the bilinear transformation between the tangent spaces of the two manifolds. The tangent spaces are finite Hilbert spaces, i.e., Euclidean spaces, which can support more complex DR models. Second, all of these algorithms utilize local analysis, while the proposed RMTSISOM-SPDDR is based on global isometry, providing a different perspective. Finally, all of these algorithms are supervised, while the proposed RMTSISOM-SPDDR is unsupervised.

Author Contributions

Conceptualization, W.G. (Wenxu Gao) and Z.M.; methodology, Z.M. and W.G. (Wenxu Gao); software, W.G. (Wenxu Gao); validation, Z.M., W.G. (Wenxu Gao) and S.L.; formal analysis, Z.M.; investigation, W.G. (Weichao Gan); resources, S.L.; data curation, Z.M.; writing—original draft preparation, W.G. (Wenxu Gao); writing—review and editing, W.G. (Weichao Gan). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Program of Guangzhou, China under Grant 68000-42050001 and the Natural Science Foundation of China under Grant 61773022.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors are grateful to Ting Gao for academic exchange and Sun Yat-sen University for the information supportive platform.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Differential Manifolds and Tangent Spaces

The so-called topological manifold [44] is a topological space which is locally homeomorphic to a Euclidean space R n . If M is an n-dimensional topological manifold, there must be a local coordinate system { ( U i , φ i ) | i U } , where U i is an open set of M , φ i is a homeomorphic mapping from U i to an open set of R n and M = i I U i . ( U i , φ i ) is then called a local coordinate of M .
With regard to any two local coordinates of the topological manifold M , ( U , φ ) and ( V , ξ ) , which satisfy U V ϕ , ξ φ 1 : φ ( U V ) ξ ( U V ) is a mapping between two open sets of R n . If ξ φ 1 is infinitely differentiable, M will be called a differential manifold [44]. Differentiable functions, smooth curves and tangent vectors can be defined on differential manifolds.
Let f : U f R ( U f is an open set of M ) and ( U , φ ) be a local coordinate of M which satisfies U U f ϕ . f φ 1 : φ ( U U f ) R is a function which is called the pull-back function of f defined in R n . If f φ 1 is infinitely differentiable, then we call f a differentiable function.
Similarity, let ρ : ( ε , ε ) M be a curve on M , t 0 ( ε , ε ) . It follows that ρ ( t 0 ) M . If ρ ( t 0 ) U for any local coordinate ( U , φ ) , the function φ ρ : ρ 1 ( ρ ( ε , ε ) U ) φ ( ρ ( ε , ε ) U ) is infinitely differentiable, ρ will be infinitely differentiable at t 0 . If ρ is infinitely differentiable at each point of ( ε , ε ) , then ρ is called a smooth curve on M .
Suppose p M , G p ( M ) represents the set of entire differentiable functions defined in the domain containing p on M . Let Φ p : G p ( M ) R . If Φ p satisfies the following two condition, for any f , g G p ( M ) and α , β R ,
  • Linear: Φ p ( f + β g ) = α Φ p ( f ) + β Φ p ( g )
  • Leibniz Law: Φ p ( f g ) = f ( p ) Φ p ( f ) + g ( p ) Φ p ( f )
then, Φ p is called a tangent vector of p .
Smooth curves can also be used to represent tangent vectors. Let ρ : ( ε , ε ) M be a smooth curve on M . If t 0 ( ε , ε ) , then ρ ( t 0 ) M . Define: ρ ( t 0 ) : G ρ ( t 0 ) ( M ) R , for any f G ρ ( t 0 ) ,
ρ ( t 0 ) ( f ) = d ( f ρ ) ( t ) d t | t = t 0
It can be proved that ρ ( t 0 ) satisfies linear and Leibniz laws. ρ ( t 0 ) is consequently called a tangent vector of ρ ( t 0 ) . Moreover, we can prove it the other way around, for any tangent vector Φ p of p , there must be a smooth curve ρ that goes through p , such that ρ ( t 0 ) = Φ p , where ρ ( t 0 ) = p .
Moreover, let T p ( M ) represent the entire tangent vectors at p , define
( Φ p + Θ p ) ( f ) = Φ p ( f ) + Θ p ( f ) ; ( α Φ p ) ( f ) = α Φ p ( f )
then T p ( M ) can form a linear space.

References

  1. Zhang, J.; Yu, J.; Tao, D. Local Deep-Feature Alignment for Unsupervised Dimension Reduction. IEEE Trans. Image Process. 2018, 27, 2420–2432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Tang, J.; Shao, L.; Li, X.; Lu, K. A Local Structural Descriptor for Image Matching via Normalized Graph Laplacian Embedding. IEEE Trans. Cybern. 2015, 46, 410–420. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Yan, S.; Wang, H.; Fu, Y.; Yan, J.; Tang, X.; Huang, T.S. Synchronized Submanifold Embedding for Person-Independent Pose Estimation and Beyond. IEEE Trans. Image Process. 2008, 18, 202–210. [Google Scholar] [CrossRef] [Green Version]
  4. Li, X.; Ng, M.K.; Cong, G.; Ye, Y.; Wu, Q. MR-NTD: Manifold Regularization Nonnegative Tucker Decomposition for Tensor Data Dimension Reduction and Representation. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 1787–1800. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, Y.-L.; Hsu, C.-T.; Liao, H.-Y.M. Simultaneous Tensor Decomposition and Completion Using Factor Priors. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 577–591. [Google Scholar] [CrossRef] [PubMed]
  6. Du, B.; Zhang, M.; Zhang, L.; Hu, R.; Tao, D. PLTD: Patch-Based Low-Rank Tensor Decomposition for Hyperspectral Images. IEEE Trans. Multimed. 2017, 19, 67–79. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  8. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  9. Huang, Z.; Gool, L.V. A riemannian network for SPD matrix learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2036–2042. [Google Scholar]
  10. Huang, Z.; Wang, R.; Shan, S.; Li, X.; Chen, X. Log-euclidean metric learning on symmetric positive definite manifold with application to image set classification. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–1 July 2015; Volume 37, pp. 720–729. [Google Scholar]
  11. Hussein, M.E.; Torki, M.; Gowayyed, M.A.; El-Saban, M. Human action recognition using a temporal hierarchy of covariance descriptors on 3D joint locations. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013; pp. 2466–2472. [Google Scholar]
  12. Faraki, M.; Harandi, M.; Porikli, F. Image set classification by symmetric positive semi-definite matrices. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2016; pp. 1–8. [Google Scholar]
  13. Chen, K.-X.; Wu, X.-J. Component SPD matrices: A low-dimensional discriminative data descriptor for image set classification. Comput. Vis. Media 2018, 4, 245–252. [Google Scholar] [CrossRef] [Green Version]
  14. Tuzel, O.; Porikli, F.; Meer, P. Region Covariance: A Fast Descriptor for Detection and Classification. In Proceedings of the Transactions on Petri Nets and Other Models of Concurrency XV, Aachen, Germany, 23–28 June 2019; Springer Science and Business Media LLC.: Berlin, Germany, 2006; pp. 589–600. [Google Scholar]
  15. Tuzel, O.; Porikli, F.; Meer, P. Pedestrian Detection via Classification on Riemannian Manifolds. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1713–1727. [Google Scholar] [CrossRef]
  16. Harandi, M.; Sanderson, C.; Hartley, R.; Lovell, B.C. Sparse Coding and Dictionary Learning for Symmetric Positive Definite Matrices: A Kernel Approach. In Transactions on Petri Nets and Other Models of Concurrency XV; Springer: Berlin/Heidelberg, Germany, 2012; pp. 216–229. [Google Scholar]
  17. Jayasumana, S.; Hartley, R.; Salzmann, M.; Li, H.; Harandi, M. Kernel Methods on the Riemannian Manifold of Symmetric Positive Definite Matrices. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 73–80. [Google Scholar]
  18. Harandi, M.; Salzmann, M.; Hartley, R. From Manifold to Manifold: Geometry-Aware Dimensionality Reduction for SPD Matrices. In Proceedings of the Transactions on Petri Nets and Other Models of Concurrency XV; Springer Science and Business Media LLC.: Berlin, Germany, 2014; pp. 17–32. [Google Scholar]
  19. Arsigny, V.; Fillard, P.; Pennec, X.; Ayache, N. Geometric Means in a Novel Vector Space Structure on Symmetric Positive-Definite Matrices. SIAM J. Matrix Anal. Appl. 2007, 29, 328–347. [Google Scholar] [CrossRef] [Green Version]
  20. Ilea, I.; Bombrun, L.B.; Said, S.; Berthoumieu, Y. Covariance Matrices Encoding Based on the Log-Euclidean and Affine In-variant Riemannian Metrics. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 506–509. [Google Scholar]
  21. Li, P.; Wang, Q.; Zuo, W.; Zhang, L. Log-Euclidean Kernels for Sparse Representation and Dictionary Learning. In Proceedings of the 2013 IEEE International Conference on Computer Vision; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2013; pp. 1601–1608. [Google Scholar]
  22. Yamin, A.; Dayan, M.; Squarcina, L.; Brambilla, P.; Murino, V.; Diwadkar, V.; Sona, D. Comparison Of Brain Connectomes Using Geodesic Distance On Manifold: A Twins Study. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019); Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2019; pp. 1797–1800. [Google Scholar]
  23. Zhang, J.; Wang, L.; Zhou, L.; Li, W. Learning Discriminative Stein Kernel for SPD Matrices and Its Applications. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1020–1033. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Wang, R.; Wu, X.-J.; Chen, K.-X.; Kittler, J. Multiple Manifolds Metric Learning with Application to Image Set Classification. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR); Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2018; pp. 627–632. [Google Scholar]
  25. Huang, Z.; Wang, R.; Shan, S.; Van Gool, L.; Chen, X. Cross Euclidean-to-Riemannian Metric Learning with Application to Face Recognition from Video. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2827–2840. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Vemulapalli, R.; Pillai, J.K.; Chellappa, R. Kernel Learning for Extrinsic Classification of Manifold Features. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2013; pp. 1782–1789. [Google Scholar]
  27. Sanin, A.; Sanderson, C.; Harandi, M.T.; Lovell, B.C. Spatio-temporal covariance descriptors for action and gesture recogni-tion. In Proceedings of the 2013 IEEE Workshop on Applications of Computer Vision (WACV), Washington, DC, USA, 15–17 January 2013; pp. 103–110. [Google Scholar]
  28. Vemulapalli, R.; Jacobs, D. Riemannian Metric Learning for Symmetric Positive Definite Matrices. arXiv 2015, arXiv:1501.02393. [Google Scholar]
  29. Zhou, L.; Wang, L.; Zhang, J.; Shi, Y.; Gao, Y. Revisiting Metric Learning for SPD Matrix Based Visual Representation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2017; pp. 7111–7119. [Google Scholar]
  30. Dong, Z.; Jia, S.; Zhang, C.; Pei, M.; Wu, Y. Deep manifold learning of symmetric positive definite matrices with application to face recognition. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4009–4015. [Google Scholar]
  31. Wang, H.; Banerjee, A.; Boley, D. Common component analysis for multiple covariance matrices. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining—KDD ’11, San Diego, CA, USA, 21–24 August 2011; pp. 956–964. [Google Scholar]
  32. Harandi, M.; Salzmann, M.; Hartley, R. Dimensionality Reduction on SPD Manifolds: The Emergence of Geometry-Aware Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 48–62. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Weinberger, K.Q.; Saul, L.K. Unsupervised Learning of Image Manifolds by Semidefinite Programming. Int. J. Comput. Vis. 2006, 70, 77–90. [Google Scholar] [CrossRef] [Green Version]
  34. Ren, J.; Wu, X.-J. Probability Distribution-Based Dimensionality Reduction on Riemannian Manifold of SPD Matrices. IEEE Access 2020, 8, 153881–153890. [Google Scholar] [CrossRef]
  35. Grey, D.R. Multivariate analysis, by K. V. Mardia, J.T. Kent and J. M. Bibby. Pp 522. £14·60. 1979. ISBN 0 12 471252 5 (Academic Press). Math. Gaz. 1981, 65, 75–76. [Google Scholar] [CrossRef]
  36. Gao, Z.; Wu, Y.; Harandi, M.; Jia, Y. A Robust Distance Measure for Similarity-Based Classification on the SPD Manifold. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 3230–3244. [Google Scholar] [CrossRef]
  37. Wang, W.; Wang, R.; Huang, Z.; Shan, S.; Chen, X. Discriminant Analysis on Riemannian Manifold of Gaussian Distributions for Face Recognition with Image Sets. IEEE Trans. Image Process. 2017, 27, 1. [Google Scholar] [CrossRef]
  38. Xu, C.; Lu, C.; Gao, J.; Zheng, W.; Wang, T.; Yan, S. Discriminative Analysis for Symmetric Positive Definite Matrices on Lie Groups. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 1576–1585. [Google Scholar] [CrossRef]
  39. Müller, M.; Röder, T.; Clausen, M.; Eberhardt, B.; Krüger, B.; Weber, A. Documentation Mocap Database HDM05. 2007. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.71.7245 (accessed on 23 August 2021).
  40. Chan, A.B.; Vasconcelos, N. Probabilistic Kernels for the Classification of Auto-Regressive Visual Processes. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05); Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2005; Volume 841, pp. 846–851. [Google Scholar]
  41. Harandi, M.; Salzmann, M.; Baktashmotlagh, M. Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV); Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2015; pp. 4112–4120. [Google Scholar]
  42. Li, Y.; Wang, R.; Shan, S.; Chen, X. Hierarchical hybrid statistic based video binary code and its application to face retrieval in TV-series. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; Volume 1, pp. 1–8. [Google Scholar]
  43. Leibe, B.; Schiele, B. Analyzing appearance and contour based methods for object categorization. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, Wisconsin, 18–20 June 2003; pp. 409–415. [Google Scholar]
  44. Munkres, J.R.; Auslander, L.; MacKenzie, R.E. Introduction to Differentiable Manifolds. Am. Math. Mon. 1964, 71, 1059. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Schematic diagram of DR for SPD data. First, original SPD data are mapped to the tangent space at identity. Subsequently, the bilinear transformation between tangent spaces is determined in line with a specific criterion. Finally, the bilinear transformation learned from tangent spaces is taken to realize the DR of SPD data.
Figure 1. Schematic diagram of DR for SPD data. First, original SPD data are mapped to the tangent space at identity. Subsequently, the bilinear transformation between tangent spaces is determined in line with a specific criterion. Finally, the bilinear transformation learned from tangent spaces is taken to realize the DR of SPD data.
Entropy 23 01117 g001
Figure 2. Samples of the Motion dataset.
Figure 2. Samples of the Motion dataset.
Entropy 23 01117 g002
Figure 3. Classification accuracy on the Motion dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Figure 3. Classification accuracy on the Motion dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Entropy 23 01117 g003
Figure 4. Convergence diagrams on the Motion dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (63, 53, 43, and 33) respectively.
Figure 4. Convergence diagrams on the Motion dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (63, 53, 43, and 33) respectively.
Entropy 23 01117 g004
Figure 5. Samples of the Traffic dataset.
Figure 5. Samples of the Traffic dataset.
Entropy 23 01117 g005
Figure 6. Classification accuracy on the Traffic dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Figure 6. Classification accuracy on the Traffic dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Entropy 23 01117 g006
Figure 7. Convergence diagrams on the Traffic dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (45, 40, 35, and 30) respectively.
Figure 7. Convergence diagrams on the Traffic dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (45, 40, 35, and 30) respectively.
Entropy 23 01117 g007
Figure 8. Samples of BBT Dataset.
Figure 8. Samples of BBT Dataset.
Entropy 23 01117 g008
Figure 9. Samples of PB Dataset.
Figure 9. Samples of PB Dataset.
Entropy 23 01117 g009
Figure 10. Classification accuracy on the BBT dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Figure 10. Classification accuracy on the BBT dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Entropy 23 01117 g010
Figure 11. Classification accuracy on the PB dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Figure 11. Classification accuracy on the PB dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Entropy 23 01117 g011
Figure 12. Convergence diagrams on the BBT dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (90, 80, 70, and 60) respectively.
Figure 12. Convergence diagrams on the BBT dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (ad) represent the results under different target dimensions (90, 80, 70, and 60) respectively.
Entropy 23 01117 g012aEntropy 23 01117 g012b
Figure 13. Convergence diagrams on the PB dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (a–d) represent the results under different target dimensions (90, 80, 70, and 60) respectively.
Figure 13. Convergence diagrams on the PB dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (a–d) represent the results under different target dimensions (90, 80, 70, and 60) respectively.
Entropy 23 01117 g013
Figure 14. Samples of the ETH-80 dataset.
Figure 14. Samples of the ETH-80 dataset.
Entropy 23 01117 g014
Figure 15. Classification accuracy on the ETH80 dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Figure 15. Classification accuracy on the ETH80 dataset. The horizontal axis represents the target dimension for DR and the vertical axis represents the classification accuracy.
Entropy 23 01117 g015
Figure 16. Convergence diagrams on the ETH80 dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (a–d) represent the results under different target dimensions (38, 33, 28, and 23) respectively.
Figure 16. Convergence diagrams on the ETH80 dataset. The horizontal axis represents the number of iterations and the vertical axis represents the value of the loss function. (a–d) represent the results under different target dimensions (38, 33, 28, and 23) respectively.
Entropy 23 01117 g016aEntropy 23 01117 g016b
Table 1. Notations and corresponding descriptions in this paper.
Table 1. Notations and corresponding descriptions in this paper.
NotationsDescriptions
T p ( M ) The tangent space of the differential manifold M at p
G p ( M ) The differentiable function germ of the differential manifold M at p
S y m + + D The entire D × D SPD matrices
S y m D The entire D × D symmetric matrices
X Capital letter denotes an SPD matrix
Φ Greek capital letter denotes a tangent vector of an SPD manifold
Table 2. Parameter settings of five comparison algorithms.
Table 2. Parameter settings of five comparison algorithms.
MethodsParameter Settings
SSPDDRNone
PDDRSPD λ w = 3 , λ b = 3 , λ s = 3
DARG-Graph σ = 1
LG-SPDDR k w = 5 , k b = 20
PSSSD ζ S = 1 , ζ D = 10
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, W.; Ma, Z.; Gan, W.; Liu, S. Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and Isometry. Entropy 2021, 23, 1117. https://doi.org/10.3390/e23091117

AMA Style

Gao W, Ma Z, Gan W, Liu S. Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and Isometry. Entropy. 2021; 23(9):1117. https://doi.org/10.3390/e23091117

Chicago/Turabian Style

Gao, Wenxu, Zhengming Ma, Weichao Gan, and Shuyu Liu. 2021. "Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and Isometry" Entropy 23, no. 9: 1117. https://doi.org/10.3390/e23091117

APA Style

Gao, W., Ma, Z., Gan, W., & Liu, S. (2021). Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and Isometry. Entropy, 23(9), 1117. https://doi.org/10.3390/e23091117

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop