Next Article in Journal
Reorganization of Resting-State EEG Functional Connectivity Patterns in Children with Cerebral Palsy Following a Motor Imagery Virtual-Reality Intervention
Previous Article in Journal
Multi-Objective Instance Weighting-Based Deep Transfer Learning Network for Intelligent Fault Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Structural Adversarial Variational Auto-Encoder for Attributed Network Embedding

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
2
Key Laboratory of Network Information System Technology (NIST), Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
3
University of Chinese Academy of Sciences, Beijing 100190, China
4
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(5), 2371; https://doi.org/10.3390/app11052371
Submission received: 20 January 2021 / Revised: 24 February 2021 / Accepted: 2 March 2021 / Published: 7 March 2021
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
As most networks come with some content in each node, attributed network embedding has aroused much research interest. Most existing attributed network embedding methods aim at learning a fixed representation for each node encoding its local proximity. However, those methods usually neglect the global information between nodes distant from each other and distribution of the latent codes. We propose Structural Adversarial Variational Graph Auto-Encoder (SAVGAE), a novel framework which encodes the network structure and node content into low-dimensional embeddings. On one hand, our model captures the local proximity and proximities at any distance of a network by exploiting a high-order proximity indicator named Rooted Pagerank. On the other hand, our method learns the data distribution of each node representation while circumvents the side effect its sampling process causes on learning a robust embedding through adversarial training. On benchmark datasets, we demonstrate that our method performs competitively compared with state-of-the-art models.

1. Introduction

Networks are ubiquitous since their strong power to model real data, such as telecommunication networks, social networks, citation networks, and biological networks. There are many crucial networks analysis tasks to deal with, such as node classification, node clustering, link prediction, community detection, etc. Since a large portion of networks contain nodes with extra information, attributed networks are really worth to study. Earlier works suffer high computation and great space cost since they usually represent networks with adjacency matrix. Then, network embedding, also known as graph embedding or network representation learning is proposed, aiming to embed vertices into a low dimensional space where the network structural information and network properties are preserved, then it is easy to exploit off-the-shelf machine learning algorithms to analysis networks efficiently [1].
First kind of network embedding models are methods such as DeepWalk [2], node2vec [3] and LINE [4], which adopt language model word2vec [5] in network representation. Second kind of models are based on matrix factorization, such as HOPE [6], TADW [7], and MNF-M [8]. They construct a matrix first, then decompose it to get embeddings.
As for attribute network embedding, since Graph Neural Networks (GCNs) have achieved great success, many models combine the GCN to do network analysis tasks. VGAE [9] uses the auto-encoder framework and combines the Variational auto-encoders (VAE) ([10,11]) and GCN to learn great node representation. Then ARVGA [12] and CAN [13] improve it by adding GAN or changing learning aim. GCN-based models inherit some common defects. VGAE and ARVGA use variational auto-encoders (VAE) to encode the data distribution of the latent codes in graphs while suffering from the drawback of the sampling process in VAE, which makes the latent representation not repeatable for the same sample and harms the performance in related tasks. These typical methods mostly ignore the global information between nodes.
To overcome the deficits mentioned before, we propose a framework (SAVGAE) to encode the graph structure and node attribute of attributed networks into compact representation. We use the higher proximity metric Rooted PageRank to include the local and global information between each pair of nodes. We design a delicate structure to employ the strong learning ability of VAE while circumventing the effect brought by the inner core sampling process through an adversarial training module. The delicate structure makes the representation exclusive for each node and the adversarial training module prevents the performance declining it may cause. We evaluate our model on three typical graph analysis tasks including link prediction, node clustering, and graph visualization meanwhile get competitive performance. In summary, our contributions are as follows:
  • We propose to use Rooted PageRank (RPR) rather than adjacency matrix as the reconstruction objective so as to preserve local and global proximity simultaneously in node representation.
  • We develop a novel graph autoencoder model combining generative adversarial networks and variational autoencoder, where we use the great learning ability of VAE and avoid the side effect it may cause on learning graph embedding.
  • We evaluate the proposed model over two tasks on three widely used graph benchmarks, and get competitive results.
We organize the rest of our paper as follows. In Section 2, we introduce the work related to our model. In Section 3, we detail our model SAVGAE. In Section 4, we introduce the experiments setup and experiments results. Finally, in Section 5, we draw our conclusions.

2. Related Work

In this section, we introduce the models related to our work. Firstly, we introduce the earlier network embedding methods, then we present attributed network embedding models. Finally, we introduce related models exploiting deep generative models.

2.1. Network Embedding

Network embedding assigns each node in the network a low dimensional representation while preserving the network information [14]. Earlier works were related to dimension reduction techniques such as Isomap [15], Locally Linear Embedding (LLE) [16], and Laplacian Eigenmap (LE) [17]. The time complexity of these methods is usually at least quadratic with respect to the number of nodes, thus suffering scalability issue [1]. Recent models are directly designed for representing networks rather than reducing the high dimensionality of the non-relational data.

2.2. Attributed Network Embedding

Attributed Network Embedding exploits the topological structure and the node content simultaneously. Yang et al. prove that DeepWalk is equivalent to matrix factorization, and propose TADW to incorporate text features of nodes into network embedding [7]. TriNDR [18] is a coupled neural network model to exploit network structure, node content, and node labels to learn node representation. However, these models shallow models are either DeepWalk variants or based on the mechanism of matrix factorization. Recently Graph Neural Networks (GNNs) achieve excellent performance due to their strong representation power, and there are some works that exploit GNN in network embedding, such as VGAE, incorporating GCN into variational autoencoder framework.

2.3. Deep Generative Models

Generative models are an important class of models in probability statistics and machine learning, referring to a set of models used to generate observable data randomly. Generative models have a wide range of applications, and can be used to model different kinds of data, such as images, text, sound, etc. Thus, it is promising to use generative models in network representation learning. Deep generative models take advantage of the ability of deep neural networks to approximate arbitrary functions to model a complex distribution, Variational Autoencoder (VAE) and Generative Adversarial Networks (GANs) are two excellent classes.
AAE [19] combines the GAN and autoencoder to perform variational inference. However, the model is not designed for networks. Recently, Dai et al. used the adversarial training in networks [20]. However, the model cannot handle networks with node attributes. ARVGA also combines the GAN with the graph auto-encoder structure. However it only uses the GAN as a regularizer to enhance an intrinsic hypothesis of prior distribution in VAE. In contrast, our model uses the adversarial mechanism to improve the defect of VAE in attribute network embedding. Ref. [21] also try to overcome the problem of VAE, but it aims at general data.

3. Model

In this section, we formalize the attribute network embedding task we focus on firstly. Then we present our Structural Adversarial Variational Graph Auto-Encoder (SAVGAE) model. Next we introduce the Structral Variational Graph Auto-Encoder(SVGA) module, which uses Rooted Pagerank to learn comprehensive topology structural information. Finally, we introduce the adversarial training module which overcomes the defects of the VAE.

3.1. Problem Definition

Definition  1
(Attibuted Network). A attributed network is defined as G = ( V , E , X ) , where V = ( v 1 , v 2 , . . . , v N ) is the set of nodes, and E denotes the sets of edges with e i , j = ( v i , v j ) E . X is the feature matrix with x i X denoting the feature information associated with node v i . Meanwhile the topology information of the network can be represented by adjacency matrix A . All notations are summarized in Table 1.
Problem  1
(Attributed Network embedding). The goal is to project each node v i V to a low-dimension vector z i R N × D :
f : V Z
where D < < N , such that embedding matrix Z should well preserve the topological structure and attribute information of the graph, given an attributed network denoted as G = ( V , E , X ) .

3.2. Overall Architecture

In this section, we introduce the overall architecture of our model. As Figure 1 shows, our model consists of two modules. Firstly, the SVGAE module learns representations for nodes while preserving their overall structural information and attribute information. Then we propose adversarial training module to learn robust embeddings for nodes.
  • Structural Variational Graph Auto-encoder module. SVGAE takes in adjacency matrix A and feature matrix X and gets embeddings Z S , then reconstructs the Rooted Pagerank matrix.
  • Adversarial training module. This module introduces the adversarial training based on the SVGAE to overcome the side effect of the VAE framework.

3.3. Structural Variational Graph Auto-Encoder Module

The Structural Variational Graph Auto-encoder consists of an inference model and a generative model. In our model, we adopt vanilla Graph Convolution Network layer [22] in our graph auto-encoder model. The layerwise transformation is denoted as follows:
Z l + 1 = f ( Z ( l ) , A | W ( l ) )
where Z l is the input layer, A is the adjacency matrix, and the W is the weights parameters to be learned in the training process. To be specific, our graph convolution here is defined as follows:
f ( Z ( l ) , A | W ( l ) ) = ϕ ( D 1 2 A D 1 2 Z ( l ) W ( l ) ) ,
where ϕ is the Relu activation function, A = A + I , D ii = A ij .

3.3.1. Rooted PageRank

Here we choose a high-order proximity Rooted PageRank (RPR) S R P R [23] as an alternative to A . As is defined in Equations (4) and (5), the ( i , j ) -th entry of S R P R defines the probability that if there is a random walk starting from node v i , it will stop at node v j in the steady state. β R P R ( 0 , 1 ) is the parameter to control the probability of the current node walking to a neighbor node randomly rather than jumping back to the start node. P is transition matrix, and P i , j = A i , j k = 1 | V | A i , k . So we can use S R P R to describe the node-to-node proximity. Specifically, the i-th row of S R P R , can be used to represent the global structural information of node v i .
S R P R = β R P R · S R P R · P + ( 1 β R P R ) · I
S R P R = ( 1 β R P R ) ( I β R P R P ) 1

3.3.2. Inference Model

In the inference model, we get hidden variable Z S from the input adjacency matrix A and attribute matrix X .
q ( Z S | X , A ) = i = 1 N q ( z s ( i ) | X , A ) , w i t h q ( z s ( i ) | X , A ) = N ( z s ( i ) | μ s ( i ) , d i a g ( σ i 2 ) )
The inference model can also be seen as an encoder. G ( Z S , A ) = q ( Z S | X , A ) which actually consists of two GCN layers.
Z I = f R e l u ( X , A | W ( 0 ) )
Z ( 21 ) = f R e l u ( Z I , A | W ( 11 ) )
Z ( 22 ) = f R e l u ( Z I , A | W ( 12 ) )
μ = Z ( 21 ) and l o g σ = Z ( 22 ) are matrix of mean vectors z i and matrix of log of the variance vectors respectively.

3.3.3. Generative Model

The generative model can also be seen as a decoder where traditional models aim at reconstructing the adjacency matrix A so the embedding matrix Z can preserve the information of graph G . However, A just encode the one-order structural information of the graph which is not enough for learning comprehensive information of the graph structure. Therefore we adopt the S R P R as the reconstruction aiming of the decoder. So the node embeddings preserve the comprehensive information of the network after loss function being optimized as it can reconstruct the Rooted Pagerank matrix to some extent.
Z S = μ ( Z I ) + σ ( Z I ) ϵ , ϵ N ( 0 , 1 )
p ( S ^ | Z S ) = i = 1 N j = 1 N p ( S i j ^ | z s ( i ) , z s ( j ) ) , w i t h p ( S i j ^ = 1 | z s ( i ) , z s ( j ) ) = θ ( z s ( i ) z s ( j ) )

3.3.4. Optimization

As a Variational graph auto-encoder model, we need optimize the variational lower bound
L = E q ( Z S | X , A ) [ log p ( S ^ | Z S ) ] K L [ q ( Z S | X , A ) | | p ( Z S ) ] ,
where the former indicates the reconstruction loss and the latter uses the Kullback-Leibler divergence as a regularizer of the variational auto-encoder.

3.4. Adversarial Training Module

In the previous section, we can find Z S is suitable for represent the nodes of the network as illustrated in Figure 1. However, locating after the sampling process makes it not repeatable, so we decide to use Z I for its stability, and using adversarial training module to decrease the information loss between Z S and Z I to guarantee the latter is representative.
In our adversarial training module, the input of the discriminator is latent code either from Z I (real) or Z S (fake) and the output is a scalar deciding whether the input is true or not. The cost of the discriminator is
1 2 E x p x [ log D ( f ( X , A ) ) ] 1 2 E x p x [ log ( 1 D ( G ( X , A ) ) ) ] .
Then the objective to train the encoder of SVGAE module with discriminator of the adversarial training module is as follows:
min G max D E x p x [ log D ( f ( X , A ) ) ] + E x p x [ log ( 1 D ( G ( X , A ) ) ) ] .

3.5. Overall Algorithm

As in Algorithm 1, the SAVGAE model takes in the attributed network G , and generates Z I and Z S through Equations (7) and (10) in step 2. Then it takes m samples from Z I and Z S and trains the discriminator in step 6. Finally, it trains the SAVGAE through updating Equation (12) in step 8 and gets the desired network embedding Z I .
Algorithm 1 Algorithm for SAVGAE
Input: G = { V , E , X } : Graph; T: the number of iterations of training process; K: the number of steps for updating discriminator; d: the dimension of the embeddings
Output: Z I R n × d
1: for iterator = 1 to T do
2:     Generate matrix Z I and Z S by Equations (7) and (10);
3:     for iterator = 1 to K do
4:       Sample m samples a ( 1 ) , , a ( m ) from Z I
5:       Sample m samples b ( 1 ) , , b ( m ) from Z S
6:       Update the discriminator with its stochastic gradient
1 m i = 1 m log D a i + log 1 D b i
7:     end for
8:     Update the SVAE module by Equation (12)
9: end for
10: return Z I R n × d

4. Experiments

In this section, we evaluate the proposed model on two widely used tasks: link prediction and node clustering over 2 real-world datasets [24]. Meanwhile, we compare our methods with baseline algorithms.

4.1. Datasets and Baselines

We report experimental results on two attributed graph datasets: Cora and Citeseer. They are citation networks with varying sizes where nodes and edges represent publications and citations. Node attributes of Cora and Citeseer are bag-of-words representing the content of the publication. The statistics of the datasets are summarized in Table 2.
To validate and evaluate SAVGAE, we compare it with 7 baseline methods including DeepWalk, Spectral Clustering [25], node2vec, GAE, VGAE, ARGA, ARVGA in link prediction task. In node clustering task, we include 6 methods K-means, Graph Encoder [26], DNGR [27], RTM [28], RMSC [29], TADW besides baselines in link prediction task.

4.2. Experimental Setup and Evaluation Methods

We detail the experimental setup and evaluation methods for link prediction and node clustering respectively. Link prediction task aims at inferring the missing edges from the network which is not complete. Following the experimental settings of VGAE, each dataset is partitioned into the training set, the validatoin set and the test set. We randomly remove 5% edges for validation and 10% edges for test which serve as positive samples. Then we sample the same number of non-existing links as negative samples from the pairs of unconnected nodes. The model is trained on the residual network where attributes of nodes are kept. The model can be evaluated by the ability to classify the positive samples and negative samples, and we employ the AUC(the area under a receiver operating characteristic curve) and AP(average precision) as the metrics for evaluation.
For the node clustering task, we employ the K-means algorithm based on the node embeddings learned from the training phase. Then we employ accuracy (Acc), normalized mutual information (NMI), F-score (F1), precision, and average rand index (ARI) as metrics, compared to the true class labels of nodes.

4.3. Main Results

In this section, we detail our model performance on link prediction and node clustering experiments. Then we validate the effectiveness of different parts of SAVGAE through ablation studies. Finally, we visualize the node embeddings.

4.3.1. Link Prediction

The experimental results on the link prediction task are reported in Table 3. We can observe that SAVGAE achieves outstanding performance: consistently performs better than all baseline models. Compared with all the baselines, SAVGAE increased the AP score from around 2.9% compared with ARVGE, 3.9% compared with VGAE, 13.1% compared with VGAE without node features; 14.3% and 12.3% compared with Spectral Clustering and DeepWalk respectively.

4.3.2. Node Clustering

After getting node embeddings, we apply K-means algorithm to do the node clustering task. The results on Cora and Citeseer datasets are reported in Table 4 and Table 5. The results illustrate that SAVGA has achieved a great improvement compared with all the other baselines. For example, on Cora dataset, SAVGAE has increased the accuracy from 40.2% compared with K-means to 88.0% compared with Spectral Clustering; increased the F1 score from 44.5% compared with TADW to 102.8% compared with DeepWalk; and increased NMI from 55.0% compared with K-means to 10.4% compared with ARVGE. The results has demonstrated the superiority of our SAVGAE.

4.3.3. Graph Visualization

Visualization is another intuitive and effective way to assess the performance of network embedding methods. We get node representations through SAVGAE model, then map these points into 2D space by t-SNE [30]. The color denotes the class label of the nodes in the dataset, which is not exploited in our unsupervised attributed network embedding. Therefore, nodes in the same color should locate closer to each other. As in Figure 2, the meaningful clustering layout demonstrates the performance of SAVGAE.

4.4. Ablation Studies

To evaluate the performance of different parts of SAVGAE, we perform ablation studies on link prediction and node clustering task. “-RPR” means dicarding high-order proximity Rooted Pagerank and using adjacency matrix as learning objective. “-Adversarial training” means discarding adversarial training module. “-RPR and adversarial training” means discarding both Rooted Pagerank and adversarial training module. The results are illustrated in Table 6, Table 7 and Table 8. We find that the hybrid model performs best.

5. Conclusions

In this paper, we propose SAVGAE, a novel attributed network embedding method, which preserves comprehensive topology structure and node content information of attributed network by modifying the VAE to handle graph data while integrating with generative adversarial networks. We use Rooted PageRank (RPR) rather than adjacency matrix as the reconstruction objective of the auto-encoder to preserve local and global proximity simultaneously in node representation while not increasing much model complexity. The sampling process is an indispensable part of VAE before getting a latent representation of the node, but it causes the problem that the node embeddings are not exclusive which may affect the downstream network analysis task. We modify the VAE structure and integrate with generative adversarial networks to learn an exclusive and robust embedding for each node to overcome the aforementioned problem. Empirically, we evaluate the generated network representations in a variety of network datasets and applications. The results demonstrate the superiority of our method compared with the state-of-the-art. Our future work will focus on proposing a generalized framework that can handle unseen nodes and networks rather than just learning embeddings on a single fixed network.

Author Contributions

Conceptualization, J.Z. and F.L.; methodology, J.Z., Y.W., D.L.; validation, J.Z.; formal analysis, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, F.L., Y.W., D.L.; supervision, G.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Strategic Priority Research Program of Chinese Academy of Sciences (grant number Y835120378).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree letter acronym
LDLinear dichroism

References

  1. Zhang, D.; Yin, J.; Zhu, X.; Zhang, C. Network representation learning: A survey. IEEE Trans. Big Data 2018, 6, 3–28. [Google Scholar] [CrossRef] [Green Version]
  2. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; ACM: New York, NY, USA, 2014; pp. 701–710. [Google Scholar]
  3. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 855–864. [Google Scholar]
  4. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  5. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2013; pp. 3111–3119. [Google Scholar]
  6. Ou, M.; Cui, P.; Pei, J.; Zhang, Z.; Zhu, W. Asymmetric transitivity preserving graph embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 1105–1114. [Google Scholar]
  7. Yang, C.; Liu, Z.; Zhao, D.; Sun, M.; Chang, E.Y. Network representation learning with rich text information. IJCAI 2015, 2015, 2111–2117. [Google Scholar]
  8. Wang, X.; Cui, P.; Wang, J.; Pei, J.; Zhu, W.; Yang, S. Community preserving network embedding. Proc. AAAI Conf. Artif. Intell. 2017, 17, 3298239–3298270. [Google Scholar]
  9. Kipf, T.N.; Welling, M. Variational Graph Auto-Encoders. NIPS Workshop on Bayesian Deep Learning; Curran Associates: New York, NY, USA, 2016. [Google Scholar]
  10. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  11. Rezende, D.J.; Mohamed, S.; Wierstra, D. Stochastic backpropagation and approximate inference in deep generative models. arXiv 2014, arXiv:1401.4082. [Google Scholar]
  12. Pan, S.; Hu, R.; Long, G.; Jiang, J.; Yao, L.; Zhang, C. Adversarially Regularized Graph Autoencoder for Graph Embedding. arXiv 2018, arXiv:1802.04407. [Google Scholar]
  13. Meng, Z.; Liang, S.; Bao, H.; Zhang, X. Co-embedding attributed networks. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, VIC, Australia, 11–15 February 2019; ACM: New York, NY, USA, 2019; pp. 393–401. [Google Scholar]
  14. Cui, P.; Wang, X.; Pei, J.; Zhu, W. A survey on network embedding. IEEE Trans. Knowl. Data Eng. 2018, 31, 833–852. [Google Scholar] [CrossRef] [Green Version]
  15. Tenenbaum, J.B.; De Silva, V.; Langford, J.C. A global geometric framework for nonlinear dimensionality reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef] [PubMed]
  16. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Belkin, M.; Niyogi, P. Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv. Neural Inf. Proc. Syst. 2001, 14, 585–591. [Google Scholar]
  18. Pan, S.; Wu, J.; Zhu, X.; Zhang, C.; Wang, Y. Tri-party deep network representation. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; AAAI Press: Menlo Park, CA, USA, 2016; pp. 1895–1901. [Google Scholar]
  19. Makhzani, A.; Shlens, J.; Jaitly, N.; Goodfellow, I.; Frey, B. Adversarial autoencoders. arXiv 2015, arXiv:1511.05644. [Google Scholar]
  20. Dai, Q.; Li, Q.; Tang, J.; Wang, D. Adversarial network embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI: Menlo Park, CA, USA, 2018. [Google Scholar]
  21. Zhang, X.; Yao, L.; Yuan, F. Adversarial variational embedding for robust semi-supervised learning. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; ACM: New York, NY, USA, 2019; pp. 139–147. [Google Scholar]
  22. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations, ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  23. Liben-Nowell, D.; Kleinberg, J.M. The link-prediction problem for social networks. J. Assoc. Inf. Sci. Technol. 2007, 58, 1019–1031. [Google Scholar] [CrossRef] [Green Version]
  24. Sen, P.; Namata, G.; Bilgic, M.; Getoor, L.; Gallagher, B.; Eliassi-Rad, T. Collective Classification in Network Data. AI Mag. 2008, 29, 93–106. [Google Scholar] [CrossRef] [Green Version]
  25. Tang, L.; Liu, H. Leveraging social media networks for classification. Data Min. Knowl. Discov. 2011, 23, 447–478. [Google Scholar] [CrossRef]
  26. Tian, F.; Gao, B.; Cui, Q.; Chen, E.; Liu, T.Y. Learning deep representations for graph clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–13 February 2016; AAAI: Menlo Park, CA, USA, 2014; Volume 14, pp. 1293–1299. [Google Scholar]
  27. Cao, S.; Lu, W.; Xu, Q. Deep neural networks for learning graph representations. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–13 February 2016; AAAI: Menlo Park, CA, USA, 2016; Volume 16, pp. 1145–1152. [Google Scholar]
  28. Chang, J.; Blei, D. Relational topic models for document networks. In Proceedings of the Artificial Intelligence and Statistics, Clearwater, FL, USA, 16–19 April 2009; pp. 81–88. [Google Scholar]
  29. Xia, R.; Pan, Y.; Du, L.; Yin, J. Robust multi-view spectral clustering via low-rank and sparse decomposition. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; AAAI: Menlo Park, CA, USA, 2014; pp. 2149–2155. [Google Scholar]
  30. Maaten, L.V.D.; Hinton, G.E. Visualizing non-metric similarities in multiple maps. Mach. Learn. 2011, 87, 33–55. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The proposed Structural Adversarial Variational Graph Auto-Encoder(SAVGAE). It consists of two modules. The SAVE module is a variational graph auto-encoder which encodes the network structure A and network attribute X into embedding Z S and reconstruct the Rooted Pagerank matrix S RPR to make Z S contain comprehensive network topological information. The adversarial training module reduces the information loss between Z I and Z S to guarantee Z I is exclusive and representative.
Figure 1. The proposed Structural Adversarial Variational Graph Auto-Encoder(SAVGAE). It consists of two modules. The SAVE module is a variational graph auto-encoder which encodes the network structure A and network attribute X into embedding Z S and reconstruct the Rooted Pagerank matrix S RPR to make Z S contain comprehensive network topological information. The adversarial training module reduces the information loss between Z I and Z S to guarantee Z I is exclusive and representative.
Applsci 11 02371 g001
Figure 2. The visualization results of SAVGAE on Cora dataset.
Figure 2. The visualization results of SAVGAE on Cora dataset.
Applsci 11 02371 g002
Table 1. Main Notations.
Table 1. Main Notations.
SymbolDescription
V nodes
E edges
G attribute network
X attribute matrix
A adjacency matrix
Z embedding matrix
P transition matrix
Table 2. Statistics of datasets.
Table 2. Statistics of datasets.
Datasets#Nodes#Edges#Attributes#Lables
Cora2708542914337
Citeseer3327473237036
Table 3. Results for Link Prediction. GAE* and VGAE* denote the variants of GAE and VGAE, which do experiments without using input features. Bold font indicates the best result.
Table 3. Results for Link Prediction. GAE* and VGAE* denote the variants of GAE and VGAE, which do experiments without using input features. Bold font indicates the best result.
ModelCoraCiteseer
AUCAPAUCAP
SC84.688.580.585.0
DW83.185.080.583.6
node2vec84.187.283.785.3
GAE*84.388.178.784.1
VGAE*84.087.778.984.1
GAE91.092.089.589.9
VGAE91.492.690.892.0
ARGE92.493.291.993.0
ARVGE92.492.692.493.0
SAVGAE(ours)95.095.394.695.7
Table 4. Clustering Results on Cora. Bold font indicates the best result.
Table 4. Clustering Results on Cora. Bold font indicates the best result.
CoraAccNMIF1PrecisionARI
K-means0.4920.3210.3680.3690.230
SC0.3670.1270.3180.1930.031
GraphEncoder0.3250.1090.2980.1820.006
DeepWalk0.4840.3270.3920.3610.243
node2vec0.5260.3430.4130.3930.284
DNGR0.4190.3180.3400.2660.142
RTM0.4400.2300.3070.3320.169
RMSC0.4070.2550.3310.2270.090
TADW0.5600.4410.4810.3960.332
GAE0.5960.4290.5950.5960.347
VGAE0.6090.4360.6090.6090.346
ARGE0.6400.4490.6190.6460.352
ARVGE0.6380.4500.6270.6240.374
SAVGAE(ours)0.6900.4970.6950.7010.429
Table 5. Clustering Results on Citeseer. Bold font indicates the best result.
Table 5. Clustering Results on Citeseer. Bold font indicates the best result.
CoraAccNMIF1PrecisionARI
K-means0.5400.3050.4090.4050.279
SC0.2390.0560.2990.1790.010
GraphEncoder0.2250.0330.3010.1790.010
DeepWalk0.3370.0880.2700.2480.092
node2vec0.3740.1050.2910.2760.125
DNGR0.3260.1800.3000.2000.044
RTM0.4510.2390.3420.3490.203
RMSC0.2950.1390.3200.2040.049
TADW0.4550.2910.4140.3120.228
GAE0.4080.1760.3720.4180.124
VGAE0.3440.1560.3080.3490.093
ARGE0.5730.3500.5460.5730.341
ARVGE0.5440.2610.5290.5490.245
SAVGAE(ours)0.5820.3520.5730.5960.364
Table 6. Ablation results of link prediction experiments.
Table 6. Ablation results of link prediction experiments.
ModelCoraCiteseer
AUCAPAUCAP
SAVGAE(ours)95.095.394.695.7
-RPR95.295.793.794.7
-Adversarial training92.692.992.794.1
-RPR and adversarial training91.492.690.892.0
Table 7. Ablation results of clustering experiments on Cora.
Table 7. Ablation results of clustering experiments on Cora.
CoraAccNMIF1PrecisionARI
SAVGAE(ours)0.6900.4970.6950.7010.429
-RPR0.6550.4650.6280.6350.393
-Adversarial training0.6420.4620.6400.6320.282
-RPR and adversarial training0.6090.4360.6090.6090.346
Table 8. Ablation results of clustering experiments on Citeseer.
Table 8. Ablation results of clustering experiments on Citeseer.
CoraAccNMIF1PrecisionARI
SAVGAE(ours)0.5820.3520.5730.5960.364
-RPR0.5470.2910.5260.5480.272
-Adversarial training0.5470.2910.5470.5820.272
-RPR and adversarial training0.3440.1560.3080.3490.093
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhan, J.; Li, F.; Wang, Y.; Lin, D.; Xu, G. Structural Adversarial Variational Auto-Encoder for Attributed Network Embedding. Appl. Sci. 2021, 11, 2371. https://doi.org/10.3390/app11052371

AMA Style

Zhan J, Li F, Wang Y, Lin D, Xu G. Structural Adversarial Variational Auto-Encoder for Attributed Network Embedding. Applied Sciences. 2021; 11(5):2371. https://doi.org/10.3390/app11052371

Chicago/Turabian Style

Zhan, Junjian, Feng Li, Yang Wang, Daoyu Lin, and Guangluan Xu. 2021. "Structural Adversarial Variational Auto-Encoder for Attributed Network Embedding" Applied Sciences 11, no. 5: 2371. https://doi.org/10.3390/app11052371

APA Style

Zhan, J., Li, F., Wang, Y., Lin, D., & Xu, G. (2021). Structural Adversarial Variational Auto-Encoder for Attributed Network Embedding. Applied Sciences, 11(5), 2371. https://doi.org/10.3390/app11052371

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop