Attribute Graph Embedding Based on Multi-Order Adjacency Views and Attention Mechanisms
Abstract
:1. Introduction
- The encoder model utilizes multi-order neighbor views to construct adjacency matrices. The DGA uses an attention aggregator to effectively assist nodes in the graph in aggregating information from different views;
- A multi-layer perceptron (MLP) decoder and inner product decoder are used simultaneously to decode the adjacency matrix A and feature matrix X. The model is jointly optimized using the reconstruction loss of the adjacency matrix, the self-supervised clustering loss, and the feature matrix reconstruction loss;
- We apply the learned vectors to graph clustering and link prediction tasks using different datasets. Experimental results obtained demonstrate that our model achieves excellent performance on the respective tasks.
2. Related Work
2.1. Graph Embedding
2.2. Graph Auto-Encoder
3. Method
3.1. Encoder
3.1.1. Building Multi-Neighbor Views
3.1.2. Learning Embedding Vectors in a Multi-Neighbor Views
3.1.3. Fusion of Embedding Vectors in a Multi-Neighbor Views
3.1.4. Attention Aggregator
3.2. Decoder
3.2.1. Attention Aggregator
3.2.2. Feature Decoder
3.3. Loss Function
3.3.1. The Loss Function for Reconstructing the Adjacency Matrix A
3.3.2. The Loss Function for Reconstructing the Attribute Matrix X
3.3.3. The Self-Supervised Clustering Loss Function
3.3.4. Joint Optimization
4. Experiments
4.1. Datasets
4.2. Baseline Methods
- (1)
- Classical Methods:
- (2)
- Classical Graph Autoencoder Methods:
- GAE: This uses graph convolutional networks (GCNs) as the encoder and is a classical graph autoencoder used for unsupervised tasks.
- VGAE [6]: This is a variational version of GAE.
- ARGA [29]: This uses an adversarial learning framework to generate node embeddings by adversarial regularization. It incorporates both graph structure and node features into the embeddings. ARVGA is the variational version of ARGA.
- (3)
- Recent Methods:
- ARGAX: [1] This extends ARGA by using GCN as the decoder to simultaneously reconstruct the graph structure and node features.
- ARVGAX [1]: This is the variational version of ARGAX.
- DBGAN [30]: This utilizes a distribution-induced bidirectional generative adversarial network, where the prior distribution is estimated through structure-aware means instead of assuming a normal distribution.
- SSGC [31]: This improves graph convolutional networks (GCNs) with an enhanced Markov diffusion kernel that balances the influences of low-pass and high-pass filters to fuse information from low-order and high-order neighboring nodes.
- R-GAE [32]: This avoids clustering noise through sampling operations and improves the GAE model by transforming the reconstruction graph step into a graph structure relevant to the reconstruction and clustering tasks.
- R-VGAE [32]: This applies the same idea to improve the VGAE model.
- DGAE [33]: This extends GAE by using GCN as the feature decoder.
- GNAE [34]: This resolves the issue of zero embeddings for isolated nodes in the graph by using normalization.
- VGNAE: [34] This is the variational version of GNAE.
- DGA: This uses an MLP feature decoder.
- DGA_EX (DGA Exclude X-decoder): This does not use a feature decoder.
- DGA_GX (DGA GCN X-decoder): This uses a GCN feature decoder.
4.3. Evaluation Metrics
4.4. Parameter Settings
4.5. Experimental Results
4.5.1. Node Clustering Results
4.5.2. Link Prediction Results
4.5.3. Experiment on Influence of Embedding Size
4.5.4. Visualization
5. Discussion
Noise Implications and Accuracy Concerns of Multi-Order Neighbor Views
- Weight Adjustment: Introduce different weights for neighbors of different orders to reduce the influence of more distant neighbors. This ensures that information from immediate neighbors is prioritized while still retaining some level of distant structural information.
- Attention Mechanisms: Utilize technologies such as graph attention networks (GATs) to dynamically assign different attention weights to different neighbors of a node. Through this method, the model can adaptively focus on more relevant neighbors, thereby minimizing the impact of noisy data.
- Neighbor Selection: Employ a strategy for constructing multi-order neighbor views that only selects nodes highly relevant or contributory to the current node. This can be achieved through preliminary feature selection or task-based backpropagation.
- Regularization Techniques: Incorporate regularization terms, such as graph regularization, to penalize the contributions of neighbors excessively distant from the central node, thus avoiding over-reliance on remote information.
6. Conclusions
Advantages of Multi-Order Neighbor Views in Diverse Graph Applications
- Heterogeneous Networks: In networks with diverse types of nodes and edges, multi-order neighbor views can better capture the complex interactions and relationships between different entity types, providing a richer representation of the network.
- Graphs with Long-range Dependencies: For applications where long-range dependencies between nodes are crucial, such as in citation networks or knowledge graphs, multi-order neighbor views can effectively capture these distant relationships, which might be missed by focusing solely on immediate neighbors.
- Community Detection and Clustering: In scenarios where identifying communities or clusters within the graph is essential, multi-order neighbor views can provide insights into the broader community structure, helping to identify not just local but also global community memberships.
- Graphs with Sparse Connections: For graphs with sparse connections, multi-order neighbor views can help in identifying relevant connections that are not immediately apparent, aiding in tasks like link prediction by providing a wider context.
- Dynamic Networks: In dynamic networks where the topology changes over time, multi-order neighbor views can provide a more stable representation by capturing relationships that persist over multiple scales, offering resilience against temporal variations.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Pan, S.; Hu, R.; Fung, S.f.; Long, G.; Jiang, J.; Zhang, C. Learning graph embedding with adversarial training methods. IEEE Trans. Cybern. 2019, 50, 2475–2487. [Google Scholar] [CrossRef] [PubMed]
- Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
- Yu, F.; Liu, Q.; Wu, S.; Wang, L.; Tan, T. Attention-based convolutional approach for misinformation identification from massive and noisy microblog posts. Comput. Secur. 2019, 83, 106–121. [Google Scholar] [CrossRef]
- Hastings, M.B. Community detection as an inference problem. Phys. Rev. E 2006, 74, 035102. [Google Scholar] [CrossRef] [PubMed]
- Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
- Kipf, T.N.; Welling, M. Variational graph auto-encoders. arXiv 2016, arXiv:1611.07308. [Google Scholar]
- Newman, M.E. Finding community structure in networks using the eigenvectors of matrices. Phys. Rev. E 2006, 74, 036104. [Google Scholar] [CrossRef] [PubMed]
- Cao, S.; Lu, W.; Xu, Q. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 18–23 October 2015; pp. 891–900. [Google Scholar]
- Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
- Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
- Wang, D.; Cui, P.; Zhu, W. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1225–1234. [Google Scholar]
- Chen, C.; Lu, H.; Wei, H.; Geng, X. Deep subspace image clustering network with self-expression and self-supervision. Appl. Intell. 2023, 53, 4859–4873. [Google Scholar] [CrossRef]
- Wang, J.; Huang, P.; Zhao, H.; Zhang, Z.; Zhao, B.; Lee, D.L. Billion-scale commodity embedding for e-commerce recommendation in alibaba. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 839–848. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
- Wang, H.; Chen, E.; Liu, Q.; Xu, T.; Du, D.; Su, W.; Zhang, X. A united approach to learning sparse attributed network embedding. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 557–566. [Google Scholar]
- Bandyopadhyay, S.; Biswas, A.; Kara, H.; Murty, M. A multilayered informative random walk for attributed social network embedding. In ECAI 2020; IOS Press: Clifton, VA, USA, 2020; pp. 1738–1745. [Google Scholar]
- Wang, H.; Lian, D.; Tong, H.; Liu, Q.; Huang, Z.; Chen, E. Decoupled representation learning for attributed networks. IEEE Trans. Knowl. Data Eng. 2021, 35, 2430–2444. [Google Scholar] [CrossRef]
- Gao, H.; Ji, S. Graph u-nets. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 2083–2092. [Google Scholar]
- Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.A. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
- Wang, C.; Pan, S.; Hu, R.; Long, G.; Jiang, J.; Zhang, C. Attributed graph clustering: A deep attentional embedding approach. arXiv 2019, arXiv:1906.06532. [Google Scholar]
- Xu, H.; Xia, W.; Gao, Q.; Han, J.; Gao, X. Graph embedding clustering: Graph attention auto-encoder with cluster-specificity distribution. Neural Netw. 2021, 142, 221–230. [Google Scholar] [CrossRef] [PubMed]
- Jin, D.; Liu, Z.; Li, W.; He, D.; Zhang, W. Graph convolutional networks meet markov random fields: Semi-supervised community detection in attribute networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 152–159. [Google Scholar]
- Shi, H.; Fan, H.; Kwok, J.T. Effective decoding in graph auto-encoder using triadic closure. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 906–913. [Google Scholar]
- Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning. PMLR, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
- Yang, L.; Cao, X.; He, D.; Wang, C.; Wang, X.; Zhang, W. Modularity based community detection with deep learning. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016; Volume 16, pp. 2252–2258. [Google Scholar]
- He, D.; Song, Y.; Jin, D.; Feng, Z.; Zhang, B.; Yu, Z.; Zhang, W. Community-centric graph convolutional network for unsupervised community detection. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3515–3521. [Google Scholar]
- Waikhom, L.; Patgiri, R. Recurrent convolution based graph neural network for node classification in graph structure data. In Proceedings of the 2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Virtual, 27–28 January 2022; pp. 201–206. [Google Scholar]
- Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 849–856. [Google Scholar]
- Pan, S.; Hu, R.; Long, G.; Jiang, J.; Yao, L.; Zhang, C. Adversarially regularized graph autoencoder for graph embedding. arXiv 2018, arXiv:1802.04407. [Google Scholar]
- Zheng, S.; Zhu, Z.; Zhang, X.; Liu, Z.; Cheng, J.; Zhao, Y. Distribution-induced bidirectional generative adversarial network for graph representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7224–7233. [Google Scholar]
- Zhu, H.; Koniusz, P. Simple spectral graph convolution. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Mrabah, N.; Bouguessa, M.; Touati, M.F.; Ksantini, R. Rethinking graph auto-encoder models for attributed graph clustering. IEEE Trans. Knowl. Data Eng. 2022, 35, 9037–9053. [Google Scholar] [CrossRef]
- Sun, D.; Li, D.; Ding, Z.; Zhang, X.; Tang, J. Dual-decoder graph autoencoder for unsupervised graph representation learning. Knowl. Based Syst. 2021, 234, 107564. [Google Scholar] [CrossRef]
- Ahn, S.J.; Kim, M. Variational graph normalized autoencoders. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual, 1–5 November 2021; pp. 2827–2831. [Google Scholar]
- Gan, G.; Ma, C.; Wu, J. Data Clustering: Theory, Algorithms, and Applications; SIAM: Philadelphia, PA, USA, 2020. [Google Scholar]
- Xie, Y.; Xu, Z.; Zhang, J.; Wang, Z.; Ji, S. Self-supervised learning of graph neural networks: A unified review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2412–2429. [Google Scholar] [CrossRef] [PubMed]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Wang, J.; Liang, J.; Yao, K.; Liang, J.; Wang, D. Graph convolutional autoencoders with co-learning of graph structure and node attributes. Pattern Recogn. 2022, 121, 108215. [Google Scholar]
- Van Der Maaten, L. Accelerating t-SNE using tree-based algorithms. J. Mach. Learn. Res. 2014, 15, 3221–3245. [Google Scholar]
Dataset | Nodes | Edges | Features | Classes |
---|---|---|---|---|
Cora | 2708 | 5429 | 1433 | 7 |
Citeseer | 3327 | 4732 | 3703 | 6 |
Pubmed | 19,717 | 44,338 | 500 | 3 |
Methods | Cora | Citeseer | Pubmed | ||||||
---|---|---|---|---|---|---|---|---|---|
ACC | NMI | ARI | ACC | NMI | ARI | ACC | NMI | ARI | |
SC | 0.367 | 0.127 | 0.031 | 0.239 | 0.056 | 0.010 | 0.403 | 0.042 | 0.002 |
DeepWalk | 0.529 | 0.384 | 0.291 | 0.390 | 0.131 | 0.137 | 0.684 | 0.279 | 0.299 |
GAE | 0.611 | 0.482 | 0.302 | 0.456 | 0.221 | 0.191 | 0.632 | 0.249 | 0.246 |
VGAE | 0.592 | 0.408 | 0.347 | 0.467 | 0.261 | 0.206 | 0.619 | 0.216 | 0.201 |
ARGA | 0.640 | 0.449 | 0.352 | 0.573 | 0.350 | 0.341 | 0.681 | 0.276 | 0.291 |
ARVGA | 0.638 | 0.450 | 0.374 | 0.544 | 0.261 | 0.245 | 0.513 | 0.117 | 0.078 |
ARGA-AX | 0.597 | 0.455 | 0.366 | 0.547 | 0.263 | 0.243 | 0.637 | 0.245 | 0.226 |
ARVGA-AX | 0.711 | 0.526 | 0.495 | 0.581 | 0.338 | 0.301 | 0.640 | 0.239 | 0.226 |
DBGAN | 0.748 | 0.576 | 0.540 | 0.670 | 0.407 | 0.414 | 0.694 | 0.324 | 0.327 |
SSGC | 0.696 | 0.547 | 0.474 | 0.688 | 0.428 | 0.445 | 0.710 | 0.332 | 0.346 |
R-GAE | 0.658 | 0.516 | 0.441 | 0.501 | 0.246 | 0.200 | 0.696 | 0.314 | 0.316 |
R-VGAE | 0.713 | 0.498 | 0.480 | 0.449 | 0.199 | 0.125 | 0.692 | 0.303 | 0.309 |
DGA | 0.767 | 0.586 | 0.587 | 0.698 | 0.441 | 0.457 | 0.672 | 0.263 | 0.267 |
Methods | Cora | Citeseer | Pubmed | |||
---|---|---|---|---|---|---|
AUC | AP | AUC | AP | AUC | AP | |
SC | 0.846 | 0.885 | 0.805 | 0.850 | 0.842 | 0.878 |
DeepWalk | 0.831 | 0.850 | 0.805 | 0.836 | 0.844 | 0.841 |
GAE | 0.910 | 0.920 | 0.895 | 0.899 | 0.964 | 0.965 |
VGAE | 0.914 | 0.926 | 0.908 | 0.920 | 0.944 | 0.947 |
ARGA | 0.924 | 0.932 | 0.919 | 0.930 | 0.968 | 0.971 |
ARVGA | 0.924 | 0.926 | 0.924 | 0.930 | 0.965 | 0.968 |
ARGA-AX | 0.913 | 0.913 | 0.919 | 0.934 | 0.966 | 0.967 |
ARVGA-AX | 0.902 | 0.892 | 0.898 | 0.904 | 0.967 | 0.971 |
DBGAN | 0.945 | 0.951 | 0.945 | 0.958 | 0.968 | 0.973 |
DGAE | 0.949 | 0.955 | 0.949 | 0.958 | 0.974 | 0.972 |
VGNAE | 0.954 | 0.958 | 0.970 | 0.971 | 0.976 | 0.976 |
GNAE | 0.956 | 0.957 | 0.965 | 0.970 | 0.975 | 0.975 |
DGA | 0.919 | 0.913 | 0.980 | 0.981 | 0.959 | 0.952 |
DGA_GX | 0.953 | 0.952 | 0.970 | 0.971 | 0.963 | 0.957 |
DGA_EX | 0.961 | 0.962 | 0.961 | 0.962 | 0.965 | 0.959 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sheng, J.; Yang, Z.; Wang, B.; Chen, Y. Attribute Graph Embedding Based on Multi-Order Adjacency Views and Attention Mechanisms. Mathematics 2024, 12, 697. https://doi.org/10.3390/math12050697
Sheng J, Yang Z, Wang B, Chen Y. Attribute Graph Embedding Based on Multi-Order Adjacency Views and Attention Mechanisms. Mathematics. 2024; 12(5):697. https://doi.org/10.3390/math12050697
Chicago/Turabian StyleSheng, Jinfang, Zili Yang, Bin Wang, and Yu Chen. 2024. "Attribute Graph Embedding Based on Multi-Order Adjacency Views and Attention Mechanisms" Mathematics 12, no. 5: 697. https://doi.org/10.3390/math12050697
APA StyleSheng, J., Yang, Z., Wang, B., & Chen, Y. (2024). Attribute Graph Embedding Based on Multi-Order Adjacency Views and Attention Mechanisms. Mathematics, 12(5), 697. https://doi.org/10.3390/math12050697