Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks
Abstract
:1. Introduction
2. Related Works
2.1. Structure Fusion
2.2. Graph Neural Networks
2.3. Node Classification Based on GCN
3. Graph Convolutional Networks
4. Multiple Structures’ Fusion
4.1. Specificity Loss of Multiple Structures
4.2. Commonality Loss of Multiple Structures
4.3. Structure Fusion by Structure Metric Losses
Algorithm 1 Fusion structure of multiple structures. |
Input: : adjacency matrices of graph ; : regularization parameter of the total loss; T: the iteration times |
Output: W: fusion structure of multiple structures |
|
5. Experiments
5.1. Datasets
5.2. Experimental Configuration
5.3. Comparison with the Base-Line Methods
5.4. Structure Fusion Generalization
5.5. Comparison with the State-of-the-Art
5.6. Incomplete Structure Influence
5.7. Experimental Results’ Analysis
- The performance of SF-GCN was superior to that of the base-line methods (multi-GCN [6], GCN [27] for multi-views, and GCN [27] for View 1 and View 2 in Section 5.3). GCN [27] constructed a general graph convolutional architecture by the first-order approximation of spectral graph convolutions for greatly improving the computation efficiency of graph convolutional networks and also provided a feasible deep mining framework for effective semi-supervised classification. For using multiple structures, GCN for multi-views could input a sparse block-diagonal matrix, each block of which corresponded to the different structures. Therefore, the relationship of each block (the different structure) was ignored for GCN, and this point lead to the poor performance (for some times, the performance of GCN for multi-view was worse than that of GCN for View 1) of GCN for multi-views. In contrast, multi-GCN [6] could capture the relationship of the different structures to preserve the significant structure of merging the subspace. However, multi-GCN [6] neglected optimizing the fusion relationship of the different structures, while the proposed SF-GCN focused on finding these relationships by jointly considering the commonality and specificity loss of multiple structures for obtaining better performance for node classification.
- SPF-GCN showed the best performance in structure fusion generalization experiments, whereas the performance of SF-GCN was better than that of SP-GCN. The main reason was that SF-GCN emphasized the complementary information by the optimizing fusion relationship of the different structures, while SP-GCN tended toward the interactive propagation by the diffusion influence between the different structures. The complement fusion played a more important role than the interactive propagation because of the specificity structure of individual view data, but both fusion and propagation could contribute the multiple structure mining for enhancing the performance of node classification.
- The performance improvements of SPF-GCN compared with six kinds of state-of-the-art methods were respectively different. A similar performance of SPF-GCN was shown in the comparison with LGCN and DGCN on Cora and PubMed. Except these situation, the better improvement of SPF-GCN could be demonstrated for other methods. The main reasons were that LGCN could emphasize neighboring nodes’ feature fusion for the stable node representation and DGCN could correlate the local and global consistency for complementing the different structures. The proposed SPF-GCN was expected not only to capture the structure commonality for complementing the different information, but also to preserve the structure specificity for mining the discriminative information. Therefore, the proposed SPF-GCN could improve the classification performance in most experiments. In the least, the proposed SPF-GCN had a similar performance as the best performance of other methods in all experiments. In addition, the proposed SPF-GCN was based on GCN frameworks, so it had an efficient implementation like GCN. In the experiments, the computation efficiency of the proposed SPF-GCN was higher than that of the state-of-the-art methods (the details of the computation efficiency are in Section 4.2).
- The structure showed the distribution of the data and was very important for training the GCN model. The incomplete structure could evaluate the robustness of the related GCN model. We selected the classical GCN, the state-of-the-art DGCN, SF-GCN, and SPF-GCN for the robustness test. The proposed SPF-GCN showed the best performance on the three datasets. On Cora, the performance of GCN was better than DGCN, while the performance of GCN was worse than DGCN on Citeseer and PubMed. This showed that local and global consistency for fusing graph information in DGCN tended toward the unstable characteristic because of the tight constraint of incomplete structure consistency. The loose constraint of GCN for incomplete structure correlation led to the worse performance. The proposed SPF-GCN could compromise these constraints for balancing the incomplete structure information by optimizing the weight of multiple structures and also connect the different structures for complementing the different information. Therefore, the proposed SPF-GCN obtained the best performance in the experiments.
- The proposed SPF-GCN was expected to mine the commonality and the specificity of multiple structures. The commonality described the similarity characteristic of structures by the Grassmann manifold metric, while the specificity narrated the different characteristics of the structures by spectral embedding. In the proposed method, the commonality was constructed based on the specificity. Therefore, we only executed the ablation experiment for preserving the specificity loss by deleting the commonality loss from the total loss. This experiment obtained the following performance: on Cora, on Citeseer, and on PubMed. These results obviously were worse than the performance of the proposed SF-GCN and SPF-GCN, which could balance the commonality and specificity for mining the suited weight of multiple structures.
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Veksler, O. Efficient Graph Cut Optimization for Full CRFs with Quantized Edges. IEEE Trans. Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [Green Version]
- Monti, F.; Bronstein, M.; Bresson, X. Geometric matrix completion with recurrent multi-graph neural networks. Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2017; pp. 3697–3707. [Google Scholar]
- Ying, R.; He, R.; Chen, K.; Eksombatchai, P.; Hamilton, W.L.; Leskovec, J. Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; ACM: New York, NY, USA, 2018; pp. 974–983. [Google Scholar]
- Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–8 December 2016; Curran Associates, Inc.: New York, NY, USA, 2016; pp. 3844–3852. [Google Scholar]
- Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural message passing for quantum chemistry. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
- Khan, M.R.; Blumenstock, J.E. Multi-GCN: Graph Convolutional Networks for Multi-View Networks, with Applications to Global Poverty. arXiv 2019, arXiv:1901.11213. [Google Scholar] [CrossRef]
- Lin, G.; Zhu, H.; Kang, X.; Fan, C.; Zhang, E. Multi-feature structure fusion of contours for unsupervised shape classification. Pattern Recognit. Lett. 2013, 34, 1286–1290. [Google Scholar] [CrossRef]
- Lin, G.; Zhu, H.; Kang, X.; Fan, C.; Zhang, E. Feature structure fusion and its application. Inf. Fusion 2014, 20, 146–154. [Google Scholar] [CrossRef]
- Lin, G.; Zhu, H.; Kang, X.; Miu, Y.; Zhang, E. Feature structure fusion modelling for classification. IET Image Process. 2015, 9, 883–888. [Google Scholar] [CrossRef]
- Lin, G.; Fan, G.; Yu, L.; Kang, X.; Zhang, E. Heterogeneous structure fusion for Target Recognition in infrared imagery. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015; pp. 118–125. [Google Scholar]
- Lin, G.; Fan, G.; Kang, X.; Zhang, E.; Yu, L. Heterogeneous feature structure fusion for classification. Pattern Recognit. 2016, 53, 1–11. [Google Scholar] [CrossRef]
- Lin, G.; Fan, C.; Zhu, H.; Miu, Y.; Kang, X. Visual feature coding based on heterogeneous structure fusion for image classification. Inf. Fusion 2017, 36, 275–283. [Google Scholar] [CrossRef]
- Lin, G.; Liao, K.; Sun, B.; Chen, Y.; Zhao, F. Dynamic graph fusion label propagation for semi-supervised multi-modality classification. Pattern Recognit. 2017, 68, 14–23. [Google Scholar] [CrossRef]
- Lin, G.; Chen, Y.; Zhao, F. Structure propagation for zero-shot learning. arXiv 2017, arXiv:1711.09513. [Google Scholar]
- Lin, G.; Fan, C.; Chen, W.; Chen, Y.; Zhao, F. Class label autoencoder for zero-shot learning. arXiv 2018, arXiv:1801.08301. [Google Scholar]
- Lin, G.; Chen, Y.; Zhao, F. Structure Fusion and Propagation for Zero-Shot Learning. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Guangzhou, China, 23–26 November 2018; pp. 465–477. [Google Scholar]
- Lin, G.; Chen, W.; Liao, K.; Kang, X.; Fan, C. Transfer feature generating networks with semantic classes structure for zero-shot learning. arXiv 2019, arXiv:1903.02204. [Google Scholar] [CrossRef]
- Yang, X.; Latecki, L.J. Affinity learning on a tensor product graph with applications to shape and image retrieval. In Proceedings of the IEEE CVPR 2011, Providence, RI, USA, 20–25 June 2011; pp. 2369–2376. [Google Scholar]
- Yang, X.; Prasad, L.; Latecki, L.J. Affinity learning with diffusion on tensor product graph. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 28–38. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Bai, S.; Zhou, Z.; Wang, J.; Bai, X.; Latecki, L.J.; Tian, Q. Automatic Ensemble Diffusion for 3D Shape and Image Retrieval. IEEE Trans. Image Process. 2019, 28, 88–101. [Google Scholar] [CrossRef] [PubMed]
- Li, Q.; An, S.; Li, L.; Liu, W. Semi-supervised Learning on Graph with an Alternating Diffusion Process. arXiv 2019, arXiv:1902.06105. [Google Scholar]
- Bai, S.; Bai, X.; Tian, Q.; Latecki, L.J. Regularized diffusion process for visual retrieval. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; AAAI Press: Palo Alto, CA, USA, 2017; pp. 3967–3973. [Google Scholar]
- Bai, S.; Zhou, Z.; Wang, J.; Bai, X.; Jan Latecki, L.; Tian, Q. Ensemble diffusion for retrieval. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 774–783. [Google Scholar]
- Gao, H.; Wang, Z.; Ji, S. Large-Scale Learnable Graph Convolutional Networks. arXiv 2018, arXiv:1808.03965. [Google Scholar]
- Zhuang, C.; Ma, Q. Dual graph convolutional networks for graph-based semi-supervised classification. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, Lyon, France, 23–27 April 2018; International World Wide Web Conferences Steering Committee: Geneva, Switzerland, 2018; pp. 499–508. [Google Scholar]
- Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral networks and locally connected networks on graphs. arXiv 2013, arXiv:1312.6203. [Google Scholar]
- Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
- Chen, J.; Ma, T.; Xiao, C. Fastgcn: Fast learning with graph convolutional networks via importance sampling. arXiv 2018, arXiv:1801.10247. [Google Scholar]
- Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: New York, NY, USA, 2017; pp. 1024–1034. [Google Scholar]
- Dai, H.; Kozareva, Z.; Dai, B.; Smola, A.; Song, L. Learning steady-states of iterative algorithms over graphs. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 1114–1122. [Google Scholar]
- Liu, Z.; Chen, C.; Li, L.; Zhou, J.; Li, X.; Song, L.; Qi, Y. Geniepath: Graph neural networks with adaptive receptive paths. arXiv 2018, arXiv:1802.00910. [Google Scholar] [CrossRef] [Green Version]
- Van Tran, D.; Navarin, N.; Sperduti, A. On Filter Size in Graph Convolutional Networks. arXiv 2018, arXiv:1811.10435. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
- Chen, J.; Zhu, J.; Song, L. Stochastic Training of Graph Convolutional Networks with Variance Reduction. arXiv 2017, arXiv:1710.10568. [Google Scholar]
- Veličković, P.; Fedus, W.; Hamilton, W.L.; Liò, P.; Bengio, Y.; Hjelm, R.D. Deep Graph Infomax. arXiv 2018, arXiv:1809.10341. [Google Scholar]
- Xia, T.; Tao, D.; Mei, T.; Zhang, Y. Multiview spectral embedding. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2010, 40, 1438–1446. [Google Scholar]
- Lin, G.F.; Zhu, H.; Fan, C.X.; Zhang, E.H.; Luo, L. Multi-cluster Feature Selection Based on Grassmann Manifold. Jisuanji Gongcheng/Comput. Eng. 2012, 38, 178–181. [Google Scholar]
- Turaga, P.; Veeraraghavan, A.; Srivastava, A.; Chellappa, R. Statistical computations on Grassmann and Stiefel manifolds for image and video-based recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2273–2286. [Google Scholar] [CrossRef] [Green Version]
- Dong, X.; Frossard, P.; Vandergheynst, P.; Nefedov, N. Clustering on multi-layer graphs via subspace analysis on Grassmann manifolds. IEEE Trans. Signal Process. 2013, 62, 905–918. [Google Scholar] [CrossRef] [Green Version]
- Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A comprehensive survey on graph neural networks. arXiv 2019, arXiv:1901.00596. [Google Scholar]
Datasets | Nodes Number | Edges Number | Classes Number | Feature Dimension | Label Rate |
---|---|---|---|---|---|
Cora | 2708 | 5429 | 7 | 1433 | |
Citeseer | 3327 | 4732 | 6 | 3703 | |
Pubmed | 19,717 | 44,338 | 3 | 500 |
Method | Cora | Citeseer | PubMed |
---|---|---|---|
GCN [27] for View 1 | |||
GCN [27] for View 2 | |||
GCN [27] for multi-view | |||
Multi-GCN [6] | NA | ||
SF-GCN |
Method | Cora | Citeseer | PubMed |
---|---|---|---|
SF-GCN | |||
PF-GCN | |||
SPF-GCN |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lin, G.; Wang, J.; Liao, K.; Zhao, F.; Chen, W. Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks. Electronics 2020, 9, 432. https://doi.org/10.3390/electronics9030432
Lin G, Wang J, Liao K, Zhao F, Chen W. Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks. Electronics. 2020; 9(3):432. https://doi.org/10.3390/electronics9030432
Chicago/Turabian StyleLin, Guangfeng, Jing Wang, Kaiyang Liao, Fan Zhao, and Wanjun Chen. 2020. "Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks" Electronics 9, no. 3: 432. https://doi.org/10.3390/electronics9030432
APA StyleLin, G., Wang, J., Liao, K., Zhao, F., & Chen, W. (2020). Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks. Electronics, 9(3), 432. https://doi.org/10.3390/electronics9030432