Research Article  Open Access
H. Zhang, J. J. Zhou, R. Li, "Enhanced Unsupervised Graph Embedding via Hierarchical Graph Convolution Network", Mathematical Problems in Engineering, vol. 2020, Article ID 5702519, 9 pages, 2020. https://doi.org/10.1155/2020/5702519
Enhanced Unsupervised Graph Embedding via Hierarchical Graph Convolution Network
Abstract
Graph embedding aims to learn the lowdimensional representation of nodes in the network, which has been paid more and more attention in many graphbased tasks recently. Graph Convolution Network (GCN) is a typical deep semisupervised graph embedding model, which can acquire node representation from the complex network. However, GCN usually needs to use a lot of labeled data and additional expressive features in the graph embedding learning process, so the model cannot be effectively applied to undirected graphs with only network structure information. In this paper, we propose a novel unsupervised graph embedding method via hierarchical graph convolution network (HGCN). Firstly, HGCN builds the initial node embedding and pseudolabels for the undirected graphs, and then further uses GCNs to learn the node embedding and update labels, finally combines HGCN output representation with the initial embedding to get the graph embedding. Furthermore, we improve the model to match the different undirected networks according to the number of network node label types. Comprehensive experiments demonstrate that our proposed HGCN and HGCN can significantly enhance the performance of the node classification task.
1. Introduction
Graph embedding aims to map highdimensional sparse network data to lowdimensional dense realvalue vector space, which can adaptively extract features and be applied to the network analysis tasks, such as node classification, link prediction, and dimensional visualization [1, 2]. Generally, graph embedding models mainly include the shallow embedding model and the deep embedding model. The shallow embedding model represented by Deepwalk is directly learned by defining the loss function for network representation, which is usually an unsupervised learning method [3]. This model is suitable for the undirected graph, but it also has the disadvantage of ignoring the feature information in graphs. On the contrary, the deep embedding model represented by the GCNs uses the additional feature and label information of the graph at the same time, which is generally a semisupervised or supervised learning method. However, the deep embedding model is not more suitable for undirected graphs with only structure information, at the same time many real networks do not include the node features and labels due to privacy protection and annotation difficulty, so the study of deep unsupervised graph embedding method for a real undirected graph is urgent.
To help address these issues, we propose a novel unsupervised graph embedding method with a hierarchical graph convolution network, which does not need the extra features and labeled data. The general process is shown in Figure 1, which includes 3 layers with the initial, update, and output layer. Among them, the input of the model is a graph, the output is the lowdimensional vector representation of nodes that relate to the number of label types, and the detailed structure of the model is partly omitted.
Our contributions are the following: (1) we propose an enhanced hierarchical graph convolution network for the undirected graph with only structure information. (2) According to the number of label types in the different experimental data sets, we introduce maxpooling to improve the proposed model. (3) We conduct the node classification experiments on Wikipedia, American airtraffic, Cora, and Citeseer data sets separately, and our proposed method achieves significant improvements than other stateoftheart baselines.
2. Related Work
Graph embedding methods have different classification systems. In this section, firstly, we introduce the traditional graph embedding methods. Then, we review the different graph embedding methods in the field of machine learning. In the end, we classify the graph embedding models according to the structure of the embedding methods, and a profound analysis of the disadvantages of the existing graph embedding methods is given. Additionally, we further introduce existing GCN variants for graphs.
2.1. Traditional Graph Embedding Methods
Traditional network representation learning methods are realized by dimension reduction technology [4]. Classic dimension reduction methods include Principal Component Analysis (PCA) [5] and Multidimensional Scale (MDS) [6]. Both methods can capture linear structure information, but they cannot acquire a nonlinear structure in input data.
From a linear algebraic perspective, all unsupervised graph embedding methods are generally represented by the various graph matrices, especially the Laplace operator and the adjacent matrix [7]. In terms of computational efficiency, the feature decomposition of the data matrix is expensive.
2.2. Graph Embedding Methods for Machine Learning
In the representation learning for machine learning, the semisupervised graph embedding method requires a set of features that can distinguish nodes [8]. Typical model mainly uses manual feature extraction to learn node representation for specific professional fields, which has the disadvantages of the inaccuracy and poor robustness. Another method mainly learns node representation such as MMDW [9] by solving optimization problems, which improves the accuracy of the extracted feature. However, the number of estimation parameters is large and the time complexity is high [10].
As for the unsupervised graph embedding method, to balance accuracy and computational efficiency, the feature representation needs to define an objective function independent of downstream tasks, such as Deepwalk and LINE. The above unsupervised graph embedding method only uses network structure information to obtain lowdimensional embedding. Moreover, the graph embedding method also learns from the rich content of node attribute and edge attribute. TADW [11] incorporates text features of vertices into network representation learning under the framework of matrix factorization. CENE [12] integrates text modelling and structure modelling in a general framework by treating the content information as a special kind of node. HSCA [13] is a network embedding method of attribution graph, which simulates homophone, network topology, and node characteristics at the same time. Inspired by pretraining methods, pretraining models based on GCN are proposed. DeepGraphInfoMax [14] relies on maximizing mutual information between patch representations and corresponding highlevel summaries of graphs, while both of them derived using established graph convolutional network architectures. The pretrained GCN model proposed by Hu et al. [15] can capture generic graph structural information that is transferable across tasks.
Compared with the methods of manually extracting features, the representation method by optimizing objective function can obtain more comprehensive features, which is closely related to the downstream predicting task [16]. Therefore, the unsupervised graph embedding method improves the disadvantages of semisupervised machine learning method, such as difficult extensibility and high training complexity.
2.3. Graph Embedding Methods with Different Structures
According to the structure of the representation methods [17], graph embedding methods are divided into shallow embedding method and deep embedding method. In the shallow embedding model, inspired by the successful application of skipgram [18] in natural language processing [19], a series of skipgram based models are proposed to encode network structures into continuous spatial vector representations in recent years, such as Deepwalk, LINE, and Node2vec.
Deepwalk [20] simulates word sequences to generate node sequences by random walk from each node, which forms a “corpus” based on those node sequences. Furthermore, Deepwalk sets the size of the background window and then imports the “corpus” into the skipgram model to get the node representation. The firstorder similarity of direct connection and the secondorder similarity of shared neighbor nodes are optimized respectively, and the two similarities are combined at the output of the nodes in LINE [21]. Node2vec [22] uses two additional parameters to control the direction of the random walk on the generation step of the “corpus” of Deepwalk. Struc2vec [23] defines the nodes that are not structurally adjacent but have the same structural roles. The shallow embedding models usually use unsupervised machine learning methods, so they have the disadvantage of ignoring the node features.
In the deep embedding model, the mainstream methods use the deep learning model to capture the nonlinear relationship between nodes [24]. Typical deep learning models are NEFLGC, SDNE, GCN, and a series of GCN variants.
NEFLGC [25] studies the problem of representation learning for network with textual information, which aims to learn lowdimensional vectors for nodes by leveraging network structure and textual information. SDNE [26] automatically captures the local relationship of the nodes by using the unsupervised learning method and takes the secondorder neighbor of the nodes as the input to learn the lowdimensional representation of graphs, but the model still does not consider the node features.
GCN [27] is a deep semisupervised graph embedding model with incorporating the extra feature and labeled data into the graph embedding learning process, which cannot be easily applied to the undirected graph with only structure information. In terms of applicability, the existing GCN variants are mainly studied in two aspects. One variant assumes the graph is attributed, such as GraphSAGE [28], GAT [29], NGCN [30], and FastGCN [31]. GraphSAGE [28] can be used to generate node embeddings for previously unseen nodes or entirely new input graphs, as long as these graphs have the same attribute schema as the training data. GAT [29] uses attention mechanism to address the shortcomings of prior methods based on graph convolutions or their approximations. NGCN [30] improves the scalability of GCN on the whole graph by setting the size of the convolution kernel. FastGCN [31] is a batch training algorithm combining importance sampling, and it can not only make GCN training more efficient but also generalize well for inference. Generally, they take the structure information or the node degree information as the feature and then use the correct label to train GCN to learn the graph embedding. However, the model still needs to incorporate the correct labeled data into the graph embedding processing. Another variant is proposed to solve the label problem, such as M3S and GMNN. M3S [32] uses the correct extra feature and enlarges the labeled data by selftraining to learn the undirected graph embedding. GMNN [33] similarly applies the correct extra feature and neighbor nodes to generate pseudolabels for unsupervised learning. As a consequence, the previous method still cannot consider the feature and the label of the node at the same time.
3. Model
In this section, we describe the proposed method and give a detailed framework for the model. Firstly, we generate the initial node embedding using Deepwalk. Simultaneously, the pseudolabels are built by the AP clustering algorithm [34]. Then, we build the hierarchical GCN to update the node embedding and labels. Finally, we introduce maxpooling to enhance our model according to the number of label types in the different experimental datasets.
3.1. Initial Node Embedding
The first step of the model is to generate the initial node embedding using Deepwalk. Deepwalk is the shallow embedding model with random walking, and it can better learn the global structure and the context of the node in the undirected graph. Therefore, Deepwalk is used as a pretraining method to generate the initial node embedding in this paper.
Let be a given undirected graph with vertex set and edge set , where denotes the number of nodes in the graphs. Our formal definition is general and can apply to any undirected and unweighted network. Firstly, we acquire a specific length random walk sequence for the node , which is shown in the following formula (1):where is the node. And then, we use a language model to learn from these random walk sequences and denote the initial embedding as follows:where is the word sequence in the language model, and is the language model. is denoted as the output function. We use to denote the word sequence, where is a specific word in the language model. The detailed processing flow is shown in Figure 2.
3.2. Pseudolabels
The second step of the model is to construct the pseudolabels and predict labels with GCN. After obtaining the initial node embedding of the undirected graph, we use the clustering method to acquire pseudolabel for the unlabeled node , where and is a pseudolabel set. The specific processing flow is shown in Figure 3.
And now, we update to denote the graph, as shown in the following formula (3):
In this paper, AP clustering model called nearneighbor propagation clustering algorithm is adopted to generate pseudolabels, which is mainly because the method contains the following three advantages. (1) Because the number of label categories is unknown, the model needs to choose a clustering method that does not need to specify the number of clusters. Unlike kmeans and kcenter algorithms, AP clustering model is not necessary to set the final number of clusters when clustering. (2) The cluster centers of AP algorithm are the actual nodes in the data set, which are the representative of each class, so AP clustering model is not sensitive to the initial negative value of the node embedding. (3) If the sum of squares of errors is used to measure the performance of the algorithms, the sum of squares of errors of AP clustering is lower than that of other methods. This evaluation index shows that AP clustering method is effective.
Based on this, we further use GCN to predict the label for the unlabeled node and define the predicted maximum value as the label , where and is the predication label set. And now, formula (3) is updated as following in formula (4):
3.3. Hierarchical GCN
In this section, we focus on the hierarchical GCN model, which includes two GCN [27] models. At the start, we introduce GCN. Next, we set up hierarchical graph convolution network to propose the HGCN model and the HGCN model by analyzing the updated undirected graphs. Additionally, we further explain the reason for the improved model.
3.3.1. Graph Convolution Network
Different from the conventional convolutional neural networks (CNN), the convolution operation for a graph is defined as the weighted average of neighbors of one particular node. Mathematically, the graph convolution layer is defined aswhere is the normalized Laplacian matrix of the graph G. is the input of the hidden layer of GCN, i.e., the output of the hidden layer. is the weight matrix in the hidden layer that would be trained. is the adjacency matrix of the undirected graph. And is the activation function. For any given node, a GCN layer aggregates the previous layer’s embedding of its neighbor with , followed by linear transformation W and nonlinear activation σ, so as to obtain a contextualized node representation. We denote as Llayer GCNs, parametrized by . For each given graph G = (V, A) with input features , can get node representations by:
3.3.2. HGCN and HGCN
After the initial node embedding section and the pseudolabels section, we have learned the graph information from formula (4). After predicting labels, we also update the graph embedding with GCN, which is as shown in formula (7):where is the updated node embedding by the first GCN.
According to formula (6), is defined as formula (8):where is the input feature, which is the initial node embedding using Deepwalk. In this paper, we use 3layer GCNs with 128dimensional hidden vectors to accomplish the pretraining and adaptation.
And then, we further use the second GCN to learn the undirected graph again. The final embedding is denoted as formula (9):
Moreover, after this second learning process, GCN will lead to too smooth for the embedding, which can affect node classification. Therefore, to relieve smoothing, we combine the secondary learning representation with the initial node embedding . Finally, we get the ultimate embedding for the undirected graph as shown in formula (10):where is a vector splicing operation.
The previous HGCN method can be applied to learn a better embedding, when the graph is an undirected graph with a few numbers of node label types. However, when the undirected graph has a large number of node label types, the proposed model will also decompose labels into multidimensional features and add them to the learning processing, which can lead to the feature redundancy. Therefore, motivated by this analysis, we introduce maxpooling to improve the embedding model for the graph with a large number of label types, namely HGCN. The exact processing flow is shown in Figure 4.
4. Experiments
4.1. Datasets
To verify the effectiveness of the proposed model, we conduct several comparative experiments on four open available graph datasets:(i)Cora dataset [35] consists of 2,708 machine learning papers classified into one of seven classes and 5,429 links between them.(ii)Citeseer dataset [35] is a link dataset built with permission from the Citeseer web database. It contains 3,327 publications from six classes and 4,732 links among them.(iii)Wikipedia dataset [36] consists of 2,405 Wikipedia pages from 17 categories and 17,981 links between them. It is much denser than Citeseer and Cora datasets.(iv)American airtraffic dataset [23] is from the United States Air Traffic Network. It contains 1190 airports and 13599 relationships between airports and airports, and the airport labels are divided into four groups.
All specific nodes and edges in our experimental datasets are shown in Table 1.

Note that all datasets are used as undirected graphs with only network structure information and the correct labels in the datasets are only used in the node classification verification task in this paper.
4.2. Baselines
For comparison, we adopt six kinds of baselines as follows:(i)Deepwalk [20] is a typical unsupervised graph embedding method which adopts the skipgram language model.(ii)LINE [21] is also a popular unsupervised method which considers the firstorder and secondorder proximity information.(iii)Node2vec [22] learns lowdimensional representations for nodes in a graph by optimizing a neighborhood preserving objective, which explores the structure of the network by controlling the and parameters, and learns the dimension feature representation by simulating biased random walks.(iv)GraRep [37] captures the longdistance nodes and uses the matrix decomposition method to learn the d dimension feature representation of the node.(v)Hope [38] introduces higherorder similarity and uses the matrix decomposition method to learn the dimension feature representation of the node.(vi)SDNE [26] learns the dimension node representation by capturing the network nonlinear structure using multilayer nonlinear functions, which is a semisupervised deep graph embedding model.
Several existing methods include shallow embedding models, matrix decomposition models, and deep embedding models. Note that since some existing work ([27–29]) utilizes extra node features, we do not compare with them directly.
4.3. Node Classification Experiments
To verify the effectiveness, we follow the same settings used in Hamilton et al. (2016) [27], including the following benchmark tasks: (1) using HGCN to classify nodes on American airtraffic, Cora, and Citeseer datasets; (2) using HGCN to classify nodes on Wikipedia dataset.
4.3.1. Experimental Settings
In this experiment, MicroF1, MacroF1, and Accuracy are used as measurements. The model uses the same proportion of training and test data as the existing published papers, i.e., 50% of the training data and 50% of the test data.
To facilitate the comparison, on the Cora and Citeseer datasets, we report the results of one experiment for comparison with the baseline models. However, on the American airtraffic network dataset, we repeat the experiment 5 times using random samples for training and demonstrate the average performance. The experimental results are given in Table 2. “” denotes the published results of the existing papers [29], and the roughened numbers represent the best results.

On the Wikipedia dataset, we use the same measurements to compare with the results of opensource framework OpenNE [39]. The best experimental results are presented in Table 3.
 
The result with is from the opensource OpenNE [39]. 
4.3.2. Results and Discussion
As shown in Table 2, it can be seen that our proposed model achieves the best results compared to all shallow baseline models, and the outcomes are very close to the deep baseline model SDNE. On the Cora and Citeseer datasets, the experimental results of our model are higher than all baseline models in all measurements, which show that the model can effectively learn the undirected graphs.
Obviously, on the American airtraffic dataset, the performances of HGCN are close to the results of SDNE. The main reason is that the nodes on American airtraffic dataset are labeled by the network global structure. In other words, the node label of the dataset is more related to their structure identity than the labels of their neighbors. HGCN uses the label information of neighbor nodes to update and learn the labels for the unlabeled data, so the node classification performances of HGCN are significantly affected. To visually display the performances of HGCN, the accuracy is presented in Figure 5.
As illustrated in Figure 5, we compared the accuracy of all baselines including the shallow and the deep embedding model on the three undirected graph datasets. The average performances of the shallow embedding model are higher than SDNE model on the Cora dataset and the Citeseer dataset. And the result of the SDNE is better than the other shallow baselines on the American airtraffic dataset. This suggests that SDNE can learn well the special structure network, which node labels are related according to the network global structure. While the proposed model HGCN has a significant increase on most experimental datasets, which further verifies the applicability of our model on undirected graphs.
As for the improved model according to the number of label types, we verify its effectiveness on the Wikipedia dataset and compare it with the results of the opensource framework OpenNE [39]. The experimental results are shown in Table 3.
According to the results of Table 3, when the HGCN model uses only vector splicing, the performance degrades in MicroF1. In MacroF1, the results are better than the other all baseline models except for Deepwalk. That suggests that the HGCN model breaks up pseudolabels into the features, while the Wikipedia dataset leads to features redundant because of more label types.
To help address this problem, the maxpooling layer is introduced to the HGCN model, which can select the features to improve the effect of node classification on the Wikipedia dataset. As can be seen from the experimental results in Table 3, compared to Deepwalk, LINE (2nd), and Node2vec, the MicroF1 of HGCN is improved by 4.8%, 14.1%, and 6.6%. And in MacroF1, the results are enhanced by 3%, 20.3%, and 4.9%, separately. Compared to GraRep, HOPE, and SDNE, the MicroF1 of HGCN is improved by 8.4%, 11.6%, and 7.4%, and in the MacroF1, the results are improved by 11.4%, 15.2%, and 9.2%, separately.
To further illustrate the interesting trends, we plot the MicroF1 and MacroF1 to investigate the changes of the different baseline methods in Figure 6.
As can be seen from the results in Figure 6, compared to shallow embedding models, matrix decomposition models, and deep embedding models, the MicroF1 and MacroF1 of HGCN are all improved on the Wikipedia dataset. Figure 6 also shows that the MicroF1 and MacroF1 of HGCN have been significantly improved by using the maxpooling. The experimental results further indicate that the improved model for the dataset with a large number of label types can get better graph embedding effectively.
5. Conclusion
In this study, we explore an unsupervised graph embedding method for the undirected graph. Comparing with conventional graph embedding models, we introduce hierarchical graph convolution network to propose the HGCN and HGCN methods respectively. Besides, we come to the following conclusions:(1)In this paper, a hierarchical GCN for the undirected graph with only structure information is proposed. The model does not need the extra features and labeled data, which improves the applicability of the GCN in the real network.(2)Moreover, we introduce maxpooling to improve the proposed model according to the number of label types. Experimental results show that our model achieves considerable improvement than all the baselines.
However, due to the limitations of the pretraining and the clustering methods, the proposed model in this paper still has the disadvantage of obtaining the initial node embedding and the poor label accuracy. In the future, we will further focus on dealing with the problems and plan to explore the impact of initialing models and clustering methods on the proposed model, which ultimately can provide a better method of unsupervised graph embedding.
Data Availability
All experimental data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
Thanks are due to the teachers and classmates of the project team for their guidance and help. This research was supported by the National Natural Science Fund of China (no. 61936012 and 61806117), the National Social Science Fund of China (no. 18BYY074), the Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi (No. 201802012), and the Innovation Project for College Graduates of Shanxi Province (no. 2020SY015).
References
 Y. Cheng, D. Wang, P. Zhou, and T. Zhang, “A survey of model compression and acceleration for deep neural networks,” 2017, https://arxiv.org/abs/1710.09282. View at: Google Scholar
 G. Palash and F. Emilio, “Graph embedding techniques, applications, and performance: a survey,” KnowledgeBased Systems, vol. 151, pp. 78–94, 2018. View at: Publisher Site  Google Scholar
 K. S. Mehran, G. Rishab, J. Kshitij et al., “Relational representation learning for dynamic (knowledge) graphs: a survey,” 2019, https://arxiv.org/abs/1905.11485. View at: Google Scholar
 I. K. Fodor, A Survey of Dimension Reduction Techniques, Neoplasia, Amsterdam, Netherlands, 2002.
 S. Wold, K. Esbensen, and P. Geladi, “Principal component analysis,” Chemometrics and Intelligent Laboratory Systems, vol. 2, no. 1–3, pp. 37–52, 1987. View at: Publisher Site  Google Scholar
 J. B. Kruskal, M. Wish, and E. M. Uslaner, Multidimensional Scaling, Methods, vol. 116, SAGE, New York, NY, USA, 1978.
 M. Belkin and P. Niyogi, “Laplacian eigenmaps and spectral techniques for embedding and clustering,” Advances in Neural Information Processing Systems, vol. 14, no. 6, pp. 585–591, 2002. View at: Google Scholar
 W. L. Hamilton, R. Ying, and J. Leskovec, “Representation learning on graphs: methods and applications,” 2017, https://arxiv.org/abs/1709.05584. View at: Google Scholar
 C. C. Tu, W. C. Zhang, Z. Y. Liu, and S. M. Sun, “Maxmargin deepwalk: discriminative learning of network representation,” in Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), pp. 3889–3895, New York, NY, USA, July 2016. View at: Google Scholar
 Y. Bengio, A. Courville, and P. Vincent, “Representation learning: a review and new perspectives,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 8, pp. 1798–1828, 2013. View at: Publisher Site  Google Scholar
 C. Yang, Z. Y. Liu, D. L. Zhao, M. S. Sun, and E. Y. Chang, “Network representation learning with rich text information,” in Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), pp. 2111–2117, Buenos Aires, Argentina, JulyAugust 2015. View at: Google Scholar
 X. F. Sun, J. Guo, X. Ding, and T. Liu, “A general framework for contentenhanced network representation learning,” 2016, https://arxiv.org/abs/1610.02906. View at: Google Scholar
 D. K. Zhang, J. Yin, X. Q. Zhu, and C. Q. Zhang, “Homophily, structure, and content augmented network representation learning,” in Proceedings of the International Conference on Data Mining (ICDM), pp. 609–618, Barcelona, Spain, December 2016. View at: Google Scholar
 P. Velickovic, W. Fedus, W. L. Hamilton, P. Lio, Y. Bengio, and R. D. Hjelm, “Deep graph infomax,” in Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, May 2019. View at: Google Scholar
 Z. N. Hu, C. J. Fan, T. Chen, K. W. Chang, and Y. Z. Sun, “Pretraining graph neural networks for generic structural feature extraction,” 2019, https://arxiv.org/abs/1905.13728. View at: Google Scholar
 J. Pennington, R. Sorcher, and C. D. Manning, “Glove: global vectors for word representation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, October 2014. View at: Google Scholar
 J. Zhou, G. Q. Cui, Z. Y. Zhang et al., “Graph neural networks: a review of methods and applications,” 2018, https://arxiv.org/abs/1812.08434. View at: Google Scholar
 T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed representations of words and phrases and their compositionality,” in Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), pp. 3111–3119, Lake Tahoe, NV, USA, December 2013. View at: Google Scholar
 T. Tang and H. Liu, “Leveraging social media networks for classification,” Data Mining and Knowledge Discovery, vol. 23, no. 3, pp. 447–478, 2011. View at: Publisher Site  Google Scholar
 B. Perozzi, R. Alrfou, and S. Skiena, “Deepwalk: online learning of social representations,” in Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 701–710, New York, NY, USA, August 2014. View at: Google Scholar
 J. Tang, M. Qu, M. Z. Wang, M. Zhang, J. Yan, and Q. Z. Mei, “Line: largescale information network embedding,” in Proceedings of the 24th International Conference on World Wide Web (WWW), pp. 1067–1077, Florence, Italy, May 2015. View at: Google Scholar
 A. Grover and J. Leskovec, “Node2vec: scalable feature learning for networks,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 855–864, San Francisco, CA, USA, August 2016. View at: Google Scholar
 L. F. R. Ribeiro, P. H. P. Saverese, and D. R. Figueiredo, “Struc2vec: learning node representations from structural identity,” in Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 385–394, Halifax, Canada, August 2017. View at: Google Scholar
 J. Y. Xie, R. Girshick, and A. Farhadi, “Unsupervised deep embedding for clustering analysis,” in Proceedings of International Conference on Machine Learning (ICML), pp. 478–487, New York, NY, USA, June 2016. View at: Google Scholar
 H. Y. Xu, H. T. Liu, W. J. Wang, Y. H. Sun, and P. F. Jiao, “NEFLGC: network embedding based on fusing local (FirstOrder) and global (SecondOrder) network structure with node content,” PacificAsia Conference on Knowledge discovery and data mining (PAKDD), pp. 260–271, 2018. View at: Publisher Site  Google Scholar
 D. X. Wang, P. Cui, and W. W. Zhu, “Structural deep network embedding,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 1225–1234, San Francisco, CA, USA, August 2016. View at: Google Scholar
 T. Kipf and M. Welling, “Semisupervised classification with graph convolutional networks,” 2016, https://arxiv.org/abs/1609.02907. View at: Google Scholar
 W. L. Hamilton, Z. T. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in Proceedings of Annual Conference Neural Information Processing Systems (NIPS), pp. 1024–1034, Long Beach, CA, USA, December 2017. View at: Google Scholar
 P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y. Bengio, “Graph attention networks,” 2017, https://arxiv.org/abs/1710.10903. View at: Google Scholar
 S. Abuelhaija, A. Kapoor, B. Perozzi, and J. Lee, “NGCN: multiscale graph convolution for semisupervised node classification,” 2018, https://arxiv.org/abs/1802.08888. View at: Google Scholar
 J. Chen, T. F. Ma, and C. Xiao, “FastGCN: fast learning with graph convolutional networks via importance sampling,” 2018, https://arxiv.org/abs/1801.10247. View at: Google Scholar
 K. Sun, Z. X. Zhu, and Z. C. Lin, “Multistage selfsupervised learning for graph convolutional networks,” 2019, https://arxiv.org/abs/1902.11038. View at: Google Scholar
 M. Qu, Y. Bengio, and J. Tang, “GMNN: graph markov neural networks,” 2019, https://arxiv.org/abs/1905.06214. View at: Google Scholar
 U. Bodenhofer, A. Kothmeier, and S. Hochreiter, “AP cluster: an R package for affinity propagation clustering,” Bioinformatics, vol. 27, no. 17, pp. 24632464, 2011. View at: Publisher Site  Google Scholar
 A. K. McCallum, K. Nigam, J. Rennie, and K. Seymore, “Automating the construction of internet portals with machine learning,” Information Retrieval, vol. 3, no. 2, pp. 127–163, 2000. View at: Publisher Site  Google Scholar
 P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Galligher, and T. EliassiRad, “Collective classification in network data,” AI Magazine, vol. 29, no. 3, pp. 93–106, 2008. View at: Publisher Site  Google Scholar
 S. S. Cao, W. Lu, and Q. K. Xu, “Grarep: learning graph representations with global structural information,” in Proceedings of the 24th ACM International on Conference on Information and Knowledge Management (CIKM), pp. 891–900, Melbourne Australia, October 2015. View at: Google Scholar
 M. D. Ou, P. Cui, J. Pei, Z. W. Zhang, and W. W. Zhu, “Asymmetric transitivity preserving graph embedding,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pp. 1105–1114, San Francisco, CA, USA, August 2016. View at: Google Scholar
 Qinghua University, OpenNE, https://github.com/thunlp/OpenNE.
Copyright
Copyright © 2020 H. Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.