Research Article  Open Access
Jibing Wu, Zhifei Wang, Yahui Wu, Lihua Liu, Su Deng, Hongbin Huang, "A Tensor CP Decomposition Method for Clustering Heterogeneous Information Networks via Stochastic Gradient Descent Algorithms", Scientific Programming, vol. 2017, Article ID 2803091, 13 pages, 2017. https://doi.org/10.1155/2017/2803091
A Tensor CP Decomposition Method for Clustering Heterogeneous Information Networks via Stochastic Gradient Descent Algorithms
Abstract
Clustering analysis is a basic and essential method for mining heterogeneous information networks, which consist of multiple types of objects and rich semantic relations among different object types. Heterogeneous information networks are ubiquitous in the realworld applications, such as bibliographic networks and social media networks. Unfortunately, most existing approaches, such as spectral clustering, are designed to analyze homogeneous information networks, which are composed of only one type of objects and links. Some recent studies focused on heterogeneous information networks and yielded some research fruits, such as RankClus and NetClus. However, they often assumed that the heterogeneous information networks usually follow some simple schemas, such as bityped network schema or star network schema. To overcome the above limitations, we model the heterogeneous information network as a tensor without the restriction of network schema. Then, a tensor CP decomposition method is adapted to formulate the clustering problem in heterogeneous information networks. Further, we develop two stochastic gradient descent algorithms, namely, SGDClus and SOSClus, which lead to effective clustering multityped objects simultaneously. The experimental results on both synthetic datasets and realworld dataset have demonstrated that our proposed clustering framework can model heterogeneous information networks efficiently and outperform stateoftheart clustering methods.
1. Introduction
Heterogeneous information networks are ubiquitous in the realworld applications. Generally, heterogeneous information networks consist of multiple types of objects and rich semantic relations among different object types. The bibliographic network extracted from the DBLP database (http://www.informatik.unitrier.de/~ley/db/) is a typical heterogeneous information network, as shown in Figure 1. The DBLP database is an open resource that contains most bibliographic information on computer science. The bibliographic network contains four types of objects: author (A), paper (P), venue (i.e., conference or journal) (V), and term (T). The edges are labeled by “write” or “written by” between author and paper or labeled by “publish” or “published by” between paper and venue or labeled by “contain” or “contained in” between paper and term.
Clustering analysis is a basic and essential method for mining such networks, which can help us better understand the semantic information and interpretable structure in the network. Unfortunately, most existing approaches, such as spectral clustering, are designed to analyze homogeneous information networks [1] that consist of only a single type of objects and links, while the realworld situations are often heterogeneous information networks [2] in nature with more than one type of objects and links. The mission of clustering such a heterogeneous information network is more difficult than that in a homogeneous information network, as we cannot directly measure the similarity among the different types of objects and relations.
Though some recent studies have focused on clustering heterogeneous information networks, such as RankClus [1] and NetClus [2], they can only be applied to some specific simple network schemas. RankClus can only be used to model bityped networks, where only two different types of objects exist in the network. NetClus was developed for the star network schema, where the links only appear between target objects and attribute objects. The network schema is a metatemplate of a heterogeneous information network, which shows how many types of objects and links are there in the network [3]. Figure 1 shows a typical star network schema, where the paper (P) is the target object and others are attribute objects.
A tensor is a generalization of the matrix in the highdimensional space. It is a natural expression of complicated and interpretable structures in highmode data. In this paper, we model a heterogeneous information network as a tensor without the restriction of network schema. Each type of objects maps onto one mode of the tensor, and the semantic relations among different object types map onto the elements in the tensor. Then, a tensor CP decomposition method is adapted to formulate the clustering problem in heterogeneous information networks. And two stochastic gradient descent algorithms are developed, which lead to effective clustering multityped objects simultaneously. The experimental results on both synthetic datasets and realworld dataset show that the proposed clustering framework can model the heterogeneous information networks efficiently and outperform the stateoftheart clustering methods.
The rest of this paper is organized as follows. In Section 2, we discuss the related work on clustering for heterogeneous information networks and the tensor factorization. Section 3 gives some notations and definitions used in this paper. In Section 4, we formulate the clustering problem and describe two stochastic gradient descent algorithms. The experimental results on both synthetic datasets and realworld dataset are presented in Section 5. Finally, the conclusions are drawn in Section 6.
2. Related Work
Our work mainly focuses on the clustering heterogeneous information networks and the tensor factorization.
2.1. Clustering Heterogeneous Information Networks
Clustering is an unsupervised learning method to recognize the distribution and hidden structures in the data, which is a basic and significant mission for pattern recognition and machine learning. Since MacQueen first proposed Kmeans [4] in 1967, many subtle algorithms have been developed for clustering in the past decades. However, most existing algorithms, such as hierarchical clustering algorithm [5], densitybased clustering [6], meshbased clustering [7], fuzzy clustering algorithm [8], and spectral clustering [9], are designed to analyze point sets or homogeneous information networks, which are composed of only one type of objects and links.
In realworld applications, the datasets are often organized as heterogeneous information networks, where objects and the relations between them are of more than one type. In recent years, researchers have made a significant progress on clustering for heterogeneous information networks [10, 11], which largely focused on four main directions: the first is to use a ranking based clustering algorithm [1]; it developed the RankClus algorithm that integrated clustering with ranking for clustering bityped networks. Its extension, NetClus [2], was developed for the star network schema. They have proven that ranking and clustering can mutually enhance each other. GPNRankClus [12] assumed that edges in heterogeneous information networks follow a Poisson distribution. This method can simultaneously achieve both clustering and ranking in a heterogeneous information network. In addition, FctClus [13] achieved a higher computational speed and had a greater clustering accuracy when applied to heterogeneous information networks. But, same as NetClus, FctClus algorithm can only handle the star network schema. For a general network schema, HeProjI [14] projected the network into a number of bityped or star schema subnetworks and performed the ranking based clustering in each subnetwork.
The second direction involves metapath based clustering algorithms [15, 16]. A metapath is a connected path defined on the network schema of a heterogeneous information network, which represents a composite semantic relation between two objects. PathSim (metapath based topk similarity search) [3] measured the similarity between the same types of objects based on metapath in heterogeneous information networks. However, it has a limitation: the metapath must be symmetric; that is, PathSim could not work on different types of objects. The PathSelClus algorithm in [15–17] integrated metapath selection with user guidance to cluster objects in networks, where user provided seeds for each cluster acted as guidance.
The third direction is structuralbased clustering. Sun et al. proposed a probabilistic clustering method [18] to deal with heterogeneous information networks with incomplete attributes, which integrated the incomplete attribute information and the network structure information. NetSim [19] is a structuralbased similarity measurement between objects for xstar network. Xu proposed a Bayesian probabilistic model based on network structural information for clustering heterogeneous information networks.
The final direction is a clustering algorithm based on social network features. Zhou and Liu designed the SICluster algorithm [20], which adopted the heat diffusion procedure to model the social influence and then measure the similarity between objects.
2.2. Tensor Factorization
A tensor is a multidimensional array, in which the elements are addressed by more than two indices. Tensor factorization has been studied since the early 20th century [21–25]. Two of the most popular tensor factorization methods are Tucker decomposition [21, 24] and canonical decomposition using parallel factors (CANDECOMP/PARAFAC) [23, 24]. The CANDECOMP/PARAFAC is also named CP decomposition. It is worth noting that CP decomposition is a special case of Tucker decomposition.
The clustering issues based on tensor factorization are often modeled as the optimization problems [26]. But it has been proven in [27] that tensor clustering formulations are NPhard. In the past years, many approximation algorithms for tensor clustering are proposed [28–30]. These theories provide a new perspective for us as to clustering heterogeneous information networks. Tensor factorization based clustering has also been used in some specific applications. Examples include link prediction in higherorder network structures [31, 32], collaborative filtering in recommendation systems [33], community detection in multigraphs [34], graph clustering [35], and modeling multisource datasets [36].
The Alternating Least Squares (ALS) algorithm [22, 37, 38] is one of the most famous and commonly used algorithms to solve the tensor factorization, which updates one component iteratively at each round, while holding the others constant. However, ALS suffers from some limitations; for example, ALS may converge to a local minimum and the memory consumption may explode when the scale of tensor is large. Nonlinear optimization approach is another option to obtain the tensor factorization, such as nonlinear conjugate gradient method [39], Newton based optimization [40], randomized block sampling method [41], and stochastic gradient descent [42]. In this paper, we adopt a stochastic gradient descent algorithm with Tikhonov regularization item loss function to process the tensor CP decomposition based clustering.
3. Preliminaries
First, we introduce some related concepts and notations of tensors used in this paper. More details about tensor algebra can be found in [24, 43]. The order of a tensor is the number of dimensions, also known as ways or modes. We will follow the convention used in [23] to denote scalars by lowercase letters, for example, , vectors (one mode) by boldface lowercase letters, for example, , matrices (two modes) by boldface capital letters, for example, , and tensors (three modes or more) by boldface calligraphic letters, for example, . Elements of a matrix or a tensor are denoted by lowercase letters with subscripts; that is, the th element of an thorder tensor is denoted by . The notations about tensor algebra used in this paper are summarized in Notations.
If an thorder tensor can be written as the outer product of vectors, that is, and , tensor is named rankone tensor. The CP decomposition represents a tensor as a sum of rankone tensors; that is, the CP decomposition of is , where is a positive integer and . The rank of a tensor is defined as the smallest number of rankone tensors for which the equality holds in the CP decomposition and denoted as . In fact, the problem of tensor rank determination is NPhard [27].
Let factor matrices , for . We denote . By minimizing the Frobenius norm of the difference between and its CP approximation, the CP decomposition can be formulated as an optimization problem:
Then, we give the definitions for the information network and the network schema, which are based on the work by Sun et al. [3].
Definition 1 (information network [3]). An information network is a graph defined on a set of objects and a set of links , where belongs to objects types and belongs to links types .
Specifically, when or , the information network is called heterogeneous information network; otherwise, it is called homogeneous information network.
We denote the object set of type as , where is the number of objects in type ; that is, and . The total number of objects in the network is given by .
Definition 2 (network schema [3]). The network schema is a metatemplate for a heterogeneous information network , which is a graph defined over object types and links types , denoted by .
A network schema shows how many types of objects are there in the network and which type the links between different object types belong to. Figure 1 shows the network schema of DBLP, which follows a star network schema.
4. Tensor CP Decomposition Based Clustering Framework
4.1. Tensor Construction
According to Definition 2, we know that the network schema is a metatemplate for the given heterogeneous information network . In other words, is an instance of . So we can find at least one subnetwork of , which follows the schema .
Definition 3 (genenetwork). A genenetwork, denoted by , is the minimum subnetwork of , which follows the schema .
It is easy to see that a genenetwork is one of the smallest instances of in the set of subnetworks of . For example, a genenetwork in DBLP network (in Figure 1, which contains four types of objects, ), denoted by , represents a semantic relation of “an Author writes a Paper published in the Venue and containing the Term .” For simplicity, we can use the subscript of each object in to mark the corresponding genenetwork. In the example, the genenetwork can be marked by .
Let be a thorder tensor of size ; each mode of represents one type of objects in the network . An arbitrary element, , for , is an indicator of whether the corresponding genenetwork exists; that is,Then, the heterogeneous information network can be represented by the form of tensor as .
4.2. Problem Formulation
Using the tensor representation of , we can partition the multityped objects into different clusters by the CP decomposition. We assume that there are clusters in and denote as the cluster indication matrix of the th type of objects. Then, the CP decomposition of isEach row of the factor matrix is the probability vector for each object from th type belonging to the th cluster. In other words, the th cluster is composed of the th rankone tensor in the CP decomposition; that is, .
Figure 2 gives an example of tensor CP decomposition method for clustering heterogeneous information network. The left one is the original network with three types of objects, the middle cube is a 3mode tensor, and the right one is the CP decomposition of the 3mode tensor and also is the partition of the original network. In addition, the three types of objects are the yellow round, the blue square, and the red triangle, respectively. The number within each object is the identifier of the object. Each element (black dot in the middle cube) in the tensor represents a genenetwork in the network (black dashed circle in the left). Each component (black dashed circle in the right) in the CP decomposition shows one cluster of the original network.
The problem in (3) is an NPhard nonconvex optimization problem, which has a continuous manifold of equivalent solutions [39]. In other words, the global minimum is drowned in many local minima, which makes it difficult to be found. In realworld scenarios, the objects in the heterogeneous information networks may belong to more than one cluster; that is, the clusters are overlapping. However, the number of clusters that the vast majority of objects may belong to is much smaller than the total number of clusters. That is, most of the elements in should be zero; that is, should be sparse. To overcome the two challenges, we can introduce a Tikhonov regularization term, proposed by Paatero in [44, 45], in the objective function and replace the objective function by the following loss function:where is a regularization parameter. Letbe the first squared loss function component in and letbe the Tikhonov regularization term in , respectively. Then
The Tikhonov regularization term in the loss function has an encouraging property, which makes the Frobenius norms of all factor matrices in the optimization be equal; that is,Therefore, the local minima of loss function become isolated, and any replacement and scaling of the satisfactory solutions will escape from the optimization. The details of proof can be found in [39]. Meanwhile, the Tikhonov regularization term can ensure the sparsity of the factor matrices by penalizing the number of nonzero elements.
Therefore, the tensor CP decomposition method for clustering heterogeneous information networks can be formalized aswhere ; ; , and is the total number of clusters. In (9), we divide into clusters and obtain the structure of each cluster, which includes the distribution of each object. The first constraint in (9) guarantees that the sum of probabilities for each object belonging to all clusters is 1. The second constraint in (9) represents that each probability should be in the range of . The last constraint in (9) makes sure that there is no empty cluster for any mode.
4.3. The Stochastic Gradient Descent Algorithms
Stochastic gradient descent is a mature and widely used tool for optimizing various models in machine learning, such as artificial neural networks, support vector machines, and logistic regression. In this section, the regularized clustering problem in (9) will be addressed by the stochastic gradient descent algorithms. The details of tensor algebra and properties used in this section can be found in [43].
First, we review the stochastic gradient descent algorithm. To solve an optimization problem, , where is a differentiable object function to be minimized and is a variable, the stochastic gradient descent method to update can be described aswhere is a positive number, named learning rate or step size. The convergence speed of stochastic gradient descent algorithm depends on the choice of learning rate and initial solution.
Though the stochastic gradient descent algorithm may converge to a local minimum at a linear speed, the efficiency of the algorithm near the optimal point is not all roses [46]. To speed up the final optimization phase, an extension method named secondorder stochastic algorithm is designed in [46], which replaces the learning rate by the inverse of secondorder derivative of the object function; that is,
Now, we apply the stochastic gradient descent and the secondorder stochastic algorithm to the clustering problem in (9) and propose two algorithms, named SGDClus (Stochastic Gradient Descent for Clustering) and SOSClus (SecondOrder Stochastic for Clustering), respectively.
4.3.1. SGDClus
In SGDClus, we apply the stochastic gradient descent algorithm to the clustering problem in (9). According to (10), each factor matrix , for , is updated by the rule
Actually, is easy to be obtained according to (6); that is,To compute , can be rewritten by matricization of along the th mode as where . Then, we haveAnother way to compute is given in [39], which has the same result as (15). Following the works in [39], we denoteTherefore, the partial derivative of is given byAnd (12) can be rewritten aswhere is an identity matrix. Note that derived by (18) do not satisfy the first and second constraints in (9). To satisfy these two constraints, we can normalize each row of by
Furthermore, the pseudocode of SGDClus is given in Algorithm 1.
4.3.2. SOSClus
In SOSClus, we apply the secondorder stochastic algorithm to the clustering problem in (9). According to (11), each factor matrix , for , is updated by the rule
According to (17), we can obtain the secondorder partial derivative of with respect to ; that is,
By substituting (17) and (21) into (20), we have
Note that is the General Gradientbased Optimization (OPT) [39] solution for updating in the regularized CP decomposition by making (17) equal to zero. See the details of proof in [39]. LetTherefore, the updating rule of in SOSClus is a weighted average of the current solution and the OPT solution intuitively; that is,
Actually, SOSClus is a general extension of General Gradientbased Optimization (OPT) [39] and the ALS with step size restriction in randomized block sampling method [41]. In (24), when the learning rate , we get the OPT solution. When the regularization parameter , SOSClus becomes the ALS with step size restriction in randomized block sampling method.
Similar to SGDClus, derived by (24) in SOSClus also do not satisfy the first and second constraints in (9). We should normalize each row of according to (19). The pseudocode of SOSClus is given in Algorithm 2.
4.4. Feasibility Analysis
Theorem 4. The CP decomposition of obtains the clustering of multityped objects in the heterogeneous information network simultaneously.
Proof. Since the proofs for different types of objects in the heterogeneous information network are similar, we will simply describe the process for a single type of objects. Without loss of generality, we detail the proof on the th type of objects.
Given a heterogeneous information network and its tensor representation , the nonzero elements in represent the input genenetworks in , which we want to partition into clusters . The centre of the cluster is denoted by . By using the coordinate format [47] as the sparse representation of , the genenetworks can be denoted as a matrix , where is the number of nonzero elements in . Each row gives the subscripts of corresponding nonzero element in . In other words, represents a genenetwork and the entries are the subscripts of the objects contained in the genenetwork.
The traditional clustering approach, such as Kmeans, minimizes the sum of differences between individual genenetwork in each cluster and the corresponding cluster centres; that is,where is the probability of genenetwork belonging to the th cluster. Also we can rewrite the problem by a new perspective of clustering individual object in the genenetwork as follows: where is the probability of object belonging to the th cluster.
In the matrix form, Kmeans can be formalized as where is the cluster indication matrix and is the cluster centres.
By matricization of along the th mode, the CP decomposition of in (3) can be rewritten as is the matricization of along the th mode and is the sparse representation of . Let and let . That is, is the cluster indication matrix for the th type of objects and is the cluster centres. So, the CP decomposition in (3) is equivalent to the Kmeans clustering for the th type of objects in heterogeneous information network .
By matricization of in (3) along different modes, we can prove that the CP decomposition is equivalent to the Kmeans clustering for other types of objects. So, the CP decomposition of obtains the clustering of multityped objects in the heterogeneous information network simultaneously.
It is worth noting that the CP decomposition based clustering is a soft clustering method; the factor matrices indicate the probability of objects belonging to corresponding clusters. The soft clustering is more in line with reality, because many objects in the heterogeneous information networks may belong to several clusters. In other words, the clusters are overlapping. In some cases, the overlapping clusters need to be translated into nonoverlapping clusters, which can be achieved by using different approaches, such as Kmeans, to cluster the rows of factor matrices. Usually, the nonoverlapping clusters can be obtained by simply assigning each object to the cluster which has the largest entry in the corresponding row of factor matrix.
4.5. Time Complexity Analysis
The main time consumption of updating each factor matrix in SGDClus is computing . According to (17), we need to calculate and , respectively. Since and , we have , , and .
Firstly, if we successively calculate the KhatriRao product of matrices and a matrixmatrix multiplication when computing , the intermediate results will be of very large size and the computational cost will be very expensive. In practice, we can reduce the complexity by ignoring the unnecessary calculation. Let us observe the element of ; that is, is an element in the matricization of , which represents a corresponding genenetwork in the heterogeneous information network. When , we can ignore the following KhatriRao product. Hence, only nonzero elements in need to be computed. Therefore, the time complexity for computing is .
Secondly, the element of is given bySo, the time complexity of computing is , where is the total number of objects in the networks.
Above all, the time complexity of each iteration in SGDClus is . Note that, in the realworld heterogeneous information networks, the number of clusters and the number of object types are usually far less than ; that is, and .
According to (22), the time complexity of updating each factor matrix in SOSClus is composed of three components: , , and the product of them. Compared to SGDClus, only the time consumption of computing an inverse matrix for is additional in SOSClus. Since is a square matrix, computing the inverse of such matrix costs . Nevertheless, is usually negligible because .
5. Experiments and Results
In this section, we present several experiments on synthetic and realworld datasets for heterogeneous information networks and compare the performance with a number of stateoftheart clustering methods.
5.1. Evaluation Metrics and Experimental Setting
5.1.1. Evaluation Metrics
In the experiments, we adopt the Normalized Mutual Information (NMI) [48] and Accuracy (AC) as our performance measurements.
NMI is used to measure the mutual dependence information between the clustering result and the ground truth. Given objects, clusters, one clustering result, and the ground truth classes for the objects, let , , be the number of objects that labeled in clustering result but labeled in the ground truth. The joint distribution can be defined as , the marginal distribution of rows can be calculated as , and the marginal distribution of columns can be calculated as . Then, the NMI is defined as The NMI ranges from 0 to 1: the larger value of NMI, the better the clustering result.
AC is used to compute the clustering accuracy that measures the percent of the correct clustering result. AC is defined aswhere is the cluster label of the object and the is the ground truth class of the object . And is an indicator function:
Since both of NMI and AC are used to measure the performance of clustering one type of object, the weighted average NMI and AC are also used to measure the performance of STFClus and other stateoftheart methods:
5.1.2. Experimental Setting
In order to compare the performance of our proposed SGDClus and SOSClus with others impartially, all methods share a common stopping condition, that is,or reaches the maximum iterations. and are the values of function at the current, that is, th, iteration, and the previous, that is, th, iteration, respectively. And we set the maximum iterations to be 1000. Throughout the experiments, the regularization parameter in SGDClus and SOSClus is fixed as .
All experiments are implemented in the MATLAB R2015a (version 8.5.0), 64bit. And the MATLAB Tensor Toolbox (version 2.6, http://www.sandia.gov/~tgkolda/TensorToolbox/) is used in our experiments. Since the heterogeneous information networks are often sparse in realworld scenarios, that is, most elements in the tensor are zeros, we use the sparse format of as proposed in [47], which has been supported by MATLAB Tensor Toolbox. The experimental results are the average values obtained by running the algorithms ten times on corresponding datasets.
5.2. Experiments on Synthetic Datasets
5.2.1. The Synthetic Datasets Description
The purpose of using synthetic datasets is to examine whether the proposed tensor CP decomposition clustering framework can work well, since the detailed cluster structures of the synthetic datasets are known. In order to make the synthetic datasets similar to a realistic situation, we assume that the distribution for different types of objects that appear in a genenetwork follows Zipf’s law (see details online: https://en.wikipedia.org/wiki/Zipf’s_law). Zipf’s law is defined by , where is the number of objects, is the object index, and is the parameter characterizing the distribution. Zipf’s law denotes the frequency of the th object appearing in the genenetwork. We set and generate 4 synthetic datasets with different parameters. The details of these synthetic datasets are shown in Table 1.
 
is the number of object types in the heterogeneous information network and also the number of modes in the tensor. is the number of clusters. is the network scale, and . is the density of the tensor, that is, the percentage of nonzero elements in the tensor, and . 
5.2.2. Experimental Results
In the beginning of the experiments, we set a common learning rate, , for SGDClus and SOSClus. We find that SOSClus has a faster convergence speed and better robustness with respect to the learning rate, which is clearly shown in Figure 3. Although SGDClus may eventually converge to a local minimum, the efficiency of the optimization near a local minimum is not all roses. As shown in Figure 3, the solutions of SGDClus swing around a local minimum. This phenomenon proves that the convergence speed of SGDClus is sensitive to the choice of learning rate .
Then, we modify the learning rate to for SGDClus, where is a constant optimized in the experiments. In practice, for SGDClus running on and for SGDClus running on . The performance comparison of SGDClus with learning rate and SOSClus with learning rate on and is shown in Figure 4. By employing the optimized learning rate, SGDClus converges to a local minimum quickly. However, compared to SOSClus, SGDClus still has no advantage. The hand drawing blue circles over the curve of SOSClus in Figure 4 shows that SOSClus can escape from a local minimum and find the global minimum, while SGDClus just obtains the first reaching local minimum.
According to (23) and (24), we accessorily obtain the solutions of OPT by running SOSClus. So, we compare the AC and NMI of OPT, SGDClus, and SOSClus on the 4 synthetic datasets, which are shown in Figure 5. With the increase of object types in the heterogeneous information networks, the AC and NMI of SOSClus and OPT increase distinctly, while performance of SGDClus almost does not change. When , AC and NMI of these three methods on and are almost equal and low. However, AC and NMI of SOSClus increase to 1 when . Since the histograms of OPT, SGDClus, and SOSClus on and are almost the same and on Syn3 and Syn4 are also similar, we know that the parameters and have no significant effect on the performance. Generally, the larger density and the number of object types in the network result in higher AC and NMI of SOSClus.
Obviously, in the experiments on the 4 synthetic datasets, SOSClus shows an excellent performance. SOSClus has a faster convergence speed and better robustness with respect to the learning rate. Meanwhile, SOSClus performs better on AC and NMI, because it can escape from a local minimum and find the global minimum.
5.3. Performance Comparison on RealWorld Dataset
5.3.1. RealWorld Dataset Description
The experiments on the realworld dataset are used to compare the performance of the tensor CP decomposition clustering framework with other stateoftheart methods.
The realworld dataset extracted from the DBLP database is the DBLPfourarea dataset, which can be downloaded from http://web.cs.ucla.edu/~yzsun/data/DBLP_four_area.zip. It is a four researcharea subset of DBLP and is used in [2, 3, 12, 13, 15, 16, 18]. The four research areas in DBLPfourarea dataset are database (DB), data mining (DM), machine learning (ML), and information retrieval (IR), respectively. There are five representative conferences in each area. And all related authors, papers published in these conferences, and terms contained in these papers’ titles are included. The DBLPfourarea dataset contains 14,376 papers with 100 labeled, 14,475 authors with 4,057 labeled, 20 labeled conferences, and 8,920 terms. The density of the DBLPfourarea dataset is , so we construct a 4mode tensor with size of 14,376 14,475 20 8,920 and 334832 nonzero elements. We compare the performance of tensor CP decomposition clustering framework with several other methods on the labeled record in this dataset.
5.3.2. Comparative Methods
(i) NetClus (see [2]). An extended version of RankClus [1], which can deal with the network, follows the star network schema. The time complexity of NetClus for clustering each object type in each iteration is , where is the number of clusters, is the number of edges in the network, and is the total number of objects in the network.
(ii) PathSelClus (see [15, 16]). A clustering method based on the predefined metapath requires a user guide. In PathSelClus, the distance between the same type objects is measured by PathSim [3], and the method starts with the given seeds by user. The time complexity of PathSelClus for clustering each object type in each iteration is , where is the number of metapath instances in the network. And, the time complexity of PathSim used by PathSelClus for clustering each object type is , where is the average degree of objects.
(iii) FctClus (see [13]). It is a recently proposed clustering method for heterogeneous information networks. As with NetClus, the FctClus method can deal with networks following the star network schema. The time complexity of FctClus for clustering each object type in each iteration is .
5.3.3. Experimental Results
As the baseline methods can only deal with specific schema heterogeneous information networks, here we must construct different subnetworks for them. For NetClus and FctClus, the network is organized as a star network schema like in [2, 13], where the paper (P) is the centre type, and author (A), conference (C), and term (T) are the attribute types. For PathSelClus, we select the metapath of PTP, APCPA, and CPTPC to cluster the papers, authors, and conferences, respectively. And in PathSelClus, we give each cluster one seed to start.
We model the DBLPfourarea dataset as a 4mode tensor, where each mode represents one object type. The 4 modes are author (A), paper (P), conference (C), and term (T), respectively. Actually, the sequence of the object types is insignificant. And each element of the tensor represents a genenetwork in the heterogeneous information network. In the experiments, we set the learning rate for SOSClus to be and an optimized learning rate with for SGDClus. By running SOSClus, we accessorily obtain the solutions of OPT. So, we compare the experimental results of OPT, SGDClus, and SOSClus on the DBLPfourarea dataset with the three baseline methods. See the details in Tables 2, 3, and 4.



In Tables 2 and 3, SOSClus performs best on average AC and NMI, and SGDClus takes the second place. All methods achieve satisfactory AC and NMI on the conference, since there are only 20 conferences in the network. SGDClus takes the shortest running time, and OPT and SOSClus have an obvious advantage on running time compared with other baselines. The time complexity of SGDClus is , and the time complexity of SOSClus is , where is the number of nonzero elements in , that is, the number of genenetworks in the heterogeneous information network. We have , , and . Compared with the time complexity of the three baselines, SGDClus and SOSClus have a little disadvantage. However, it is worth noting that the three baselines can only cluster one type of objects in the network in each running, while OPT, SGDClus, and SOSClus can obtain the clusters of all types of objects simultaneously by running once. This is the reason why only the total time is shown for OPT, SGDClus, and SOSClus in Table 4. Moreover, the total time of OPT, SGDClus, and SOSClus is on the same order of magnitude as the running time of other baselines for clustering each type of objects, which is consistent with the comparison of the time complexity.
6. Conclusion
In this paper, a tensor CP decomposition method for clustering heterogeneous information networks is presented. In tensor CP decomposition clustering framework, each type of objects in heterogeneous information network is modeled as one mode of tensor, and the genenetworks in the network are modeled as the elements in tensor. In other words, tensor CP decomposition clustering framework can model different types of objects and semantic relations in the heterogeneous information network without the restriction of network schema. In addition, two stochastic gradient descent algorithms, named SGDClus and SOSClus, are designed. SGDClus and SOSClus can cluster all types of objects and the genenetworks simultaneously by running once. The proposed algorithms outperformed other stateoftheart clustering methods in terms of AC, NMI, and running time.
Notations
:  A scalar (lowercase letter) 
:  A vector (boldface lowercase letter) 
:  A matrix (boldface capital letter) 
:  A tensor (calligraphic letter) 
:  The th element of an thorder tensor 
:  Matricization of along the th mode 
:  Hadamard product (elementwise product) of two tensors (or matrices or vectors) with the same dimension 
:  KhatriRao product of two matrices 
:  Frobenius norm of a tensor (or matrices or vectors) 
:  Outer product of two vectors. 
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This study was supported by the National Natural Science Foundation of China (no. 61401482 and no. 61401483).
References
 Y. Sunt, J. Hant, P. Zhao, Z. Yin, H. Cheng, and T. Wu, “RankClus: integrating clustering with ranking for heterogeneous information network analysis,” in Proceedings of the 12th International Conference on Extending Database Technology: Advances in Database Technology (EDBT '09), pp. 565–576, Saint Petersburg, Russia, March 2009. View at: Publisher Site  Google Scholar
 Y. Sun, Y. Yu, and J. Han, “Rankingbased clustering of heterogeneous information networks with star network schema,” in Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '09), pp. 797–805, Paris, France, July 2009. View at: Publisher Site  Google Scholar
 Y. Sun, J. Han, X. Yan, P. S. Yu, and T. Wu, “PathSim: meta pathbased topK similarity search in heterogeneous information networks,” PVLDB, vol. 4, no. 11, pp. 992–1003, 2011. View at: Google Scholar
 J. MacQueen, “Some methods for classification and analysis of multivariate observations,” in Proceedings of the 5th Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Statistics, pp. 281–297, University of California Press, Berkeley, Calif, USA, 1967. View at: Google Scholar
 S. Guha, R. Rastogi, and K. Shim, “Cure: an efficient clustering algorithm for large databases,” Information Systems, vol. 26, no. 1, pp. 35–58, 2001. View at: Publisher Site  Google Scholar
 M. Ester, H. Kriegel, J. Sander, and X. Xu, “A densitybased algorithm for discovering clusters in large spatial databases with noise,” in Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD96), E. Simoudis, J. Han, and U. M. Fayyad, Eds., pp. 226–231, AAAI Press, Portland, Ore, USA, 1996. View at: Google Scholar
 W. Wang, J. Yang, and R. R. Muntz, “STING: a statistical information grid approach to spatial data mining,” in Proceedings of the 23rd International Conference on Very Large Data Bases (VLDB '97), M. Jarke, M. J. Carey, K. R. Dittrich, F. H. Lochovsky, P. Loucopoulos, and M. A. Jeusfeld, Eds., pp. 186–195, Morgan Kaufmann, Athens, Greece, August 1997, http://www.vldb.org/conf/1997/P186.PDF. View at: Google Scholar
 E. H. Ruspini, “New experimental results in fuzzy clustering,” Information Sciences, vol. 6, no. 73, pp. 273–284, 1973. View at: Publisher Site  Google Scholar
 U. von Luxburg, “A tutorial on spectral clustering,” Statistics and Computing, vol. 17, no. 4, pp. 395–416, 2007. View at: Publisher Site  Google Scholar  MathSciNet
 C. Shi, Y. Li, J. Zhang, Y. Sun, and P. S. Yu, “A survey of heterogeneous information network analysis,” IEEE Transactions on Knowledge and Data Engineering, vol. 29, no. 1, pp. 17–37, 2017. View at: Publisher Site  Google Scholar
 J. Han, Y. Sun, X. Yan, and P. S. Yu, “Mining knowledge from data: an information network analysis approach,” in Proceedings of the IEEE 28th International Conference on Data Engineering (ICDE '12), pp. 1214–1217, IEEE, Washington, DC, USA, April 2012. View at: Publisher Site  Google Scholar
 J. Chen, W. Dai, Y. Sun, and J. Dy, “Clustering and ranking in heterogeneous information networks via gammapoisson model,” in Proceedings of the SIAM International Conference on Data Mining (SDM '15), pp. 424–432, May 2015. View at: Google Scholar
 J. Yang, L. Chen, and J. Zhang, “FctClus: a fast clustering algorithm for heterogeneous information networks,” PLoS ONE, vol. 10, no. 6, Article ID e0130086, 2015. View at: Publisher Site  Google Scholar
 C. Shi, R. Wang, Y. Li, P. S. Yu, and B. Wu, “Rankingbased clustering on general heterogeneous information networks by network projection,” in Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management (CIKM ’14), pp. 699–708, ACM, Shanghai, China, November 2014. View at: Publisher Site  Google Scholar
 Y. Sun, B. Norick, J. Han, X. Yan, P. S. Yu, and X. Yu, “Integrating metapath selection with userguided object clustering in heterogeneous information networks,” in Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '12), pp. 1348–1356, ACM, Beijing, China, August 2012. View at: Publisher Site  Google Scholar
 Y. Sun, B. Norick, J. Han, X. Yan, P. S. Yu, and X. Yu, “PathSelClus: integrating metapath selection with userguided Object clustering in heterogeneous information networks,” ACM Transactions on Knowledge Discovery from Data, vol. 7, no. 3, pp. 723–724, 2013. View at: Publisher Site  Google Scholar
 X. Yu, Y. Sun, B. Norick, T. Mao, and J. Han, “User guided entity similarity search using metapath selection in heterogeneous information networks,” in Proceedings of the 21st ACM International Conference on Information and Knowledge Management (CIKM '12), pp. 2025–2029, ACM, November 2012. View at: Publisher Site  Google Scholar
 Y. Sun, C. C. Aggarwal, and J. Han, “Relation strengthaware clustering of heterogeneous information networks with incomplete attributes,” Proceedings of the VLDB Endowment, vol. 5, no. 5, pp. 394–405, 2012. View at: Google Scholar
 M. Zhang, H. Hu, Z. He, and W. Wang, “Topk similarity search in heterogeneous information networks with xstar network schema,” Expert Systems with Applications, vol. 42, no. 2, pp. 699–712, 2015. View at: Publisher Site  Google Scholar
 Y. Zhou and L. Liu, “Social influence based clustering of heterogeneous information networks,” in Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 338–346, ACM, Chicago, Ill, USA, August 2013. View at: Publisher Site  Google Scholar
 L. R. Tucker, “Some mathematical notes on threemode factor analysis,” Psychometrika, vol. 31, pp. 279–311, 1966. View at: Publisher Site  Google Scholar  MathSciNet
 R. A. Harshman, “Foundations of the parafac procedure: model and conditions for an ‘explanatory’ multimode factor analysis,” UCLA Working Papers in Phonetics, 1969. View at: Google Scholar
 H. A. L. Kiers, “Towards a standardized notation and terminology in multiway analysis,” Journal of Chemometrics, vol. 14, no. 3, pp. 105–122, 2000. View at: Publisher Site  Google Scholar
 T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” SIAM Review, vol. 51, no. 3, pp. 455–500, 2009. View at: Publisher Site  Google Scholar  MathSciNet
 A. Cichocki, D. Mandic, L. De Lathauwer et al., “Tensor decompositions for signal processing applications: from twoway to multiway component analysis,” IEEE Signal Processing Magazine, vol. 32, no. 2, pp. 145–163, 2015. View at: Publisher Site  Google Scholar
 W. Peng and T. Li, “Tensor clustering via adaptive subspace iteration,” Intelligent Data Analysis, vol. 15, no. 5, pp. 695–713, 2011. View at: Publisher Site  Google Scholar
 J. Hastad, “Tensor rank is NPcomplete,” Journal of Algorithms. Cognition, Informatics and Logic, vol. 11, no. 4, pp. 644–654, 1990. View at: Publisher Site  Google Scholar  MathSciNet
 S. Metzler and P. Miettinen, “Clustering Boolean tensors,” Data Mining & Knowledge Discovery, vol. 29, no. 5, pp. 1343–1373, 2015. View at: Publisher Site  Google Scholar
 X. Cao, X. Wei, Y. Han, and D. Lin, “Robust face clustering via tensor decomposition,” IEEE Transactions on Cybernetics, vol. 45, no. 11, pp. 2546–2557, 2015. View at: Publisher Site  Google Scholar
 I. Sutskever, R. Salakhutdinov, and J. B. Tenenbaum, “Modelling relational data using Bayesian clustered tensor factorization,” in Proceedings of the 23rd Annual Conference on Neural Information Processing Systems (NIPS '09), pp. 1821–1828, British Columbia, Canada, December 2009. View at: Google Scholar
 B. Ermiş, E. Acar, and A. T. Cemgil, “Link prediction in heterogeneous data via generalized coupled tensor factorization,” Data Mining & Knowledge Discovery, vol. 29, no. 1, pp. 203–236, 2015. View at: Publisher Site  Google Scholar
 A. R. Benson, D. F. Gleiche, and J. Leskovec, “Tensor spectral clustering for partitioning higherorder network structures,” in Proceedings of the SIAM International Conference on Data Mining (SDM '15), pp. 118–126, Vancouver, Canada, May 2015. View at: Google Scholar
 L. Xiong, X. Chen, T.K. Huang, J. G. Schneider, and J. G. Carbonell, “Temporal collaborative filtering with Bayesian probabilistic tensor factorization,” in Proceedings of the 10th SIAM International Conference on Data Mining (SDM '10), pp. 211–222, Columbus, Ohio, USA, May 2010. View at: Google Scholar
 E. E. Papalexakis, L. Akoglu, and D. Ience, “Do more views of a graph help? Community detection and clustering in multigraphs,” in Proceedings of the 16th International Conference of Information Fusion (FUSION '13), pp. 899–905, Istanbul, Turkey, July 2013. View at: Google Scholar
 M. Vandecappelle, M. Bousse, F. V. Eeghem, and L. D. Lathauwer, “Tensor decompositions for graph clustering,” Internal Report 16170, ESATSTADIUS, KU Leuven, Leuven, Belgium, 2016, ftp://ftp.esat.kuleuven.be/pub/SISTA/sistakulak/reports/2016_Tensor_Graph_Clustering.pdf. View at: Google Scholar
 W. Shao, L. He, and P. S. Yu, “Clustering on multisource incomplete data via tensor modeling and factorization,” in Proceedings of the 19th PacificAsia Conference on Knowledge Discovery and Data Mining (PAKDD '15), pp. 485–497, Ho Chi Minh City, Vietnam, 2015. View at: Google Scholar
 J. D. Carroll and J.J. Chang, “Analysis of individual differences in multidimensional scaling via an nway generalization of “EckartYoung” decomposition,” Psychometrika, vol. 35, no. 3, pp. 283–319, 1970. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 L. De Lathauwer, B. De Moor, and J. Vandewalle, “On the best rank1 and rank(R1, R2, . . . , RN) approximation of higherorder tensors,” SIAM Journal on Matrix Analysis and Applications, vol. 21, no. 4, pp. 1324–1342, 2000. View at: Publisher Site  Google Scholar
 E. Acar, D. M. Dunlavy, and T. G. Kolda, “A scalable optimization approach for fitting canonical tensor decompositions,” Journal of Chemometrics, vol. 25, no. 2, pp. 67–86, 2011. View at: Publisher Site  Google Scholar
 S. Hansen, T. Plantenga, and T. G. Kolda, “Newtonbased optimization for KullbackLeibler nonnegative tensor factorizations,” Optimization Methods & Software, vol. 30, no. 5, pp. 1002–1029, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 N. Vervliet and L. De Lathauwer, “A randomized block sampling approach to canonical polyadic decomposition of largescale tensors,” IEEE Journal on Selected Topics in Signal Processing, vol. 10, no. 2, pp. 284–295, 2016. View at: Publisher Site  Google Scholar
 R. Ge, F. Huang, C. Jin, and Y. Yuan, “Escaping from saddle points—online stochastic gradient for tensor decomposition,” http://arxiv.org/abs/1503.02101. View at: Google Scholar
 T. G. Kolda, “Multilinear operators for higherorder decompositions,” Tech. Rep. SAND20062081, Sandia National Laboratories, 2006, http://www.osti.gov/scitech/biblio/923081=0pt. View at: Google Scholar
 P. Paatero, “A weighted nonnegative least squares algorithm for threeway ‘PARAFAC’ factor analysis,” Chemometrics & Intelligent Laboratory Systems, vol. 38, no. 2, pp. 223–242, 1997. View at: Publisher Site  Google Scholar
 P. Paatero, “Construction and analysis of degenerate PARAFAC models,” Journal of Chemometrics, vol. 14, no. 3, pp. 285–299, 2000. View at: Publisher Site  Google Scholar
 B. L. Bottou and N. Murata, “Stochastic approximations and efficient learning,” in The Handbook of Brain Theory and Neural Networks, 2nd edition, 2002. View at: Google Scholar
 B. W. Bader and T. G. Kolda, “Efficient MATLAB computations with sparse and factored tensors,” SIAM Journal on Scientific Computing, vol. 30, no. 1, pp. 205–231, 2007. View at: Publisher Site  Google Scholar  MathSciNet
 A. Strehl and J. Ghosh, “Cluster ensembles—a knowledge reuse framework for combining multiple partitions,” Journal of Machine Learning Research, vol. 3, no. 3, pp. 583–617, 2003. View at: Publisher Site  Google Scholar  MathSciNet
Copyright
Copyright © 2017 Jibing Wu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.