Analysis and Applications of LocationAware Big Complex Network Data
View this Special IssueResearch Article  Open Access
Weiwei Yuan, Jiali Pang, Donghai Guan, Yuan Tian, Abdullah AlDhelaan, Mohammed AlDhelaan, "Sign Prediction on Unlabeled Social Networks Using Branch and Bound Optimized Transfer Learning", Complexity, vol. 2019, Article ID 4906903, 11 pages, 2019. https://doi.org/10.1155/2019/4906903
Sign Prediction on Unlabeled Social Networks Using Branch and Bound Optimized Transfer Learning
Abstract
Sign prediction problem aims to predict the signs of links for signed networks. Currently it has been widely used in a variety of applications. Due to the insufficiency of labeled data, transfer learning has been adopted to leverage the auxiliary data to improve the prediction of signs in target domain. Existing works suffer from two limitations. First, they cannot work if there is no target label available. Second, their generalization performance is not guaranteed due to that fact that the solution of their objective functions is not global optimal solution. To solve these problems, we propose a novel sign prediction on unlabeled social networks using branch and bound optimized transfer learning (SP_BBTL) sign prediction model. The main idea of SP_BBTL is to use target feature vectors to reconstruct source domain feature vectors based on relationship projection, which is a complicated optimal problem and is solved by proposed optimization based on branch and bound that can obtain global optimal solution. With this design, the target domain label information is not required for classifier. Finally, the experimental results on the large scale social signed networks validate the superiority of the proposed model.
1. Introduction
Sign prediction predicts signs for links of signed networks, in which signed networks are networks whose edges have signs representing the relationship between nodes. The sign of a link is either positive or negative. A link with a positive sign is also called a positive link, which means the two end nodes of this link trust or like each other. A link with a negative sign is also called a negative link, which means the two end nodes of this link distrust or dislike each other. Compared with unsigned networks which only contain values representing the existence of links, signed networks contain more valuable node relationship information. Because of this rich preserved information, signed networks have been widely used in many applications, such as recommender systems [1, 2] and community detection [3, 4].
Most link prediction methods for signed networks are supervised or semisupervised. It is difficult for them to predict unlabeled target networks without any prior target label information. As for those link prediction methods using transfer learning or ensemble learning technologies, there is nonneglectable knowledge loss in knowledge transferring or domain mapping. The main challenge of the link prediction in signed networks is the data insufficiency problem. And this is the motivation to use an auxiliary labeled network to predict signs for unlabeled target network. Signs are manually labeled by experts, which is time consuming and expensive. This leads to the insufficient number of labeled signs in the real applications. Transfer learning, which is able to transfer knowledge from other domains to assist sign prediction, has therefore been used to address this problem [5, 6]. The domain containing the signs for prediction is called the target domain, and the domain whose knowledge is transferred to the target domain is called the source domain.
Though transfer learning based sign prediction methods have good performances on labeled signed networks, they are unable to predict signs on unlabeled signed networks. In [7–9], they map the feature vectors both in source domain and target domain into a high dimensional space to get the common knowledge as the transferable knowledge. These mapping approaches lost knowledge in the calculation course of high dimensional space like reproducing kernel Hilbert space. And these methods also lost knowledge in inverse transformation when calculating the reconstruction errors. Existing methods need a number of labeled signs of the target domain to train the sign classifier. Labeled signs of the target domain are used together with the labeled signs of the source domain to map feature vectors of both domains. These feature vectors are mapped to a common feature vector space, and the mapped feature vectors are called the common knowledge for transferring. The common knowledge along with the labeled signs of both domains is then used to train the sign classifier. However, in the real applications, it is sometimes unable to get any labeled signs of the target domain, which makes it impossible for the exiting methods [10] to map both domains for the common knowledge.
In addition, sign prediction performances of existing works need further improvement since the optimization of existing objective functions always lead to local optimal solutions or illcondition solutions. Transferring knowledge between different domains is a complicated process, so the objective functions of sign prediction are usually nonconvex. Most existing works like [11] use gradient descent algorithms to optimize their nonconvex objective functions to predict signs. However, since the integration length of the existing works is fixed for gradient descent algorithms, it is not always possible for the optimal solution to be selected as the final solution of the objective function. This leads to the local optimal solution of these works’ objective functions. So, the sign prediction performances of these works are not stable for use. Some other existing works [12–14] optimize their objective function with least angle regression methods or iterations. However, when getting the analytic solution for the objective function, the error of mapping transformations is usually large with these optimal methods [15–17]. This leads to the great loss of the common knowledge for knowledge transferring.
To solve the problems of existing works, we proposed a novel sign prediction model using branch and bound optimized transfer learning (SP_BBTL). SP_BBTL is different from existing works [18–20] which rely on the target labels to establish relationship between source domain and target domain; SP_BBTL establishes a direct projection from source feature vectors to target feature vectors to obtain the reconstruction errors as the relationship between source domain and target domain. Direct mapping can preserve more original and specific information and knowledge in source domain and it is enough to train the classifier and predict target labels without any prior information of target labels. Besides, SP_BBTL adapts branch and bound optimization to calculate the global optimal solution. Specifically, the proposed model optimizes the objective functions via branch and bound (BB), which can get the global optimal solution by ensuring the bounds of solutions and BB can be applied in many combination optimization problems.
There are three main advantages in SP_BBTL. First, it does not require any sign labels in the target domain because of feature vectors mapping. Secondly, the BB based model can be used to compute the global optimal solution of a nonconvex mixed optimization problem with feature vectors in social networks. Third, the proposed method performs well in the imbalance networks compared with existing works because SP_BBTL gets the global optimal solution in the course of source feature vectors reconstruction that has preserved more complete and original transferable knowledge in source domain.
The rest of this paper is organized as follows. Section 2 gives a brief review of the related works; Section 3 presents the details of the proposed method; Section 4 demonstrates the experimental results; Section 5 concludes this paper and points out the future works.
2. Related Works
In this paper, we propose a novel sign prediction method via transfer learning technology. Thus, the relative works are mainly separated into two parts: sign prediction and transfer learning.
2.1. Sign Prediction
There are mainly three categories for sign prediction approaches. The first type constructs the nonbayesian model based on a set of vertex attributes. The second type derives the joint probability of each sample based on the knowledge of probabilities. The third type leverages linear algebra methods to calculate the similarities between network nodes based on rankreduced similarity matrices. References [7–9, 21, 22] are supervised which requires a sufficient number of training samples to construct the sign prediction model. All of the existing approaches require some prior knowledge to train classifiers, yet the cost of getting the prior knowledge is expensive in the real applications. Besides, many sign prediction problems face up to class imbalanced problem in reality. Reference [23] is not suitable for the class imbalanced problem, yet [24] utilized adjacency matrix and Laplace matrix to train and test the classifier. But the objective functions in this approach are optimized by iteration computing, which generates considerable error in calculation course.
In addition to model design, another focus of sign prediction is the extraction of useful feature vectors to construct the sign prediction model. There are mainly two types of features: vertex features and edge features. Vertex features consist of neighborhood node based features, path based features, Katz value [25, 26], cluster coefficient scores, etc. Edge features are actually the features of a pair of nodes, which mainly include kernel features conjunction, extended graph formation, and generic SimRank.
2.2. Transfer Learning
In the real social networks, it is very hard or expensive to obtain the label for our target problem which results in the insufficiency of available data. To solve this problem, transfer learning has been adopted in the sign prediction problem, which tries to utilize the knowledge from source domain to predict the signs in the target domain. Currently transfer learning based sign prediction approaches can be divided into three types: transferring knowledge of instances, transferring knowledge of parameters, and transferring knowledge of feature representations [27]. Reference[28] belongs to instances based methods but they cannot work without target labels. The approaches of transferring knowledge of parameters [29] assume that the model parameters of related learning task can always be shared, which is actually hard for the real networks.
To deal with data insufficiency problem, there are some unsupervised transfer learning approaches [12, 30, 31], which do not require any sign label in the target domain. But they require designing the pivot feature to achieve the good performance for the model. Unfortunately, the design of pivot feature is usually very challengeable. Another related approach SO [32] showed good performance on the regular datasets, yet it has a substantial performance degradation for the class imbalanced datasets.
In general, the analysis of related work shows that traditional sign predictions require a sufficient number of sign labels for training. To alleviate this, transfer learning based approaches have been proposed, yet most of these approaches still need some number of sign labels in the target domain. The existing unsupervised sign prediction approaches based on transfer learning do not need any sign labels in the target domain, but they are usually designed to solve a certain sign prediction problem and hard to use as a universal solution. Therefore, a novel transfer learning based approach for sign prediction is required. In this paper, we propose a novel sign prediction model named sign prediction on unlabeled social networks using branch and bound optimized transfer learning (SP_BBTL). The detailed introduction of SP_BBTL is presented in next section.
3. The Proposed Approach
3.1. Problem Definition
A signed network can be represented as a directed graph G = (V,E,Y), where represents the nodes, represents the edges, and is the sign of . , if there is an edge pointing from to , , ; if there is no connection between and , . If , and trusts or likes , , ; if , and distrusts or dislikes , , . An adjacency matrix A is used to describe the connection of G, in which is the connection between and , here .
Sign prediction predicts for of . To predict signs of links, a feature vector F is extracted from A to described E. F is used to train the sign classifier and then predict signs of target links. Link prediction is a learning task that predicts whether a link exists in a labeled or unlabeled network. Sign prediction is a learning task that predicts the signs of links, which also is called labels or weights of links. Labels used in this work consist of positive label and negative label. Predicting links of a network is the same as predicting the labels of links.
Transfer learning based sign prediction transfers the knowledge of the source domain to the target domain to predict the signs of links for the target domain. Let two signed networks and denote the source domain and the target domain of transfer learning based sign prediction, in which , , , , and . For , SP_BBTL predicts () for with the collaborative representation of , , and .
3.2. The Proposed SP_BBTL Model
The main idea of the proposed SP_BBTL model is presented in Figure 1. SP_BBTL first constructs a sign classifier based on the knowledge of S, i.e., and . SP_BBTL discovers the relationship between and . This relationship projects to and generates a new representation , which is used to establish the relationship between and . is then used as the input of the trained sign classifier to get the output: predicted signs .
The key step of SP_BBTL is to achieve domain adaption from to and to establish a mapping from to :where is the projection of into , and H is the mapping function. The mapping in (1) should maximize the similarity while minimize the difference between and .
The detailed architecture of the proposed SP_BBTL model is shown in Figure 2, in which the grey rectangle represents the functional module and the white rectangle represents the data. The inputs of SP_BBTL are the adjacency matrices of T and S. The output is the predicted signs Y of T. Feature vectors extraction module extracts and from the input matrices. A branch bound option based domain reconstruction module is proposed to establish the mapping from to . The reconstructed feature vectors will be used collaboratively with to predict by the sign classification module. The technical details of SP_BBTL are given in Figure 2.
3.2.1. Feature Vectors Extraction
As shown in Figure 2, given the adjacency matrices of source domain and target domain, feature vectors and are firstly extracted to describe links of S and T for link prediction. In this work, five features are extracted for each feature vectors. These features include link positive outdegree, link negative outdegree, link positive indegree, link negative indegree, and link embeddedness [33].
Link positive outdegree denotes the number of positive edges pointing from to other nodes. reflects the likelihood that gives positive sign to a connected link. The higher value has, the more probably . Link negative outdegree is the number of negative edges pointing from to other nodes. reflects the likelihood that gives negative sign to a connected link. The higher value has, the more probably .
Link positive indegree is the number of positive edges pointing to . reflects the likelihood that gets positive sign from a connected link. The higher value has, the more probably . Link negative indegree is the number of negative edges pointing to . reflects the likelihood that gets negative sign from a connected link. The higher value has, the more probably .
Link embeddedness is the number of common neighbors of and . The link embeddedness of each edge (or link) contains the essential characteristic relationship among its neighbor nodes, which reflects the global structural feature of a substructural network in the whole network. also represents the structural balance of : according to the structural balance theory, the higher value has, the more probably a positive relationship between and exists and the more probably .
3.2.2. Branch and Bound Optimized Domain Reconstruction
Domain reconstruction is the key part of SP_BBTL model. It will build up the latent relationship between source feature vectors and target feature vectors collaboratively. Reconstructing domain from to can be represented aswhere is the solution of (2). In essence, is the bridge to enable the collaborative use of knowledge in S and T. is denoted by in Figure 1, and (2) can be also written asA branch and bound based method is proposed for solving (3) with the minimum globalized error to get a global optimal solution. With minimum error, (3) can be rewritten aswhere is the optimum solution of (3). 1norm sums the matrix along the column to select the maximum numerical value. If the sum of each column of is minimized, the reconstruction error is minimized. This ensures the divergence of the transfer learning task to be minimized.
To minimize the error of (4), motivated by the idea of sparse coding, the constrained condition is set to bewhere is the number of nonzero elements in each column of . can control the sparsity of for the reconstruction. ensures the nonzero elements corresponding to the selected samples are neighbors of .
Calculating (2) is solving a problem of mixed optimization. However, existing methods can only calculate the local optimal solution, which cannot get the global optimal solution of (2). Therefore, the branch and bound method is proposed to achieve the optimal solution calculation. Branch and bound is a generalized search algorithm which includes searching and iterating. Specifically speaking, it compares the size relationship between the given error and the difference about upper bound and lower bound of the feature vectors, and then it adjusts the upper bound and lower bound according to the size relationship. This method controls the complexity of solution vectors in different network scales via parameter α. It can calculate out the global optimal solution for mixed optimization problem. [34] The details are shown in Algorithm 1.

3.2.3. Sign Classification
With the extracted global optimal solution , which is in (3), is predicted aswhere C is a sign classifier that is trained by and and . is the output of C to get the predicted sign .
4. Experimental Results
Four datasets extracted from the realworld applications are used in the experiments to verify the performances of the proposed method. These datasets (http://snap.stanford.edu/data/index.html) are Bitcoinotc [35] (denoted as OTC), Bitcoinalpha (denoted as ALP) [35], Epinions (denoted as EPI)[36], and Slashdot (denoted as SLA)[37]. The values of data label for EPI and SLA belong to . The values of data labels for OTC and ALP are mapped into by setting the label to be 1 if the original label is less than 0 and setting the label to be 1 if the original label is greater than 0. The details of the experimental datasets are given in Table 1. Accuracy and F1score [38] are used to measure the sign prediction performances of the proposed method.

Two baseline methods are used in this paper to compare with SP_BBTL. The first method is SourceOnly (SO) model. It predicts signs of links with data merely from source domain [32]. The second method is Nonnegative Matrix Trifactorization (NMTF) [14]. NMTF predicts signs of links by matrix trifactorization using source domain labels and target domain feature vectors. By factorizing the adjacency matrix of the source domain and the target domain, NMTF gets the latent feature vectors of each domain, which are used together with the explicit feature vectors of each domain to transfer knowledge from the source domain to the target domain.
Sign prediction performances of SP_BBTL are firstly measured with various network sizes. In the experiments, the number of samples in the target domain is fixed to 3000, while the number of source domain samples varies from 3500 to 9500. The proportion of positive link to negative link is set to be 7:3 ( 5%). The experimental results are given in Figure 3, in which AAABBB means AAA is the source domain of sign prediction and BBB is the target domain of sign prediction. For example, OTCALP means predicting signs of ALP by using OTC as the source domain.
As shown in Figure 3, when transferring knowledge from each of the source domaintarget domain groups, the tendency of the performance about network size is slightly decreasing because the accuracy is negative related to network size. The proposed SP_BBTL performs better than the baselines because SP_BBTL really can transfer more useful knowledge from source domain to target domain. In addition, the more efficient optimization also contributes to the superior sign prediction performance of SP_BBTL. Compared with our proposed model, the two baseline methods failed to decrease the transfer loss. This means the solution of their objective function is not globally optimal, which leads to their limited link prediction performances.
Sign prediction performances of SP_BBTL are then measured with the various negative link ratios. In the experiments, the number of samples in the target domain is set to be 3000, while the number of samples in the source domain is set to be 4500 (OTCALP and ALPOTC) and 6500 (EPISLA and SLAEPI) respectively. The ratio of negative links varies from 10% to 90%. The experimental results are given in Figure 4. It is shown that the accuracy and F1score of SP_BBTL are superior to baselines with different negative link ratios. The accuracy of SP_BBTL and the baseline method tends to be microWlike distribution on OTCALP and ALPOTC dataset, while the accuracy of SP_BBTL and the baseline method tends to be microVlike distribution on EPISLA and SLAEPI datasets. F1score of SP_BBTL and baseline methods decreases with the increasing of negative link ratios. In addition, SP_BBTL is insensitive to the decreasing of negative link ratios, while the baseline methods decrease significantly with the increasing of negative link ratios, especially on OTCALP and ALPOTC datasets.
The influence of the constraint parameter on sign prediction performances of SP_BBTL is further analyzed. In the experiments held for OTCALP and ALPOTC, the scale of the source domain and the scale of the target domain are 4500 and 3000, respectively. In the experiments held for EPISLA and SLAEPI, the scale of the source domain and the scale of the target domain are 6500 and 3000, respectively. The value of the constraint parameter varies from 0.1 to 2.5, and the negative link ratio is set to be 0.7 in both source domain and target domain. The experimental results are given in Figure 5. Based on the experimental results, the performance of SP_BBTL is relatively stable when α is larger than 0.5. When is close to 0, that means the zero solution for (3) which is meaningless in sign prediction. So, is suggested to be a value around 0.5 and this contributes to the best prediction performances of SP_BBTL.
5. Conclusion and Future Work
In this paper, a novel method named sign prediction on unlabeled social networks using branch and bound optimized transfer learning (SP_BBTL) is proposed to solve a sign prediction problem via feature vectors projections. In SP_BBTL, labeled source feature vectors are mapped into unlabeled target feature vectors and then the relationship between two domains can be established so that the classifier can be trained without any target label. In addition, the proposed optimization based on branch and bound (BB) performs efficient on social networks because the branch and bound optimization method adapted in the proposed model can ensure the global optimal solution of the objective function. Branch and bound can get global optimal solution by highly efficient searching and iteration. It can maximize the transferable knowledge of the source domain, while minimize the transfer loss. Experimental evaluation validates the superior effectiveness and stability of SP_BBTL in real social networks. At last we give the suggested value for parameter in proposed model.
In the future, we will try to improve the proposed method from several aspects. Firstly, we will try to develop a generalized algorithm, which could not only minimize the influence of negative transfer, but also discover transferable knowledge with different categories of source domains, such as the text data and the image data. Secondly, we will improve the model to minimize the number of the source domain instances used for knowledge transfer, only with little cost in link prediction performances. Lastly, we will extend our model from solving binary sign prediction problem to multilabel sign prediction problem.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that there is no conflict of interest regarding the publication of this paper.
Acknowledgments
This research was supported by Nature Science Foundation of China (Grant No. 61672284), Natural Science Foundation of Jiangsu Province (Grant No. BK20171418), China Postdoctoral Science Foundation (Grant No. 2016M591841), Jiangsu Planned Projects for Postdoctoral Research Funds (No. 1601225C). The authors extend their appreciation to the Deanship of Scientific Research at King Saud University for funding this work through research group no. RGPVPP264.
References
 E. Gündoǧan and B. Kaya, “A recommendation method based on link prediction in drugdisease bipartite network,” in Proceedings of the 2nd International Conference on Advanced Information and Communication Technologies, AICT 2017, pp. 125–128, Ukraine, July 2017. View at: Google Scholar
 P. Moradi and S. Ahmadian, “A reliabilitybased recommendation method to improve trustaware recommender systems,” Expert Systems with Applications, vol. 42, no. 21, pp. 7386–7398, 2015. View at: Publisher Site  Google Scholar
 V. Lyzinski, M. Tang, A. Athreya, Y. Park, and C. E. Priebe, “Community detection and classification in hierarchical stochastic blockmodels,” IEEE Transactions on Network Science and Engineering, vol. 4, no. 1, pp. 13–26, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 L. Yang, X. Cao, D. Jin, X. Wang, and D. Meng, “A unified semisupervised community detection framework using latent space graph regularization,” IEEE Transactions on Cybernetics, vol. 45, no. 11, pp. 2585–2598, 2015. View at: Publisher Site  Google Scholar
 C. Yunfang, W. Tongli, and Z. Wei, “Social link prediction based on the users’ information transfer,” in AsiaPacific Web Conference, pp. 64–76, 2016. View at: Publisher Site  Google Scholar
 Y. Chen, T. Wang, and W. Zhang, “Link prediction analysis of internet public opinion transfer from the individual perspective,” New Technology of Library Information Service, 2016. View at: Google Scholar
 A. Daniely, R. Frostig, and Y. Singer, “Toward deeper understanding of neural networks: The power of initialization and a dual view on expressivity,” Advances In Neural Information Processing Systems, pp. 2261–2269, 2016. View at: Google Scholar
 Y. Zhang, S. Wang, P. Phillips, J. Yang, and T.F. Yuan, “Threedimensional eigenbrain for the detection of subjects and brain regions related with alzheimer's disease,” Journal of Alzheimer's Disease, vol. 50, no. 4, pp. 1163–1179, 2016. View at: Publisher Site  Google Scholar
 R. Marcotte, A. Sayad, K. R. Brown et al., “Functional genomic landscape of human breast cancer drivers, vulnerabilities, and resistance,” Cell, vol. 164, no. 12, pp. 293–309, 2016. View at: Publisher Site  Google Scholar
 Z. Shi, P. Siva, and T. Xiang, “Transfer learning by ranking for weakly supervised object annotation,” 2017, https://arxiv.org/abs/1705.00873. View at: Google Scholar
 O. Sener and V. Koltun, “Multitask learning as multiobjective optimization,” in in Neural Information Processing Systems, pp. 524–535, 2018. View at: Google Scholar
 V. M. Patel, R. Gopalan, R. Li, and R. Chellappa, “Visual Domain Adaptation: A survey of recent advances,” IEEE Signal Processing Magazine, vol. 32, no. 3, pp. 53–69, 2015. View at: Publisher Site  Google Scholar
 M. Kan, J. Wu, S. Shan, and X. Chen, “Domain adaptation for face recognition: Targetize source domain bridged by common subspace,” International Journal of Computer Vision, vol. 109, no. 12, pp. 94–109, 2014. View at: Publisher Site  Google Scholar
 J. Ye, H. Cheng, Z. Zhu, and M. Chen, “Predicting positive and negative links in signed social networks by transfer learning,” in Proceedings of the 22nd International Conference on World Wide Web, pp. 1477–1488, ACM, May 2013. View at: Google Scholar
 D. Wollmann and M. T. A. Steiner, “The strategic decisionmaking as a complex adaptive system: a conceptual scientific Model,” Complexity, vol. 2017, Article ID 7954289, 13 pages, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 P. Liu and F. Teng, “Multiple criteria decision making method based on normal intervalvalued intuitionistic fuzzy generalized aggregation operator,” Complexity, vol. 21, no. 5, pp. 277–290, 2016. View at: Publisher Site  Google Scholar
 O. Abedinia, N. Amjady, and A. Ghasemi, “A new metaheuristic algorithm based on shark smell optimization,” Complexity, vol. 21, no. 5, pp. 97–116, 2016. View at: Publisher Site  Google Scholar  MathSciNet
 F. Zhuang, X. Cheng, P. Luo, S. J. Pan, and Q. He, “Supervised representation learning with double encodinglayer autoencoder for transfer learning,” ACM Transactions on Intelligent Systems and Technology (TIST), vol. 9, no. 2, p. 16, 2018. View at: Publisher Site  Google Scholar
 J. Tang, T. Lou, J. Kleinberg, and S. Wu, “Transfer learning to infer social ties across heterogeneous networks,” ACM Transactions on Information Systems (TOIS), vol. 34, no. 2, p. 7, 2016. View at: Google Scholar
 M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Deep transfer learning with joint adaptation networks,” 2016, https://arxiv.org/abs/1605.06636. View at: Google Scholar
 D. Pathak, P. Agrawal, A. A. Efros, and T. Darrell, “uriositydriven exploration by selfsupervised prediction,” in International Conference on Machine Learning (ICML), pp. 488489, 2017. View at: Google Scholar
 Y. Kuznietsov, J. Stückler, and B. Leibe, “Semisupervised deep learning for monocular depth map prediction,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6647–6655, 2017. View at: Google Scholar
 J. Kunegis and A. Lommatzsch, “Learning spectral graph transformations for link prediction,” in Proceedings of the 26th International Conference On Machine Learning, ICML 2009, pp. 561–568, Canada, June 2009. View at: Google Scholar
 R. Pech, D. Hao, L. Pan, H. Cheng, and T. Zhou, “Link prediction via matrix completion,” EPL (Europhysics Letters), vol. 117, no. 3, p. 38002, 2017. View at: Publisher Site  Google Scholar
 L. Katz, “A new status index derived from sociometric analysis,” Psychometrika, vol. 18, no. 1, pp. 39–43, 1953. View at: Publisher Site  Google Scholar
 P. Goyal and E. Ferrara, “Graph embedding techniques, applications, and performance: A survey,” KnowledgeBased Systems, vol. 151, pp. 78–94, 2018. View at: Publisher Site  Google Scholar
 S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 10, pp. 1345–1359, 2010. View at: Publisher Site  Google Scholar
 R. Wang, M. Utiyama, L. Liu, K. Chen, and E. Sumita, “Instance weighting for neural machine translation domain adaptation,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1482–1488, Copenhagen, Denmark, September 2017. View at: Publisher Site  Google Scholar
 S. Kumar, X. Gao, and I. Welch, “Learning under data shift for domain adaptation: a modelbased coclustering transfer learning solution,” in Pacific Rim Knowledge Acquisition Workshop, pp. 43–54, 2016. View at: Google Scholar
 P. Peng, T. Xiang, Y. Wang et al., “Unsupervised crossdataset transfer learning for person reidentification,” in Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, pp. 1306–1315, USA, July 2016. View at: Google Scholar
 B. Tan, Y. Song, E. Zhong, and Q. Yang, “Transitive transfer learning,” in Proceedings of the 21st ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2015, pp. 1155–1164, Australia, August 2015. View at: Google Scholar
 Y. Ganin and V. Lempitsky, “Unsupervised domain adaptation by backpropagation,” 2014, https://arxiv.org/abs/1409.7495. View at: Google Scholar
 J. Leskovec, D. Huttenlocher, and J. Kleinberg, “Predicting positive and negative links in online social networks,” in Proceedings of the 19th International World Wide Web Conference (WWW '10), pp. 641–650, April 2010. View at: Publisher Site  Google Scholar
 D. R. Morrison, S. H. Jacobson, J. Sauppe, and E. C. Sewell, “Branchandbound algorithms: a survey of recent advances in searching, branching, and pruning,” Discrete Optimization, vol. 19, pp. 79–102, 2016. View at: Publisher Site  Google Scholar  MathSciNet
 S. Kumar, F. Spezzano, V. S. Subrahmanian, and C. Faloutsos, “Edge weight prediction in weighted signed networks,” in Proceedings of the 16th IEEE International Conference on Data Mining, ICDM 2016, pp. 221–230, Spain, December 2016. View at: Google Scholar
 M. Richardson, R. Agrawal, and P. Domingos, “Trust management for the semantic web,” in International Semantic Web Conference, pp. 351–368, 2003. View at: Google Scholar
 J. Leskovec, K. J. Lang, A. Dasgupta, and M. W. Mahoney, “Community structure in large networks: natural cluster sizes and the absence of large welldefined clusters,” Internet Mathematics, vol. 6, no. 1, pp. 29–123, 2009. View at: Publisher Site  Google Scholar  MathSciNet
 W. Yuan, C. Li, G. Han, D. Guan, L. Zhou, and K. He, “Negative sign prediction for signed social networks,” Future Generation Computer Systems, 2017. View at: Google Scholar
Copyright
Copyright © 2019 Weiwei Yuan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.