Exploring the Combination of Dempster-Shafer Theory and Neural Network for Predicting Trust and Distrust
In social media, trust and distrust among users are important factors in helping users make decisions, dissect information, and receive recommendations. However, the sparsity and imbalance of social relations bring great difficulties and challenges in predicting trust and distrust. Meanwhile, there are numerous inducing factors to determine trust and distrust relations. The relationship among inducing factors may be dependency, independence, and conflicting. Dempster-Shafer theory and neural network are effective and efficient strategies to deal with these difficulties and challenges. In this paper, we study trust and distrust prediction based on the combination of Dempster-Shafer theory and neural network. We firstly analyze the inducing factors about trust and distrust, namely, homophily, status theory, and emotion tendency. Then, we quantify inducing factors of trust and distrust, take these features as evidences, and construct evidence prototype as input nodes of multilayer neural network. Finally, we propose a framework of predicting trust and distrust which uses multilayer neural network to model the implementing process of Dempster-Shafer theory in different hidden layers, aiming to overcome the disadvantage of Dempster-Shafer theory without optimization method. Experimental results on a real-world dataset demonstrate the effectiveness of the proposed framework.
With the pervasiveness of social media, more and more users participate in various online activities in an unprecedented rate. Trust plays an important role in helping online users make decisions, dissect relevant and reliable information, and receive recommendations. Distrust represents strong negative feelings and insecurity about a user’s motivation, intention, and behavior. Fortunately, online social media have made it easy for users to indicate whom they trust and whom they do not. However, trust and distrust relations are often too sparse to be accurately predicted . In addition, the finding in  shows that the number of negative links is much smaller compared to positive links. The imbalance and sparsity of social relations bring great difficulties and challenges in predicting trust and distrust.
In essence, constructing trust and distrust relations is a complex cognitive psychology process, which is involved with many theories, such as sociological theory, cognitive theory, and psychological theory. Meanwhile, there are numerous inducing factors to determine trust and distrust relations. However, classical classifiers do not fully take into account the relationship among inducting factors and result in the performance degrading dramatically. The relationship among inducing factors may be dependency, independence, and conflict. Dempster-Shafer theory  is an effective and efficient strategy for reasoning with uncertainty, with understood connections to other frameworks such as probability, possibility, and imprecise probability theories. The theory allows combining evidences from different sources and arrives at a degree of belief that takes into account all the available evidences from multiple information sources. While neural network is generally presented as systems of interconnected “neurons” which exchange messages between each other, the connections have numeric weights that can be tuned based on experience, making neural nets adaptive to inputs and capable of learning. Therefore, exploring the combination of Dempster-Shafer theory and neural network can potentially improve the performance and bring new opportunities for trust and distrust prediction.
In this paper, we study predicting trust and distrust based on Dempster-Shafer theory and neural network. In essence, we investigate how to model Dempster-Shafer theory and neural network mathematically and how to incorporate them into predicting trust and distrust. Our contributions are summarized as follows:(i)Analyze and quantify inducing factors about trust and distrust, namely, homophily, status theory, and emotion tendency, and verify the research motivations that these inducing factors are contributed to predict trust and distrust.(ii)Construct evidence prototype based on these inducing factors as input nodes of multilayer neural network, whose advantage is to improve the reliability of evidence features and simplify the complexity of establishing Basic Belief Assignment (BBA) in Dempster-Shafer theory.(iii)Investigate multilayer neural network to simulate the combining evidences process from different inducing factors in Dempster-Shafer theory, which not only achieves complex decision-making process, but also overcomes the disadvantage of Dempster-Shafer theory without optimization method.
The rest of the paper is organized as follows: Section 2 introduces the related works. Section 3 depicts the inducing factors about trust and distrust. Section 4 explores the novel approach for trust and distrust prediction. Section 5 reports experimental results and performance evaluation. Finally, we summarize our conclusion and suggest possible future directions in Section 6.
2. Related Works
2.1. Trust Prediction
Trust prediction, which explores unknown relations between online users, is an important research topic in social network analysis. Existing approaches can be grouped into two categories: unsupervised methods and supervised methods.
Most unsupervised methods are based on trust propagation. Trust propagation models focus on developing a trust prediction model which propagates trust values through web of trust. Guha et al.  propose a framework of trust propagation schemes that several atomic propagations are introduced, such as direct propagation, cogitation propagation, transpose propagation, and trust coupling propagation, to reduce the sparsity of a web of trust. Kim and Song  propose a trust prediction model by discovering reliable trust paths between two users based on reinforcement learning. Oh et al.  propose a probability-based trust prediction model based on trust-message passing, which takes advantage of two kinds of information: explicit information and implicit information. However, due to just relying on web of trust, these works based on trust propagation are often too sparse to predict trust with high accuracy.
Supervised methods first extract features from available sources and then construct a classifier based on the labeled trust relations. Zolfaghar and Aghaie  first develop a framework of social trust inducing factors and map these factors to corresponding measurable features and then, respectively, apply C5.0 tree and neural net into predicting trust relations. Ma et al.  extract various features based on writer-reviewer interaction and apply these features into personalized and cluster-based classification methods. Liu et al.  propose a SVM based prediction model to infer the trust relation between two users solely based on their individual actions and interactions in an online community.
Most researches about trust prediction have not paid sufficient attention to the importance of distrust relation . Therefore, links prediction in signed network is becoming an emerging research field.
2.2. Links Prediction in Signed Network
Signed network is a special network that creates positive and negative links between nodes. Link prediction in signed network infers new positive and negative links by giving old positive and negative links. Leskovec et al. [2, 11] first apply status theory into better explaining the differences between positive and negative links in online social networks and also use contrasts between balance theory and status theory to draw inferences about how links are being used in particular social computing applications. Yang et al.  investigate both unsupervised and semisupervised models to infer signed social ties by capturing the interplay between social relations and users’ behavior of decision making and extend the models to encode general principles form social psychology. Hsieh et al.  explore the low-rank matrix completion problem to infer sighed networks based on both balance theory and practical points of view. Agrawal et al.  propose a more efficient matrix factorization approach for sign inference that recovers the missing links by matrix completion algorithms under certain conditions.
2.3. Dempster-Shafer Theory
The Dempster-Shafer theory  is a mathematical theory of evidence, which allows one to combine evidences from different sources and arrive at a belief function by taking into account all the available evidences. As a more flexible mathematical tool, Dempster-Shafer theory not only combines with other mathematical frameworks [15–18], but also combines with classifiers [19–21] for dealing with imprecise and uncertain data. In , the authors propose an approach that addresses the issue of managing imprecise and vague information in evidential reasoning by combining the Dempster-Shafer theory with the fuzzy set theory. Xiao et al.  extend the research of multiple predictions, which use rough set to determine the weight of each single prediction method and utilize Dempster-Shafer theory as the combination method. Basir et al.  propose a systematic framework of DSET based on supervised learning for data fusion. They use Dempster-Shafer theory to address the issues of constructing evidence structures and handling dependence between information sources based on neural networks. Denoeux et al. [20, 21] propose an evidential neural network method for pattern classification problems. The assignment of a pattern to a class is made by the degree of support which is defined as a function of the distance between two vectors.
3. Data Analysis and Motivating Observations
3.1. Data Collection
We take Epinions as experimental dataset which is a famous online social network. Users in Epinions not only write a review by rating items with 1–5 stars, but also rate other reviews using some tags which range from “not helpful” to “most helpful.” In addition to user’s behavioral data, the trust and distrust relations are another advantage that users can express their web of trust and block list. Therefore, we first collect the available dataset for this research and pay more attention to direct or indirect social correlations, such as rating the same items, rating other users’ reviews, and common neighbours. Then, we delete these users with less than two in-degrees and further filter the users with less than two reviews and ratings, aiming to obtain datasets that are large enough and have sufficient information for the purpose of evaluation. Table 1 shows some statistics of the collected dataset.
3.2. Inducing Factors of Trust and Distrust
In view of sociology and psychology, we divide the inducing factors of trust and distrust into different types of features which include homophily, status theory, and emotion tendency. These features refer to network structures and interaction behaviors in Epinions.
Homophily is one of the most important social theories that a contact between similar people occurs at a higher rate than among dissimilar people . We study homophily via the correlation between social relations and users similarity. In this paper, we investigate the following four widely used similarity measures for homophily .
Rating Similarity (RS). Rating similarity is equivalent to user’s behavior similarity, which can be measured by the cosine similarity of two rating vectors and based on distance. We define rating similarity as
Pearson Correlation Coefficient (PCC). Due to different users with different rating styles, namely, some users have the propensity to give higher rating to all items while others probably tend to rate lowly, the similarity of Pearson Correlation Coefficient is defined aswhere is the rating to the th item from and is the set of items rates. represents the average rate value of , and denotes the subset of items rated by both and .
Jaccard’s Coefficient (JC). To calculate preference similarity between and , we model the set of items rated by and as two vectors and , respectively. Then, we utilize Jaccard’s Coefficient  to measure the similarity of user preference:where denotes the size of a set.
Common Neighbours (CN). We exploit user’s in-degree and out-degree to measure similarity feature based on common neighbours:where denotes the set of incoming links from other users to and denotes the set of outgoing links from to other users. denotes the weight coefficient of correlation between incoming links and outgoing links.
3.2.2. Status Theory
Status theory is developed to help us understand the important role of social status in the formation of trust and distrust relations. In the context of social networks with graph-based representation, we exploit some measures, such as PageRank, PolarityRank , and PolarityIndegree, which are suitable for calculating user social status.
PolarityRank is defined aswhere and , respectively, denote the positive PolarityRank and negative PolarityRank of a node ; both are calculated by (6)
PolarityIndegree. In the context of trust and distrust, users with larger positive in-degree not only are interpreted as a form of popularity, but also are a symbol of likely enough reliable information source. On the contrary, negative in-degree is also a major factor that leads to a negative impact. Therefore, we employ sigmoid function to measure PolarityIndegree based on positive and negative in-degree:
Finally, we normalize and as the final strength of user’s social status by
3.2.3. Emotion Tendency
Social interaction with emotion tendency (ET) is an effective and important way to overcome the problems of imbalance and sparsity of social relations. We explore two features, namely, ET of and ET of , to quantify emotion tendency through comparing user helpfulness ratings with average score of trustor’s ratings and trustee’s ratings. The helpfulness rating set is represented as . We, respectively, define positive emotion tendency and negative emotion tendency by measuring higher and lower than the average rating scores and :where denotes a helpfulness rating that gives the th review of and denotes the average rating of or .
4. Framework of Predicting Trust and Distrust
4.1. Problem Statement
The problem we study in this paper is to predict trust and distrust based on the binary classification through combining neural network and Dempster-Shafer theory in a supervised way. In this subsection, we first present the notations and then formally define the problem of classification on trust and distrust and depict the general framework in Figure 1.
We use uppercase letter to denote an input set , lowercase letter to denote an inducing factor of trust and distrust, and uppercase to denote an output set . The final predicted trust and distrust relations are two class labels, denoted, respectively, as class and class . In Dempster-Shafer theory, we firstly define a frame of discernment about trust and distrust, which denotes . Then, each inducting factor in input set is mapped to an evidence prototype, which denotes . Finally, we construct Basic Belief Assignment (BBA, or called mass function) between evidence prototype and target classes , which denotes . In order to overcome the disadvantage of Dempster-Shafer theory without optimization method, we introduce multilayer neural network to dynamically adjust weights in the supervised learning way; all weights are denoted as .
With the notations above, we formally define classification problem of predicting trust and distrust as follows: given a dataset of all messages with input set of inducing factors and corresponding class labels , on the basis of original features of inducing factors, a series of basic units in our framework are constructed, namely, input unit, evidence processing unit, mass combining unit, fusing unit, and decision unit. Evidence processing unit and fusing unit are, respectively, two smaller multilayer neural networks for handling potential dependence and conflict among evidences and fusing multisource evidences. Finally, all units constitute a larger multilayer neural network as classifier. We aim to learn a classifier to automatically assign class labels for unknown social relations (i.e., test data).
4.2. Basic Units of Framework
4.2.1. Input Unit
Input unit (IU) achieves the mapping process from initial input set to evidence prototype, where evidence prototype is a representative feature set of quantitative intervals of all inducing factors, denoted as , where denotes the quantitative interval of the th feature. is determined by prediction accuracy of trust and distrust in Section 5.2. The power set of evidence prototype with inducing factors and quantitative intervals for each inducing factor has evidence prototype.
4.2.2. Evidence Processing Unit
Evidence processing unit (EPU) is a smaller multilayer neural network with the form of for handling potential dependence and conflict among evidences with dynamically adjusting weights in the supervised learning, where denotes the number of input nodes and denotes the number of hidden nodes. Each input node corresponds to a focal element of one evidence source. The output node is the processing result of evidence sources. and are the connection weight set from input node to hidden node and from hidden node to output node, respectively. is the activation function associated with hidden node and output node. At the input-to-hidden and the hidden-to-output layer, the activation function is chosen as logistic sigmoid function, aiming to deal with dependence and conflict between evidence sources through the nonlinear approximation capability of the multilayer neural network. Thus, the activation function of hidden-to-output layer is the processed result of evidence processing unit.
4.2.3. Mass Combining Unit
Mass combining unit (MCU) is composed of sum nodes and normalization nodes, where each input node corresponds to the output result of an evidence processing unit; the output node is the normalized combined mass. Due to the requirement of independent sources, we exploit the operation of sum and normalization to achieve Dempster’s combination rule on the basis of all evidence processing units. The combination rule of multiple evidences is defined as
4.2.4. Fusing Unit
Fusing unit (FU) combines multiple same kind (or level) evidences by using multilayer neural network, which is composed of evidence processing units and a mass combining unit, where is the number of evidences. In practice, fusing unit can iteratively combine multiple evidences in a hierarchical form, where the local or low-level evidences are fused at the previous stages and the global or top-level evidences are achieved at the final stages. Each fusing unit is trained separately using the outputs from its previous stages.
4.2.5. Decision Unit
Decision unit (DU) is used to make the final decision based on the maximum of pignistic probability, where input nodes are the normalized masses of the fusing unit, and output nodes are associated with target class labels. In our framework, DU is applied into the training process of local fusing and the final decision of global fusing.
4.3. Framework Architecture
The framework architecture includes an input layer, a fusing layer, and a decision layer, respectively, which is shown in Figure 2.
Layer 1: Input Layer. We firstly choose the representative features from each type of inducing factor as evidences. Then, we calculate the distance between input set and all evidence prototype and select of the minimum distance instead of the initial input set by using
Finally, we establish the Basic Belief Assignment (BBA, or mass function) summarizing the information provided by the decreasing function of distance and the degree of class membership . The monotonically decreasing function of the distance between the th element in input set and the quantitative interval of the th feature is defined with an exponential form aswhere the decreasing function and . The distance function is defined as , where and are the upper bound and lower bound of quantitative interval of the evidence prototype , respectively.
In addition, we focus on the correlation between evidence prototype and class membership. We regard class membership as discriminant degree for pattern classification and assume that each evidence prototype is assigned a value which represents a degree of class membership about how much feature prototype belongs to a class . The final can be written aswhere the constraint condition of class membership is . If an evidence prototype fully belongs to a class, it is a special case and is denoted as ; for other class , .
Thus, each mass function is regarded as the output node , which corresponds to a focal element of one evidence source. There are input nodes in the input layer, where is the number of evidences and is the number of classes.
Layer 2: Fusing Layer. We utilize some fusing units to combine multiple evidences in a hierarchical form that begins from local and low-level evidences and then fuses global and top-level evidences. The advantage of fusing unit lies in flexible fusing strategies according to different types of evidences. Therefore, we adopt two strategies of fusing evidences: one is based on the different types of inducing factors; another is according to different types of data attributes, which include network-based and rating-based. The difference of two strategies is that all initial evidences are combined in the local or low-level fusing stage, while both are the same in the global fusing. For the first fusing strategy, three two-source FU modules are applied into combining the same type of inducing factors in local fusing stage. For the second fusing strategy, evidences are fed into two two-source FU for fusing the same type of data attributes in local or low-level fusing stage. The evidence types of two different fusing strategies are shown in Table 2. In global or top-level fusing stage, all evidences obtained from local or low-level fusing stage are combined into the global fusing unit FU simultaneously.
The number of units for two different fusing ways is given in Table 3.
Layer 3: Decision Layer. After fusing the local and global evidences, we obtain three mass functions , , and . Each output node in this layer is associated with a class. Each node consisted of two incoming masses from fusing layer. The final decision is made by selecting the class with the maximum probability as
4.4. Parameter Learning
To apply the proposed framework, the parameter learning is required in each FU. The learning process is achieved in a cascading way. In the local or low-level fusing stage, the FU combined with DU is trained by using corresponding training dataset. In the global or top-level fusing stage, each FU is trained based on the estimated masses, which are transferred from the best outputs of FUs in the local or low-level stage.
Each FU in the fusing layer has similar parameters and learning procedure. Let us take any FU as an example. and are the connection weight set from input node to hidden node and from hidden node to output node, respectively. Each hidden node performs the weighted sum of its inputs to form its net activation, , which is defined as a formation of inner product of the inputs with the weights at the hidden node:where denotes the th mass function on the input layer. denotes the weight of hidden node of input-to-hidden layer. and are the vector formation of both. The weight settings are contributed to improve prediction accuracy and fast converge of iterative algorithm. The initial weights of are set according to the feature effectiveness of inducing factors in our experiments. Each hidden node emits an output that is a nonlinear function of its activation :where we choose a logistic sigmoid function as nonlinear activation function.
Each output node also calculates its net activation in the hidden-to-output layer:where denotes the index of node in the output layer and denotes the number of hidden nodes, and .
Similar to the hidden node, each output node is calculated by using the nonlinear activation function:
The learning parameters are estimated and adjusted by using a multilayer neural network with least squared error function. The error function is based on minimizing the sum of squared difference between the actual output and the desired output , which is defined aswhere denotes all the weights in the multilayer neural network. To avoid overfitting, we add two smoothness regularizations on weight and weight ; then we havewhere and are the regularization parameters which trade off the error function cost with the larger weights penalization.
Then, we derive the back propagation algorithm that is based on gradient descent of error function on the weight and is given bywhere is the learning rate and merely determines how much an updating step influences the current weights. The new terms and cause the weights to decay in proportion to its size when the weights are updated in the process of iteration.
Next, we turn to the problem of evaluating and on (21). As the operations in MCU and DU are fixed without change, the weight changing is only limited to EPU with multilayer neural network. Thus, we can apply the chain rule to obtain partial derivatives of the error function with respect to the weights and :
5.1. Experimental Settings and Evaluation Metric
The experimental settings of trust and distrust prediction are described as follows: we firstly filter some user pairs, such as the similarity that is equal to zero in homophily and users without any helpfulness ratings in emotion tendency. Then, we adopt the Guha methodology  and construct a balance dataset with equal numbers of trust and distrust predictions. Finally, let be the set of user pairs with trust and distrust. We randomly divide set into two parts and . % of denoted as is chosen as the training dataset. The remaining % of denoted as is designated for the testing dataset, whose signs of relations are hidden by setting .
We follow the common metric to evaluate the effectiveness of inducing factors and the proposed framework. In detail, we take the signs of relations obtained by learning model as the set of predicted results, denoted as . Then, the prediction accuracy (PA)  is defined as , where denotes the size of a set. We conduct all experiments 5 times to ensure that our results are reliable and average the result on the evaluation metric.
5.2. Feature Effectiveness of Inducing Factors
We rank all user pairs in in a descending order based on feature weights of inducing factors and take the first % of user pairs as trust relations and take the tail % of user pairs as distrust relations. is varied as . We compare the predicted results with the original signs of relations to measure the prediction accuracy. For comparison, we set a random guessing program as a baseline method. All the features of inducing factors help to construct mass functions and initialize the weights of hidden nodes in local fusing state. The evaluated results are illustrated in Figures 3–5, respectively. We draw the following observations.
(a) Trust relations
(b) Distrust relations
(a) Trust relations
(b) Distrust relations
(a) Trust relations
(b) Distrust relations
(1) Homophily. In Figures 3(a) and 3(b), RS obtains the best performance among the four measures of homophily coefficient. We also note that RS, PCC always obtain better performance than CN, JC, and random. It suggests that helpfulness rating is more effective than network topology in predicting social relations. Meanwhile, since those user pairs with distrust relations also have a certain degree of similarity, the performance of trust prediction is superior to distrust prediction for homophily.
(2) Social Status. In Figures 4(a) and 4(b), the effectiveness of predicting trust relations is worse than distrust relations for social status. The main reason is that users with high statuses are more active; the number of their constructing trust and distrust relations is greatly more than that of users with low statuses. We also note that the performance of and is very similar, which is much better than that of .
(3) Emotion Tendency. In Figures 5(a) and 5(b), since users are likely to give positive ratings (the majority of ratings are with 4 or 5 stars), predicting trust relations based on emotion tendency leads to a bit of confusion and has a certain degree of error rate. However, negative emotion tendency is obviously more discriminated than positive one; the effectiveness of predicting distrust relations is relatively clear. In addition, the performance of ET of is better than ET of in terms of trust and distrust prediction.
On the whole, with the increase of , feature effectiveness of inducing factors reduces, however, still much better than that of random guessing. It is consistent with the previous observations that the task of predicting trust and distrust is more obvious at first and tail region; on the contrary, trust and distrust are confused and difficult at middle region. In addition to comparing the effectiveness of different inducing factors, our other objective is to discrete features of inducing factors and to determine the quantitative interval of evidence prototype in our proposed framework. We stratify features of inducing factors into levels according to the acceptable results of prediction accuracy. In our experiments, we choose for all cases and divide feature strength into three levels, such as high, medium, and low, respectively. The discretization result of feature prototype is shown in Figure 6, if the acceptable prediction accuracy is assigned as 80%, which we can get the order of feature effectiveness. For trust relations, the order of feature effectiveness is ET of , ET of , RS, PCC, , , and so on. For distrust relations, ET of , ET of , , and are several more effect features. In local fusing state, the weight settings of hidden nodes in EPU are based on the rank of feature effectiveness.
5.3. Impact in Different Types of Evidences
On the basis of analyzing all features effectiveness, we use the logistic regression method of LIBLINEAR (http://www.csie.ntu.edu.tw/~cjlin/liblinear/) to evaluate the impact in different types of evidences. The evaluated results help to initialize weight settings of hidden nodes in global fusing state. In detail, we compare two fusing ways to evaluate the impact in different types of evidences: one way is based on the different types of inducing factors; another one is based on the different types of data sources, such as network structure and interaction behaviors. In two groups of experiments, we vary the percentage of labeled relations from 10% to 50% for training.
Table 4 lists the classification performance on different inducing factors. We can see that emotion tendency achieves the highest performance, which is 0.845 at the 30% labeled relations and gradually stabilizes. The performance on homophily is even slightly worse than that of other methods. It is consistent with our previous observations that homophily does not have good discriminating ability regarding trust and distrust.
We, respectively, integrate the features of CN and PolarityRank into network structure and the features of RS and emotion tendency into interaction behaviors. The classification performance of two different types of data sources is shown in Table 5. We can see that the classification method achieves better performance when using interaction behaviors compared to network structure data in terms of prediction accuracy. Specifically, the performance with interaction behaviors improves the prediction accuracy by 0.05–0.12 compared to the network structure. In addition, from both Tables 4 and 5, the fusing way based on inducing factors outperforms the fusing way of data sources.
5.4. Performance of the Proposed Framework
In order to demonstrate the effectiveness of the proposed framework, we compare with several baseline methods:(i)Random: a random guessing result yields 50% prediction accuracy, since trust and distrust in dataset are balanced.(ii)SVM: we apply SVM classifier on the features of inducing factors, which is similar to the logistic regression model.(iii)DT: C5.0 decision tree is constructed by exploiting information gain on features of inducing factors.(iv)NN: we regard neural network with multilayer as the multisources data fusion tool, where input nodes are corresponding to feature of inducing factors.(v)DS: previous work  applies Dempster-Shafer theory on Epinions to predict trust and distrust, which only includes rating data, without a web of trust.(vi)CDN: our proposed approach combines Dempster-Shafer theory and neural network to predict trust and distrust.
As the sparsity of social relations and interaction data, we only use 40% of the labeled data for training. Table 6 shows the performance comparison of the different classification methods. Our approach, CDN, gives 0.067–0.113 improvements on prediction accuracy over DT and SVM methods. When all features of inducing factors are integrated into the SVM classifier, there is no significant improvement of performance compared with the previous logistic regression. In addition, since feature ranking technique is applied into the C5.0 decision tree, the performance of DT is better than SVM (+0.046).
Table 7 lists the performance comparison of the different methods for data fusion technique. It is clear that our proposed framework CDN significantly outperforms the other baseline methods. Specifically, CDN improves the prediction accuracy by 0.045–0.079 compared with the NN and DS methods. From both Tables 6 and 7, we can arrive at a conclusion that our approach can clearly improve performance on predicting trust and distrust.
6. Conclusion and Future Work
In this paper, we firstly analyze and quantify inducing factors about trust and distrust and verify the research motivations that these inducing factors are contributed to predict trust and distrust relations. Then, we devise the evidence prototype based on these inducing factors, which improve the reliability of evidence features and simplify the complexity of establishing the Basic Belief Assignment in Dempster-Shafer theory. Finally, we propose a framework to predict trust and distrust based on combining Dempster-Shafer theory and neural network, which investigate a multilayer neural network to model the implementing process of Dempster-Shafer theory of evidence in different hidden layers, aiming to overcome the disadvantage of Dempster-Shafer theory without optimization method. Experimental results on a real-world dataset demonstrate the effectiveness of the proposed framework.
There are several directions needing further investigation. Firstly, we will incorporate our framework to other social media applications, such as tie strength prediction, sign prediction, and recommendation system. Secondly, we plan to explore new models and algorithms on predicting trust and distrust, such as sparse learning and deep learning.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
This work is supported by the National Natural Science Foundation of China under Grant no. 61300148, the Science and Technology Development Program of Jilin Province of China under Grant no. 20130522112JH, Guangxi Key Laboratory of Trusted Software no. kx201533, Scientific and Technological Planning Project of Jilin Province (20140414064GH), Scientific Research Fund of Jilin Provincial Education (20140307), and Jilin Province Development and Reform Commission Projects (2013C040).
T. Dubois, J. Golbeck, and A. Srinivasan, “Predicting trust and distrust in social networks,” in Proceedings of the IEEE 3th Inernational Conference on Social Computing (SocialCom '11), pp. 418–424, Boston, Mass, USA, October 2011.View at: Google Scholar
G. Shafer, “A mathematical theory of evidence,” Technometrics, vol. 20, no. 1, pp. 1–242, 1978.View at: Google Scholar
H.-K. Oh, J.-W. Kim, S.-W. Kim, and K. Lee, “A probability-based trust prediction model using trust-message passing,” in Proceedings of the 22nd International Conference on World Wide Web (WWW '13), pp. 161–162, Rio de Janeiro, Brazil, May 2013.View at: Google Scholar
N. Ma, E.-P. Lim, V.-A. Nguyen, A. Sun, and H. Liu, “Trust relationship prediction using online product review data,” in Proceedings of the 1st ACM International Workshop on Complex Networks Meet Information & Knowledge Management (CNIKM '09), pp. 47–54, ACM, Hong Kong, November 2009.View at: Publisher Site | Google Scholar
S.-H. Yang, A. J. Smola, B. Long, H. Zha, and Y. Chang, “Friend or frenemy?: predicting signed ties in social networks,” in Proceedings of the 35th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '12), pp. 555–564, ACM, Portland, Ore, USA, August 2012.View at: Publisher Site | Google Scholar
P. Agrawal, V. K. Garg, and R. Narayanam, “Link label prediction in signed social networks,” in Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI '13), pp. 2591–2597, Beijing, China, August 2013.View at: Google Scholar
G. Salton and M. J. Mcgill, Introduction to Modern Information Retrieval, Computer Series, McGraw-Hill, New York, NY, USA, 1983.