Abstract

Issues about opinion diffusion have been studied for decades. It has so far no empirical approach to model the interflow and formation of crowd’s opinion in elections due to two reasons. First, unlike the spread of information or flu, individuals have their intrinsic attitudes to election candidates in advance. Second, opinions are generally simply assumed as single values in most diffusion models. However, in this case, an opinion should represent preference toward multiple candidates. Previously done models thus may not intuitively interpret such scenario. This work is to design a diffusion model which is capable of managing the aforementioned scenario. To demonstrate the usefulness of our model, we simulate the diffusion on the network built based on a publicly available bibliography dataset. We compare the proposed model with other well-known models such as independent cascade. It turns out that our model consistently outperforms other models. We additionally investigate electoral issues with our model simulator.

1. Introduction

Huge success of viral marketing nowadays clearly shows that acquaintances indeed greatly influence people adopting a new or different opinion. This implicates that people, in a way, attempt to plant their intrinsic ideas, opinions, or preferences in others’ minds through exchanging opinions over and over in different circumstances. One interesting and long-discussed scenario is election. Elections in the modern world are an essential mechanism to aggregate the opinions of the masses and to make joint decisions for a variety of purposes. People share thoughts and even attempt to convince others to adopt their attitudes during the election season.

As social media such as Facebook are widely utilized, it becomes quite convenient for people to manifest themselves. Social media exposure grants people a hitherto wide range to deliver their views. Social media extremely accelerates and facilitates such opinion-exchange interactions among individuals. As opinions interflow, the intrinsic opinions of an irresolute person could eventually be assimilated to those of the determined ones. Then a consensus or a public opinion appears.

From a research aspect, understanding the progress of human negotiation benefits the real world applications. For instance, social scientists would wonder to what extent the opinions’ exchange among friends can affect each other’s viewpoints. Campaign companies would inquire how to promote a candidate with limited budgets. Such questions are not easy to answer via a user study, particularly when the number of participants becomes huge.

Opinion diffusion on social networks has been studied for decades. Unfortunately, many previous models, such as the Independent Cascade Model, Linear Threshold Model, SIR/SIS model, and heat diffusion model, cannot manage the election scenario intuitively due to the following two reasons. First, people have their intrinsic opinions more or less, which is absent in the aforementioned models. In such a manner, people may not serve as neutral relays. People amplify the opinions they stand for and deamplify opinions they stand against. Second, opinions could be multidimensional, for example, a viewpoint for multiple electoral candidates. Most diffusion models adopt single values to represent the opinions for simplicity. A single value clearly cannot directly represent the views about multiple candidates. Our goal is to design a suitable diffusion model which is capable of managing the propagation of viewpoints. Up to date we have not yet seen too many computational approaches with systematic and quantifiable studies on this issue.

Inspired by the real world phenomena, we have realized several preferable properties to manage the information diffusion like opinion polling. The properties are high-dimensional media, input dependence, deterministic convergence, and consensus. A summary of the properties is as follows. First, we prefer the media (which represents preference toward candidates) propagated throughout the process being a unit vector because, democratically, individuals (or nodes) have equal rights in casting votes. Second, the preference distribution should be significantly affected by the initial intrinsic preference as well as the neighbors through social network. Finally, we hope the propagation converges eventually, and a common trend appears after numerous interactions [1]. In this paper, we show that our model is the only one satisfying all properties among the existing models.

The novelty and contributions of this paper can be viewed from several different angles.(1)This work strategically demonstrates a plausible process to answer a set of real world problems.(a)We start by designing a preference negotiation model (with theoretical guarantees) to manage high-dimensional information. We assess the quality of this model by proving its convergence and several other important properties.(b)We conduct an experiment to demonstrate the validity of our model in predicting the change of citation preference among authors through collaboration networks.(c)To build the diffusion simulation framework, we further devise suite of satellite algorithms for preference profile sampling, deployment, and seed-voter selection.(2)Our case studies, in practice, have provided the concrete solutions to the real world problems of concern.(a)The simulation shows that election outcomes can be significantly affected by the social factor. We find that each individual preference profile extremely changes after a preference diffusion. Using Kendal coefficient to measure the similarity before and after the diffusion, it usually results in a coefficient below .(b)With the simulation, we additionally examine several well-known voting schemas to verify their vulnerability to vote-buying. Among them, Borda Count voting schema performs best to resist vote-buying. Plurality voting is the most vulnerable to manipulation.

2. Preliminary

We review the previously done works related to the information diffusion and electoral issues.

2.1. Diffusion Model

To develop models for diffusion simulation or prediction, researchers unearth the underlying mechanisms or the inherent patterns of information diffusion from real word phenomena and utilize these findings.

The Linear Threshold Model (LT model for short) [2, 3] and Independent Cascade Model (IC model for short) [3, 4] are the most well-known and fundamental ones to describe how the information propagates step by step in a network. Inspired by the ideas of the two models, various models have been proposed later for more specific scenarios.

The LT model at first intends to describe the process of shutdowns due to chain effect of energy overload in a power grid. The concept is then adopted for simulating the information diffusion. In the LT model, nodes in a network are the containers of energy (information) and the amounts carried are represented as real values. Each node has a predesignated carrying capacity and initially carries no energy. Once the simulation proceeds, some nodes are assigned as the early adopters, the first groups gaining energy (information), to carry energy, and the carried amount increases progressively. Once the amount of carried energy exceeds their capacity, the nodes become active (overload) and pass excessive energy to other linked nodes. This leads to a propagation of power overload.

With an operation similar to the LT model, the IC model further simplifies the carried information as a binary value. Nodes become active once they receive the information passed from the linked neighbors. There is no predetermined capacity for nodes in IC model. Instead, each edge is associated with a real number representing the probability of successful information pass along it. During the diffusion, the active nodes continually attempt to send out the information through edges until all linked neighbors become active.

Kempe et al. (2003) [3] generalized the IC model by introducing a General Cascade Model. Gruhl et al. (2004) [5] and Leskovec et al. (2006) [6] proposed generative model to simulate blog essay generation based on the IC Model. These models assume nodes can turn from inactive to active given a certain probability for cascading. Based on the LT model and the IC model, Saito et al. (2010) [7] proposed the Asynchronous Linear Threshold Model and Asynchronous Independent Cascade Model.

Another influential line of research, following the success of the PageRank algorithm, puts the propagation process in an explicit recursive mathematical form. Heat diffusion [8, 9] is a physics phenomenon describing heat flows from high temperature positions to low temperature positions. Inspired by the heat diffusion, Ma et al. (2008) [8] proposed a model to analyze candidate selection strategies for market promotion. The process is formulated as where is the heat of node at time and is the thermal conductivity, namely, the heat diffusion coefficient.

In heat diffusion process, each vertex receives heat from its neighbors, which is similar to our model. The major difference, which will be discussed in detail in the following section, is that heat diffusion model lacks a normalization phase (since it considers only the propagation of one value) and a fusion phase (because the heat itself can disappear after diffusion, so there is no need to fuse on heat diffusion model).

Inspired by these previous works, our model takes the strong points of these approaches, namely, their focus on mimicking social interaction traits such as forming consensus as well as their incorporation of structural information into the propagation process, and blends them into a more coherent framework that could be used to answer real world problems mentioned in the introduction.

2.2. Electoral Issues

In 1992, Bartholdi III et al. [10] first studied the complexity of the process to determine needed actions by organizer to add or remove candidates to manipulate election results (where it is recognized as the classical social choice theory). However, they did not propose any model for the interactions between voters. Gibbard [11] and Satterthwaite [12] showed that every election scheme with at least three possible outcomes is subject to individual manipulation. This means the minority has a chance to manipulate the group decision to secure a preferred outcome. Gibbard and Satterthwaite also addressed the computational difficulty in minority manipulation. However, their model assumes the independence of voters, which does not consider nor discuss the effect from other voters on voter’s preference. Existing studies in this direction mainly focus on the complexity and feasibility issues, which is very different from our goal.

Liu (2009) [13] attempted to check whether the preference distribution changes if the number of political experts in a communication network increases. They use an agent-based model for simulation. Each agent in the model maintains a binary value toward a candidate (instead of a real value or ranking) and simply disseminates the values to other agents in the nearby 3 by 3 matrix.

Yoo et al. (2009) [14] proposed semisupervised importance propagation model. Their idea is, to some extent, similar to our “fusion phase” by adding the original score into the accumulated score obtained from the neighbor. The difference between their model and ours is that theirs deals with a single value instead of a vector, and therefore they do not perform the normalization over candidate scores like we do.

The election manipulation is a long-discussed issue. Nevertheless, the social factor is absent in these works. Here we bring a marriage between the social network analysis and the electoral issues.

3. The Proposed Model

We here propose the diffusion model to unearth how the communications affect the individual decisions. Abbreviations section lists the notations used in the rest of the paper.

3.1. Preference Propagation Model

We first define a preference profile of an individual , which is a -dimensional vector that represents ’s preference toward different candidates. The th element in is an integer in indicating this individual’s preference for candidate (smaller numbers denote higher ranks). To facilitate the operation of the preference profiles, we translate into a score vector , for all , using the following equation: where . This transformation can be regarded as a normalization process as in not only does the preferred candidate receive higher score but also the sum of all elements equals 1. Using the score vector of each individual, we can create an by matrix denoted by the preference matrix. We denote the preference matrix of a given time stamp since the propagation process starts as .

The information propagates one iteration after the other in our model, and each iteration consists of three phases: propagation, normalization, and fusion.

In the propagation phase, each node synchronically propagates the preference score vector to the neighboring nodes. To describe such operation mathematically, we define an forward transition matrix such that the multiplication of and represents the score of each node obtained from all neighbors after this phase. We denoted it by .

We assume the edge directions in a network reveal the direction of influence. Therefore, , where is a diagonal matrix with the inverse of degree of each node in the diagonal and is the adjacency matrix of . Note that is identical to the forwarding matrix of a random walk algorithm. The only difference is that in a random walk algorithm is multiplied by a vector instead of by a matrix .

In , each row represents the neighbors’ accumulated preference scores toward each candidate. Unlike , the elements in each row of do not add up to one. To ensure that every individual has equal influence while casting votes, we normalize each row of so that its elements add up to one. Therefore, in the second phase, is multiplied by an diagonal normalization matrix , where each element in the diagonal of is equal to the sum of all elements in the corresponding row of . After the second phase, we will obtain a new scoring matrix .

The major difference between our propagation model and the diffusion models for electricity/heat (see Section 2 for more details) lies in the intrinsic difference of the media that are propagated. Electricity or heat flows from one place to another (that is, a flow from node to node implies the material does not exist in anymore). Opinions, by contrast, do not vanish after propagation (that is, ’s inclination towards a candidate does not disappear even after bringing his or her opinions to ). Therefore we add a third phase to include a fusion model that integrates an individual’s own opinions with the opinion gathered from its neighbors.

In the fusion phase, we introduce a parameter for each individual: the susceptible ratio, a real number that represents how easily an individual can be affected by others. Given a susceptibility parameter for each individual, we can then create a susceptible matrix , an diagonal matrix with the value of each individual in the diagonal. If is equal to the identity matrix , which would imply all individuals are equally and highly susceptible to one another, then should be equivalent to its neighbors’ opinion . On the opposite side, if is equal to the zero matrix, implying all individuals are impervious to one another, then should be identical to . Thus, after one iteration of propagation, the preference score matrix can be represented as

Note that we assume that does not change over time, and neither does (which is only dependent upon topology). Interestingly, at first glance one might assume that changes iteratively; it actually does not. Because the sum of each column in equals 1 and the scores are always normalized for all candidates, it is not hard to prove that which depends only on . Therefore, we can rewrite as , where is a time-independent matrix, which becomes an important feature for the proof of convergence in the next section.

The above concludes one iteration of propagation. In the next iteration, becomes the initial preference score for the individuals and the same process can be executed to obtain another round of propagation results . Algorithm 1 is the algorithm for our model.

: iteration number; : initial preference profiles
: susceptible matrix; : forwarding matrix
: normalization matrix
for     to   do
end for
return  

3.2. Proof of Convergence and Consensus

In this section, we show the convergent property of our proposed scheme. The score matrix becomes invariant after a sufficient number of propagations. Moreover, we show that given certain conditions all rows in the converged score matrix are identical. In other words, a consensus within a community will eventually be reached through information propagations in our model.

Let denote the overall preference propagation operation of all three phases explicitly laid down in the previous section, To provide intuition for the forthcoming deductions and to borrow results of the properties of from Section 3.1, we start by pointing out the similarities as well as the differences between and the PageRank matrix . First, the entity acting on is actually a matrix consisting of the vectors of probabilities instead of a simple vector of probabilities. As a result, the columns of do not add up to 1 (only the rows do) and therefore it is not a stochastic matrix. Furthermore, a social personal relationship network is intrinsically more localized compared to the World Wide Web, and, as such, the favorable positive definite property enjoyed by does not necessarily hold for . That said, these complexities, while no doubt complicating the theoretical treatment of our algorithm, are in fact a natural manifestation of the increased richness of our target of research in hand—social networks.

We start our deduction of the convergence of by enlisting the Perron-Frobenius theorem [15] which states that an irreducible, acyclic matrix has a single eigenvalue that is strictly larger than the others. Under the assumption that the graph being induced by , , is strongly connected and that the weights matrix has entries smaller than one but not all zeros, is irreducible and acyclic and thus applies to the Perron-Frobenius theorem. We denote the dominant real positive eigenvalue of by . Armed with this fact, we are able to transform into its Jordan canonical form by which the leading block is a matrix , and other ’s correspond to their strictly smaller eigenvalues . Since, by the rules of matrix multiplication, the effects of on can be analyzed one by one with respect to ’s column vectors without loss of generality, we will proceed with our proof of ’s convergence by concentrating on ’s column vectors which we denote by lower case . Decomposing into the sum of ’s eigenvectors, , we obtain the general form of the time evolution of , where The above shows that converges to zero when is large, and therefore converges to . To get an intuition for the speed of this convergence, we turn to a special case where the susceptible ratios are identical; that is, is a scalar . In this case, we apply the Perron-Frobenius theorem again on and we again obtain ’s Jordan form However, note that since it needs not to be acyclic, it is strictly larger than the other. Now, using this basis we find that is equal to

Since a Jordan canonical form is unique, we obtain . From this result, we confirm that when , degenerates to the trivial diagonal case and that as approaches 1, the rate of convergence is geometrically proportional to .

We are now one step away from the final proof of ’s convergence. Recalling that , once is established, converges. To prove this, we take advantage of the Collatz-Wielandt theorem which gives the following formula for : , where and .

We begin by asserting that the upper bound of is 1. To prove this, we suppose the opposite holds, that means there exists such that . This implies the following list of equations: Note that . Thus, the above list of equations can be arranged into However, by denoting by the subscript that has and remembering that is a nonnegative matrix, one of the above equations would not hold: This justifies the assertion that . Combining this result with the observation that the trivial vector yields , we conclude that . Therefore, , and converges to .

For networks that are not strongly connected we can always find the SCCs in linear time, and the problem reduces to the smaller “source SCCs” of the network since the matrices of all the other SCCs have a Perron root smaller than 1 and their elements eventually vanish. For the remaining source SCCs, since no vertices have susceptibility ratios equal to 1, according to the above results they all converge. The net effect is exemplified by the stark difference between the individuals belonging to the various source SCCs and the rest. Whereas source SCC vertices will converge to their own respective common values, the others may converge to different values and act as followers in terms of aligning their own preferences to the weighted average of those belonging to the sources. Figure 1 gives an example of such phenomenon. Let the initial preference matrix of all the nodes in Figure 1 be where each row in the preference matrix denotes each node’s preference for candidates 1, 2, and 3, respectively. Then after infinite number of propagations, the preference matrix will become in which the preferences of nodes A and B in Figure 1 for candidate 1 converge to the common value , for candidate 2 converge to the common value , and for candidate 3 converge to the common value . However, for nodes and , given that the SCC composed by them is under the influences of both opinion leaders SCC and , their eventual preferences instead of aligning themselves to a common value become a linear combination of the preferences of and . The exact details of this combination depend on the structure of the network.

The preference propagation model simulates this unique behavior of people by projecting the preferences’ vector onto the leading uniform eigenvector denoting equilibrium. In addition, it also attempts to mimic the real world by distinguishing the opinion leaders from the followers. As with its real world counterpart, this process is solely determined by the initial preferences of every individual and the structure of the embedding social network.

Another example is shown in Figure 2, time evolution of preferences held by nodes in a social network demonstrating the effects of opinion leaders, creating their own consensus, and passing it down to opinion followers in a cascading manner. We see that the opinion follower SCC composed by nodes 15 to 20 is colored with various shades of gray depending on its distance to the two opinion leader SCCs composed by nodes 1 to 3 and 4 to 7. We observe that the preference of the opinion leader SCCs 1 to 3 is first passed to the opinion follower SCCs to (in 10th propagation round) and then subsequently passed to the opinion follower SCCs 12 to 14 through the efforts of SCCs to in a cascaded manner.

This simple example demonstrates that the strongly connected source components form the opinion leader groups, while each follower node is affected by (i.e., linear combination) the opinions of its surrounding opinion leader groups. Our framework models the real world observation about how less-convinced personnel are affected by the mass opinions they encountered.

3.3. Comparison with Other Models

We here discuss what the most salient characteristics of a successful social model are based on common observations and beliefs, in an attempt to contrast the most distinguishing features of our model with the other previously proposed frameworks.

High Dimension Media. Since a personal preference describes the order of preference of all possible candidates, the media in an ideal model should be represented as ordered lists instead of as a single value. Most of the propagation models such as Linear Threshold Model, Cascade Independent Model, or DiffusionRank model, unfortunately, only handle binary or real value in propagation.

Topology Dependence and Input Dependence. The word of mouth is the main strategy for a person to affect others. The real world process of guiding friends toward the adoption of self-preference goes mutually and simultaneously. To state such phenomenon, the outgoing persuasions of a person should ideally become a combination of self-preference and the incoming preferences. An ideal model should take into account both network structure and initial personal preference. Moreover, we would like a model’s way of incorporating these two factors to be as natural as possible, instead of relying on ad hoc stopping designs or simply restricting the number of time nodes or individuals’ interactions.

Deterministic Convergence. Of course an ideal model should converge or end eventually, or else it would be difficult for the modeler to interpret the results. As far as we know, there are currently two kinds of designs to achieve such a convergence. The first one, such as LT model and IC model, attaches a binary status to each node in a network to determine whether it is visited. The inactive status means the node is not yet visited while the active status means the node is visited. With such design, preference propagation to inactive nodes can be easily monitored. Moreover, the propagation converges in such model when none of the existing nodes can change the status anymore.

Following the success of the PageRank algorithm, the second popular approach is building the convergence mechanism into a model inherently, so that after sufficient iterations the model converges and produces a definite result.

To make results easily analyzable, convergent models that can generate identical results, given both the same initial preferences of nodes and network structure, are preferred.

Consensus. The problem of reaching a consensus among agents has been studied since around 1970 [16, 17] with simulation models such as the voter model [1]. Mossel et al. gave a theoretical proof that the consensus could be reached with the voter model [18]. Thus an ideal model should be able to reflect specific common traits. In particular, we observe that one such universal trait is people in the same community (i.e., SCC) having the tendency to align their preference after sufficient exchanges. This translates into the fact that an ideal model should contain some kind of homogeneity inside a group.

To see how our model and other proposed frameworks capture the above characteristics of real world social interactions, we conducted several experiments and recorded their results in Table 1 for ease of comparison. We particularly chose models that are most representative in their own stance, namely, the Linear Threshold Model, Independent Cascade Model, PageRank model, and DiffusionRank model, for comparison. Note that since the propagating media in these models are not a vector of preference, we made the following enhancements for each of them to handle such cases. For the LT and IC models, we assume that each vertex initially held approval for its top preferred candidates (nonapproval for the others), and thus, for every candidate, we can obtain a list of seeds as inputs into the LT and IC models. We then execute the model separately on each candidate, gather their results, and normalize them to form the final preference of each vertex. For the PageRank and DiffusionRank models, given that they can take real values as inputs, we simply executed these models separately for each candidate in the preference list and then integrated the results to be a vector of real numbers.

As shown in Table 1, we see that our model is the only model that operates directly on a list of preferences, whereas other models work restrictively on single boolean or real values and have to be executed separately to obtain a joint preference, which fail to consider the correlation of the preference score among candidates. We note that all models provide convergent results. Besides, since the IC model carries a random component, it does not deliver repeatable final preference results.

To examine whether these models can give a kind of consensus to nodes that belong to a strongly connected network, we execute all models on a strongly connected graph until they naturally stop or converge. It turns out that, except for our model, none showed signs of reaching consensus among the final output preferences. Note that our model does not produce consensus given non-SCC components.

To see whether these models take into account the initial preferences held by nodes, we fed all models with six different initial preferences and see whether they give six different results. It is not surprising that the PageRank model returns identical results regardless of the input, indicating that it takes into account only the structure of the network but ignores the initial preferences held by each node or individual. In conclusion, our model is the only framework that supports all five criteria set by observations from real world social networks.

4. Experiment

In this section, we compare our model with well-known diffusion models to evaluate the performance. We examine whether all the aforementioned algorithms, including ours, can capture the preference transition in social networks to a certain extent. To conduct such validation, we require the information such as the network structure and the node preferences over time.

4.1. Preference Data

The citations of scientific research papers implicitly reveal the research interests of the authors. In other words, we believe that the acts such as citing or submitting to the journals or the conferences would be an indicator of the authors’ interests. By utilizing this fact, we can infer the researchers’ preference from their corresponding top frequently cited conferences and journals. We have further realized that one author’s preference could be influenced by the other coauthors. It is particularly correct for advisor-student relationship since the advisors and students usually affect each other’s research interests. We therefore have designed an experiment to model how researchers’ preferences can be affected by the collaborators.

We use KDD Cup 2003 ArXiv HEP-TH (High Energy Physics Theory) citation network [19] with the corresponding paper meta information as our evaluation dataset. This dataset contains the citations from 1992 to 2003. We select the top 16 journals that possess most papers as the candidates to construct the preference lists and construct the yearly preference lists for all authors. A preference list consists of the citation count of the corresponding journals within one year. Thus, for each author, we have 12 lists representing their preferences.

The reason we use the citations rather than the publications of authors is that the publications imply not only preference but also capability. To fairly present the interests, we use the citations. In addition, we construct a collaborative network from this dataset as the underlying social preference diffusion backbone. To easily perceive the changes in interests, we remove the authors who had fewer than 5 publications in the dataset, which results in a network with 2683 nodes.

4.2. Model Comparison

Since we already have all the required information including network structure and preference transition, the next step is to study which diffusion model predicts the preference transition better. We assume a good diffusion model could capture the progression of the authors’ research interests through collaborations. To do so, we initially set up the node preference according to the actual data in year and then compare the predicting results with the actual preference in year . The following issues are noted in the experiment.

High Dimension Media. To represent the order in preference toward all candidates, the media in an ideal model ought to be an ordered list instead of a single value. Nonetheless, most well-known diffusion models, such as LT, IC, and DiffusionRank, only treat the media as boolean or real number. For comparison, we exploit these models in our problem by executing them independently for each candidate. We evaluate the candidate rank based on each independent diffusion result.

Determinism of the Final State. Except for the IC model, outcome of all the models mentioned above is deterministic. Because the parameter (i.e., diffusion probability) in IC model is a nondeterministic factor, we execute the experiment times and average the results.

Initialization. Because the media in LT and IC models are not native for high dimension, we singly process the propagation for each candidate. That means, in our experiment, the active mode of top 1% authors to a specific publisher is initially set active in LT and IC models, while the rest of publishers are set inactive. We further set the diffusion probability of each edge as , where is the degree of its source node in IC model. In LT model, we assign links with identical weight and nodes with the same threshold. The parameters in LT and IC are then tuned to find the optimal outcome. The propagation process is executed multiple times with different thresholds and the performance is averaged. For DiffusionRank model, we use the parameter settings suggested by the authors of [9].

4.3. Experiment Result

Diffusion models are evaluated by comparing their predictions about preference in 1997, 1998, and 1999, while using the real preference during the period from 1993 to 1996 as initial status. To measure the similarity between predicting and real results, we adopt Kendall’s tau coefficient [20] and the Jaccard coefficient. We individually measure the similarity for each author, each node in the network, and then average them as a performance indicator. Because Kendall’s tau coefficient is not well defined with tie scores, we manually set Kendall’s tau score as 0 when there is a tie on all 16 publishers. Furthermore, we calculate the Jaccard coefficient performing on the top 3 highest scored publishers.

Firstly, for the sake of knowing the correspondence between the extent of changes in iterations and in years, we execute one-iteration propagation in each model and then compare the results with the ground truth in 1997, 1998, and 1999, respectively. We also try different susceptible ratio in our model, as implies the authors stick to their own preferences without considering the effect from the neighbors. Table 2 shows the results, which we find quite suitable to take one iteration as a period of a year. The results demonstrate that our model consistently outperforms the 2nd best model DiffusionRank, regardless of which susceptible ratio is used as long as it is not .

Secondly, we execute the diffusion algorithms for multiple rounds and compare them with the ground truth of years 1997–1999. Table 3 shows the average of the scores for 1997, 1998, and 1999. Note that LT and IC models stop when there is no possible activation (regarded as one round), which implies that authors are not affected by their neighbors after the first round is completed. Tables 2 and 3 additionally show that the impervious preferences () reach a performance similar to the best result, which might reveal the slowly changing nature. Nevertheless, the results show that our model can faithfully capture the trait of the social influence even if the authors’ interests change slowly.

5. A Social-Based Simulation Framework for Election Behavior

Based on the preference negotiation model, we implement a simulation framework, shown in Figure 4, granting us to analyze the social impact to elections.

5.1. System Architecture

Voters possess their own preference profiles to each candidate in the early stage of an election. A faithful simulation ought to produce preference profiles that are similar to the practical cases. Thus, we produce profiles satisfying certain distributions according to the data collected from a historical election.

Although we have the voter’s preference data, unfortunately, there is no information telling us the relationship between the voters. To deal with this, we propose several plausible scenarios to deploy the profile on a given social network. Because people generally become friends due to similar tastes and thoughts, the profiles should not be distributed randomly. We design several plausible ways to distribute preference profiles on a social network.

Once the preference profiles are assigned to each node, the system starts executing the preference negotiation model. Nodes in the network start to persuade others and be persuaded. It is not necessary for a simulation to run until convergence, because in the real world there exist some elections which leave insufficient time for people to negotiate, debate, or exchange opinions before having to cast their votes.

Finally, since in each stage there are several parameters we can adjust, we have designed a user-interface that allows the easy execution for experiments (see Figure 3).

5.2. Preference Profile Generation

In the Preference Profile Generation stage, we create preference profiles based on historical election data obtained from OpenSTV, an online voting record database. We choose to use the “Melbourne City Council Victoria Australia 2008—Lord Mayor Leadership team” data set because it is by far the most complete dataset we have found. This dataset was recorded in November 29, 2008, and has the ballot size of 57,961 and 11 candidates. It consists of the preference lists for all voters. A preference list is a sorted list of candidates revealing the preference order of this particular ballot.

We propose a ranking-preserved sampling method to produce the preference profiles based on the historical data, with the aim of preserving the rank of each candidate. Given our historical data, we first learn a matrix , where is equal to the total number of candidates. The th elements of encode the probability that the th position in ranking belongs to candidate according to the historical dataset, . Each column of yields a probability distribution for each candidate of a given rank. Given , we can iteratively sample candidates in each rank (from higher to lower) based on the distribution (with the natural restriction of prohibiting the same candidate in different positions in a single ballot).

5.3. Preference Deployment

As the old saying goes “birds of a feather flock together,” we presume that the people of similar preference profiles have a higher likelihood of being close to each other in the network. Below are three algorithms to realize such idea.

Greedy Deployment. It can be realized by first randomly picking a profile and assigning it to a node in the social network; then we assign the most similar unassigned profiles to their neighbors. Then iteratively for each unassigned node, the algorithm allocates it a profile that is the most similar to its neighbors. To measure the similarity, we exploit the commonly used Kendall coefficient:

Community-Based Deployment. Our underlying idea is to match social network communities with clusters of preference profiles. We first conduct a community detection algorithm to identify communities in a social network (in the experiment, we apply [21]), which groups the community and determines the number of communities automatically based on the maximization of the modularity. Next, we apply a clustering algorithm (in this experiment, we used -means) to group preference profiles into clusters based on Kendall similarity.

In the final step, we assign profiles to each node. The main idea is to assign each profile in the th largest cluster to each node in the th largest community. This, however, is not a straightforward task because the th community and the corresponding cluster are likely to have different sizes. Here we propose a method to adjust the cluster sizes to match those of the communities. To accomplish this, we first sort both the sets of communities and their clusters by their respective sizes. Then from the largest community to the smallest, we compare the size of the th community with that of the th cluster. When the th community is of the same size as the th cluster, we randomly assign the profile in the cluster to a node in the community. When the cluster size is smaller, we add the unassigned points outside this cluster but closest to the cluster center into the cluster until its size matches the size of the community. When the cluster is larger, by contrast, we remove the points from the cluster that are farthest away from the center and join them to the closest unassigned neighbor cluster. Doing this iteratively will gradually assign profiles to nodes and guarantee that nodes in the same community have similar profiles. An example is illustrated in Figure 5, and pseudo code is listed in Algorithm 2.

: social-networks; : preference-matrix
: community-number
: the th largest community
: the th largest cluster
(
for     to     do
if     then
      next
else if     then
      add the unassigned preferences which is outside
but closest to into to
      match
else
      remove the preferences most away from
       to the closest unassigned neighbor
      cluster
end if
 randomly assign the preferences in to nodes in
end for
return the pairs of (node, preference)

Segregation Deployment. The idea is from the setting of Schelling’s segregation model [22]: blacks and whites may not mind, even prefer, each other’s presence, but people will move if they are the minority. At the beginning, we deploy profiles randomly. Then, in each iteration, the nodes which have fewer than 30% neighbors with positive Kendall similarity values will be selected, and their preference profiles will be shuffled. In this experiment, 1000 iterations are performed.

5.4. Preference Negotiation

Once profiles and communities have been assigned, the framework executes the core preference negotiation model to the network. There are two parameters that can have some impact on the process: the iterations of negotiations and the susceptible ratio matrix . controls the negotiation iterations taken before the voters have to cast the votes. In the experiments below, we set .

We set as for all voters by default for our case studies. As suggested in our proof, if is a constant matrix , then the resulting converged preference matrix is indifferent no matter what is ( only controls the speed of convergence).

6. Case Studies

In this section, we answer two questions based on the proposed simulation framework.(1)To what extent does the negotiation process in a social network affect election results?(2)Among the widely known voting schemas, Borda Count, -approval, and plurality, which is the most vulnerable to vote-buying (i.e., easiest to be manipulated)?

In order to construct simulations, we use three collaboration networks (ca-GrQc containing 5,242 nodes and 2,890 edges, ca-HepPh containing 12,008 nodes and 237,010 edges, and ca-HepTh containing 9,877 nodes and 51,971 edges) as the underlying social network dataset. Once the negotiation process ends, the preference scoring vectors will be examined to determine the final rank of the candidates using different voting schemas: Borda Count, -approval, and plurality [23]. We conduct experiments on all three plausible deployments proposed in Section 5.3. Ideally, we hope the simulation on all three deployment methods can produce similar conclusions, which would consequently offer users higher confidence about the results.

The Borda Count determines the final rank of the candidates by giving each candidate a certain number of points corresponding to the position in which it is ranked by each voter. Once all votes have been counted the candidate can be ranked by their total points. For each ballot, the top-ranked candidate will receive points, the second points, and so on.

The ranking in -approval and plurality schemas is determined similar to Borda Count. The only difference among the three voting schemas is the definition of points to be given to each candidate. In the -approval voting schema, the top candidates in each ballot will each receive one point, while the rest will not receive any points. In the plurality voting schema, only the top candidate receives one point.

6.1. Effectiveness of Negotiation

The first question is whether the social-network-based negotiation process can significantly affect the election results. To quantify changes in a voters’ preference profile through negotiations, we compute the average Kendall coefficient between the preference orders before and after each negotiation.

As shown in Figure 6, no matter which deployment method is used, the average Kendall coefficient generally decreases as we enter deeper rounds of negotiation. The slope is steepest in the beginning, revealing the fact that the effect of negotiation reaches the peak in the beginning and gradually declines, which matches the real world experience. Eventually the Kendall value decreases to below 0.5, implying that negotiation through social networks can significantly change the election results.

6.2. Vulnerability to Vote-Buying

This section discusses a key question about elections: if an organization can boost the vote count of a candidate through manipulating certain seed nodes’ preference profiles (pejoratively, we can call this “vote-buying”). We define one successful vote-buying to a voter as “raising the target candidate’s preference score to slightly higher than the score required to obtain a vote from the voter.” For example, if each voter is allowed to cast 3 votes, then the buyers would attempt raising its score to slightly above the 3rd place candidate. To quantify such manipulation cost, we defined it as the difference between the scores before and after manipulating. The scoring vector after vote-buying must be renormalized before further computation can proceed.

Discarding the effect of negotiation, it is intuitive to assume that the best promotion strategy is aiming at the voters whose costs are the smallest. Under such attack, which voting schema is the most vulnerable to vote-buying? This answer can be deduced from Figure 7, where the -axis stands for the budget spent while the -axis stands for the rank of a given candidate for promotion. Note that the higher value of stands for the less favorable, and a candidate has to move to lower position in order to be elected. The results show that, regardless of the deployment, Borda Count schema consistently requires more budget to advance a target candidate, while the plurality schema is the most vulnerable to vote-buying since it generally requires less budget to advance a candidate to higher rank.

7. Conclusion

Analyzing the effect of social networks upon group decisions outcomes is a difficult problem because it is both costly and time consuming to perform user studies to collect people’s private preferences. Indeed, it is the change of preferences through social propagation in particular that we care most about, and to our knowledge this is the first ever study that provides not only theoretical analysis but also the empirical justification of this problem. This study provides an example of how to perform such research with limited data through exploiting algorithm and model design, theoretical justification, and computer simulation.

Our other significant contribution is that we provide an alternative evaluation plan and data to verify a preference propagation model. Acknowledging the lack of real world data to evaluate how the voter’s preference can change through social diffusion, we have come up with a novel idea to identify a publicly available bibliography dataset to evaluate how researchers gradually change their research fields according to the influence of their collaborators. Our evaluation plan opens a new possibility that allows researchers working on preference diffusion problems to be able to evaluate their models without having to identify a highly private voter preference dataset.

Abbreviations

:Individuals
:Candidates
:Number of individuals
:Number of candidates
:Preference profile vector of individual
:Preference score vector for individual
:Preference scoring matrix with size
:Social network layer.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported by the National Science Council, National Taiwan University, and Intel Corporation under Grants NSC101-2911-I-002-001, NSC101-2628-E-002-028-MY2, and NTU102R7501.