- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Computational and Mathematical Methods in Medicine

Volume 2012 (2012), Article ID 303601, 8 pages

http://dx.doi.org/10.1155/2012/303601

## Causal Information Approach to Partial Conditioning in Multivariate Data Sets

^{1}Department of Data Analysis, Faculty of Psychology and Pedagogical Sciences, University of Gent, 9000 Gent, Belgium^{2}Dipartimento Interateneo di Fisica “Michelangelo Merlin”, University of Bari, 70126 Bari, Italy^{3}TIRES-Center of Innovative Technologies for Signal Detection and Processing, University of Bari, 70125 Bari, Italy^{4}INFN, Sezione di Bari, 70125 Bari, Italy

Received 2 November 2011; Revised 15 March 2012; Accepted 18 March 2012

Academic Editor: Dimitris Kugiumtzis

Copyright © 2012 D. Marinazzo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

When evaluating causal influence from one time series to another in a multivariate data set it is necessary to take into account the conditioning effect of the other variables. In the presence of many variables and possibly of a reduced number of samples, full conditioning can lead to computational and numerical problems. In this paper, we address the problem of partial conditioning to a limited subset of variables, in the framework of information theory. The proposed approach is tested on simulated data sets and on an example of intracranial EEG recording from an epileptic subject. We show that, in many instances, conditioning on a small number of variables, chosen as the most informative ones for the driver node, leads to results very close to those obtained with a fully multivariate analysis and even better in the presence of a small number of samples. This is particularly relevant when the pattern of causalities is sparse.

#### 1. Introduction

Determining how the brain is connected is a crucial point in neuroscience. To gain better understanding of which neurophysiological processes are linked to which brain mechanisms, structural connectivity in the brain can be complemented by the investigation of statistical dependencies between distant brain regions (functional connectivity) or of models aimed to elucidate drive-response relationships (effective connectivity). Advances in imaging techniques guarantee an immediate improvement in our knowledge of structural connectivity. A constant computational and modelling effort has to be done in order to optimize and adapt functional and effective connectivity to the qualitative and quantitative changes in data and physiological applications. The paths of information flow throughout the brain can shed light on its functionality in health and pathology. Every time that we record brain activity we can imagine that we are monitoring the activity at the nodes of a network. This activity is dynamical and sometimes chaotic. Dynamical networks [1] model physical and biological behaviour in many applications; also, synchronization in dynamical network is influenced by the topology of the network itself [2]. A great need exists for the development of effective methods of inferring network structure from time series data; a dynamic version of the Bayesian networks has been proposed in [3]. A method for detecting the topology of dynamical networks, based on chaotic synchronization, has been proposed in [4].

Granger causality has become the method of choice to determine whether and how two time series exert causal influences on each other [5, 6]. This approach is based on prediction: if the prediction error of the first time series is reduced by including measurements from the second one in the linear regression model, then the second time series is said to have a causal influence on the first one. This frame has been used in many fields of science, including neural systems [7–10], reochaos [11], and cardiovascular variability [12].

From the beginning [13, 14], it has been known that if two signals are influenced by third one that is not included in the regressions, this leads to spurious causalities, so an extension to the multivariate case is in order. The conditional Granger causality analysis (CGCA) [15] is based on a straightforward expansion of the autoregressive model to a general multivariate case including all measured variables. CGCA has been proposed to correctly estimate coupling in multivariate data sets [16–19]. Sometimes though, a fully multivariate approach can entrain problems that can be purely computational or even conceptual: in the presence of redundant variables the application of the standard analysis leads to underestimation of causalities [20].

Several approaches have been proposed in order to reduce dimensionality in multivariate sets, relying on generalized variance [16], principal components analysis [19], or the Granger causality itself [21].

In this paper we will address the problem of partial conditioning to a limited subset of variables, in the framework of information theory. Intuitively, one may expect that conditioning on a small number of variables should be sufficient to remove indirect interactions if the connectivity pattern is sparse. We will show that this subgroup of variables might be chosen as the most informative for the driver variable and describe the application to simulated examples and a real data set.

#### 2. Materials and Methods

We start by describing the connection between the Granger causality and information-theoretic approaches like the transfer entropy in [22]. Let be a time series that may be approximated by a stationary Markov process of order , that is, . We will use the shorthand notation and , for , and treat these quantities as realizations of the stochastic variables and . The minimizer of the risk functional represents the best estimate of , given X, and corresponds [23] to the regression function . Now, let be another time series of simultaneously acquired quantities, and denote . The best estimate of , given and , is now . If the generalized Markov property holds, that is, then and the knowledge of does not improve the prediction of . Transfer entropy [22] is a measure of the violation of 2: it follows that the Granger causality implies nonzero transfer entropy [24]. Under the Gaussian assumption it can be shown that the Granger causality and transfer entropy are entirely equivalent and just differ for a factor two [25]. The generalization of the Granger causality to a multivariate fashion, described in the following, allows the analysis of dynamical networks [26] and to discern between direct and indirect interactions.

Let us consider time series ; the state vectors are denoted: being the window length (the choice of can be done using the standard cross-validation scheme). Let be the mean squared error prediction of on the basis of all the vectors (corresponding to linear regression or nonlinear regression by the kernel approach described in [24]). The multivariate Granger causality index is defined as follows: consider the prediction of on the basis of all the variables but and the prediction of using all the variables, then the causality measures the variation of the error in the two conditions, that is, Note that in [24] a different definition of causality has been used, The two definitions are clearly related by a monotonic transformation: Here, we first evaluate the causality using the selection of significative eigenvalues described in [26] to address the problem of overfitting in (5); then we use (6) and express our results in terms of because it is with this definition that causality is twice the transfer entropy, equal to , in the Gaussian case [25].

Turning now to the central point of this paper, we address the problem of coping with a large number of variables, when the application of the multivariate Granger causality may be questionable or even unfeasible, whilst bivariate causality would detect also indirect causalities. Here, we show that conditioning on a small number of variables, chosen as the most informative for the candidate driver variable, is sufficient to remove indirect interactions for sparse connectivity patterns. Conditioning on a large number of variables requires a high number of samples in order to get reliable results. Reducing the number of variables, that one has to condition over, would thus provide better results for small data sets. In the general formulation of the Granger causality, one has no way to choose this reduced set of variables; on the other hand, in the framework of information theory, it is possible to individuate the most informative variables one by one. Once that it has been demonstrated [25] that the Granger causality is equivalent to the information flow between the Gaussian variables, partial conditioning becomes possible for the Granger causality estimation; to our knowledge this is the first time that such approach is proposed.

Concretely, let us consider the causality ; we fix the number of variables, to be used for conditioning, equal to . We denote by the set of the variables, in , most informative for . In other words, maximizes the mutual information among all the subsets of variables. Then, we evaluate the causality Under the Gaussian assumption, the mutual information can be easily evaluated, see [25]. Moreover, instead of searching among all the subsets of variables, we adopt the following approximate strategy. Firstly, the mutual information of the driver variable, and each of the other variables, is estimated, in order to choose the first variable of the subset. The second variable of the subsets is selected among the remaining ones as those that, jointly with the previously chosen variable, maximize the mutual information with the driver variable. Then, one keeps adding the rest of the variables by iterating this procedure. Calling the selected set of variables, the set is obtained adding, to , the variable, among the remaining ones, with the greatest information gain. This is repeated until variables are selected. This greedy algorithm, for the selection of relevant variables, is expected to give good results under the assumption of sparseness of the connectivity.

#### 3. Results and Discussion

##### 3.1. Simulated Data

Let us consider linear dynamical systems on a lattice of nodes, with equations, for where ’s are the couplings, is the strength of the noise, and ’s are unit variance i.i.d. Gaussian noise terms. The level of noise determines the minimal amount of samples needed to assess that the structures recovered by the proposed approach are genuine and are not due to randomness as it happens for the standard Granger causality (see discussions in [24, 26]); in particular noise should not be too high to obscure deterministic effects. Firstly we consider a directed tree of 16 nodes depicted in Figure 1; we set equal to 0.9 for each directed link of the graph thus obtained and zero otherwise. We set . In Figure 2 we show the application of the proposed methodology to data sets generated by (8), 100 samples long, in terms of quality of the retrieved network, expressed in terms of sensitivity (the percentage of existing links that are detected) and specificity (the percentage of missing links that are correctly recognized as nonexisting). The bivariate analysis provides 100% sensitivity and 92% specificity. However, conditioning on a few variables is sufficient to put in evidence just the direct causalities while still obtaining high values of sensitivity. The full multivariate analysis (obtained as tends to 16) gives here a rather low sensitivity due to the low number of samples. This is a clear example where conditioning on a small number of variables is better than conditioning on all the variables at hand.

As another example, we now fix and construct couplings in terms of the well-known Zachary data set [27], an undirected network of 34 nodes. We assign a direction to each link, with equal probability, and set equal to 0.015, for each link of the directed graph thus obtained, and zero otherwise. The noise level is set as . The network is displayed in Figure 3: the goal is again to estimate this directed network from the measurements of time series on nodes.

In Figure 4 we show the application of the proposed methodology to data sets generated by (8), in terms of sensitivity and specificity, for different numbers of samples. The bivariate analysis detects several false interactions; however, conditioning on a few variables is sufficient to put in evidence just the direct causalities. Due to the sparseness of the underlying graph, we get a result that is very close to the one by the full multivariate analysis; the multivariate analysis here recovers the true network, and indeed the number of samples is sufficiently high. In Figure 5, concerning the stage of selection of variables upon which conditioning is performed, we plot the mutual information gain as a function of the number of variables included : it decreases as increases.

##### 3.2. EEG Epilepsy Data

We consider now a real data set from an -electrode grid that was implanted in the cortical surface of the brain of a patient with epilepsy [28]. We consider two 10-second intervals prior to and immediately after the onset of a seizure, called, respectively, the preictal period and the ictal period. In Figure 6 we show the application of our approach to the preictal period; we used the linear causality. The bivariate approach detects many causalities between the electrodes; most of them, however, are indirect. According to the multivariate analysis there is just one electrode that is observed to influence the others, even in the multivariate analysis: this electrode corresponds to a localized source of information and could indicate a putative epileptic focus. In Figure 6 it is shown that conditioning on or variables provides the same pattern corresponding to the multivariate analysis, which thus appears to be robust. These results suggest that the effective connectivity is sparse in the preictal period. As a further confirmation, in Figure 7 we plot the sum of all causalities detected as a function of the number of conditioning variables, for the preictal period; a plateau is reached already for small values of .

In Figure 8 the same analysis is shown w.r.t. the ictal period: in this case conditioning on or variables does not reproduce the pattern obtained with the multivariate approach. The lack of robustness of the causality pattern w.r.t. seems to suggest that the effective connectivity pattern, during the crisis, is not sparse. In Figures 9 and 10 we show, for each electrode and for the preictal and ictal periods, respectively, the total outgoing causality (obtained as the sum of the causalities on all the other variables). These pictures confirm the discussion above: looking at how the causality changes with may provide information about the sparseness of the effective connectivity.

#### 4. Conclusions

We have addressed the problem of partial conditioning to a limited subset of variables while estimating causal connectivity, as an alternative to full conditioning, which can lead to computational and numerical problems. Analyzing simulated examples and a real data set, we have shown that conditioning on a small number of variables, chosen as the most informative ones for the driver node, leads to results very close to those obtained with a fully multivariate analysis and even better in the presence of a small number of samples, especially when the pattern of causalities is sparse. Moreover, looking at how causality changes with the number of conditioning variables provides information about the sparseness of the connectivity.

#### References

- A. Barabasi,
*Linked*, Perseus, 2002. - S. Boccaletti, D. U. Hwang, M. Chavez, A. Amann, J. Kurths, and L. M. Pecora, “Synchronization in dynamical networks: evolution along commutative graphs,”
*Physical Review E*, vol. 74, no. 1, Article ID 016102, 5 pages, 2006. View at Publisher · View at Google Scholar · View at Scopus - Z. Gharhamani, “Learning dynamic bayesian networks,”
*Lecture Notes in Computer Science*, vol. 1387, pp. 168–197, 1997. - D. Yu, M. Righero, and L. Kocarev, “Estimating topology of networks,”
*Physical Review Letters*, vol. 97, no. 18, Article ID 188701, 4 pages, 2006. View at Publisher · View at Google Scholar · View at Scopus - K. Hlaváčková-Schindler, M. Paluš, M. Vejmelka, and J. Bhattacharya, “Causality detection based on information-theoretic approaches in time series analysis,”
*Physics Reports*, vol. 441, no. 1, pp. 1–46, 2007. View at Publisher · View at Google Scholar · View at Scopus - S. L. Bressler and A. K. Seth, “Wiener-Granger causality: a well established methodology,”
*NeuroImage*, vol. 58, no. 2, pp. 323–329, 2011. - M. Kamiński, M. Ding, W. A. Truccolo, and S. L. Bressler, “Evaluating causal relations in neural systems: granger causality, directed transfer function and statistical assessment of significance,”
*Biological Cybernetics*, vol. 85, no. 2, pp. 145–157, 2001. View at Publisher · View at Google Scholar · View at Scopus - K. J. Blinowska, R. Kuś, and M. Kamiński, “Granger causality and information flow in multivariate processes,”
*Physical Review E*, vol. 70, no. 5, Article ID 050902, 4 pages, 2004. View at Publisher · View at Google Scholar · View at Scopus - A. Seth, “Causal connectivity of evolved neural networks during behavior,”
*Network*, vol. 16, no. 1, pp. 35–54, 2005. View at Publisher · View at Google Scholar · View at Scopus - A. Roebroeck, E. Formisano, and R. Goebel, “Mapping directed influence over the brain using Granger causality and fMRI,”
*NeuroImage*, vol. 25, no. 1, pp. 230–242, 2005. View at Publisher · View at Google Scholar · View at Scopus - R. Ganapathy, G. Rangarajan, and A. K. Sood, “Granger causality and cross recurrence plots in rheochaos,”
*Physical Review E*, vol. 75, no. 1, Article ID 016211, 6 pages, 2007. View at Publisher · View at Google Scholar · View at Scopus - L. Faes, G. Nollo, and K. H. Chon, “Assessment of granger causality by nonlinear model identification: application to short-term cardiovascular variability,”
*Annals of Biomedical Engineering*, vol. 36, no. 3, pp. 381–395, 2008. View at Publisher · View at Google Scholar · View at Scopus - C. W. J. Granger, “Investigating causal relations by econometric models and cross-spectral methods,”
*Econometrica*, vol. 37, no. 3, pp. 424–438, 1969. - N. Wiener,
*The Theory of Prediction*, vol. 1, McGraw-Hill, New York, NY, USA, 1996. - J. F. Geweke, “Measures of conditional linear dependence and feedback between time series,”
*Journal of the American Statistical Association*, vol. 79, no. 388, pp. 907–915, 1984. - A. B. Barrett, L. Barnett, and A. K. Seth, “Multivariate Granger causality and generalized variance,”
*Physical Review E*, vol. 81, no. 4, Article ID 041907, 14 pages, 2010. View at Publisher · View at Google Scholar · View at Scopus - Y. Chen, S. L. Bressler, and M. Ding, “Frequency decomposition of conditional Granger causality and application to multivariate neural field potential data,”
*Journal of Neuroscience Methods*, vol. 150, no. 2, pp. 228–237, 2006. View at Publisher · View at Google Scholar · View at Scopus - G. Deshpande, S. LaConte, G. A. James, S. Peltier, and X. Hu, “Multivariate granger causality analysis of fMRI data,”
*Human Brain Mapping*, vol. 30, no. 4, pp. 1361–1373, 2009. View at Publisher · View at Google Scholar · View at Scopus - Z. Zhou, Y. Chen, M. Ding, P. Wright, Z. Lu, and Y. Liu, “Analyzing brain networks with PCA and conditional granger causality,”
*Human Brain Mapping*, vol. 30, no. 7, pp. 2197–2206, 2009. View at Publisher · View at Google Scholar · View at Scopus - L. Angelini, M. de Tommaso, D. Marinazzo, L. Nitti, M. Pellicoro, and S. Stramaglia, “Redundant variables and granger causality,”
*Physical Review E*, vol. 3, Article ID 037201, 4 pages, 81. View at Publisher · View at Google Scholar - D. Marinazzo, W. Liao, M. Pellicoro, and S. Stramaglia, “Grouping time series by pairwise measures of redundancy,”
*Physics Letters, Section A*, vol. 374, no. 39, pp. 4040–4044, 2010. View at Publisher · View at Google Scholar · View at Scopus - T. Schreiber, “Measuring information transfer,”
*Physical Review Letters*, vol. 85, no. 2, pp. 461–464, 2000. View at Publisher · View at Google Scholar · View at Scopus - A. Papoulis,
*Proability, Random Variables, and Stochastic Processes*, McGraw-Hill, New York, NY, USA, 1985. - D. Marinazzo, M. Pellicoro, and S. Stramaglia, “Kernel method for nonlinear Granger causality,”
*Physical Review Letters*, vol. 100, no. 14, Article ID 144103, 4 pages, 2008. View at Publisher · View at Google Scholar · View at Scopus - L. Barnett, A. B. Barrett, and A. K. Seth, “Granger causality and transfer entropy Are equivalent for gaussian variables,”
*Physical Review Letters*, vol. 103, no. 23, Article ID 238701, 4 pages, 2009. View at Publisher · View at Google Scholar · View at Scopus - D. Marinazzo, M. Pellicoro, and S. Stramaglia, “Kernel-Granger causality and the analysis of dynamical networks,”
*Physical Review E*, vol. 77, no. 5, Article ID 056215, 9 pages, 2008. View at Publisher · View at Google Scholar · View at Scopus - W. Zachary, “An information flow model for conflict and fission in small groups,”
*Journal of Anthropological Research*, vol. 33, no. 2, pp. 452–473, 1977. - M. A. Kramer, E. D. Kolaczyk, and H. E. Kirsch, “Emergent network topology at seizure onset in humans,”
*Epilepsy Research*, vol. 79, no. 2-3, pp. 173–186, 2008. View at Publisher · View at Google Scholar · View at Scopus