- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Computational and Mathematical Methods in Medicine
Volume 2012 (2012), Article ID 635103, 16 pages
Quantification of Effective Connectivity in the Brain Using a Measure of Directed Information
Department of Electrical and Computer Engineering, Michigan State University, East Lansing, MI 48824, USA
Received 18 November 2011; Revised 20 February 2012; Accepted 14 March 2012
Academic Editor: Luca Faes
Copyright © 2012 Ying Liu and Selin Aviyente. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Effective connectivity refers to the influence one neural system exerts on another and corresponds to the parameter of a model that tries to explain the observed dependencies. In this sense, effective connectivity corresponds to the intuitive notion of coupling or directed causal influence. Traditional measures to quantify the effective connectivity include model-based methods, such as dynamic causal modeling (DCM), Granger causality (GC), and information-theoretic methods. Directed information (DI) has been a recently proposed information-theoretic measure that captures the causality between two time series. Compared to traditional causality detection methods based on linear models, directed information is a model-free measure and can detect both linear and nonlinear causality relationships. However, the effectiveness of using DI for capturing the causality in different models and neurophysiological data has not been thoroughly illustrated to date. In addition, the advantage of DI compared to model-based measures, especially those used to implement Granger causality, has not been fully investigated. In this paper, we address these issues by evaluating the performance of directed information on both simulated data sets and electroencephalogram (EEG) data to illustrate its effectiveness for quantifying the effective connectivity in the brain.
Neuroimaging technologies such as the electroencephalogram (EEG) make it possible to record brain activity with high temporal resolution and accuracy. However, current neuroimaging modalities display only local neural activity rather than large-scale interactions between different parts of the brain. Assessment of the large-scale interdependence between these recordings can provide a better understanding of the functioning of neural systems [1, 2]. Three kinds of brain connectivity are defined to describe such interactions between recordings: anatomical connectivity, functional connectivity, and effective connectivity . Anatomical connectivity is the set of physical or structural connections linking neuronal units at a given time and can be obtained from measurements of the diffusion tensor [3, 4]. Functional connectivity captures the statistical dependence between scattered and often spatially remote neuronal units by measuring their correlations in either time or frequency domain. Effective connectivity describes how one neural system affects another [2, 4, 5], which can provide information about both the magnitude and the direction of the interaction.
The main approaches used to quantify the effective connectivity between two time series are model-based measures and information-theoretic measures . Granger-causality-based methods and dynamic causal modeling  are two widely used model-based measures. Granger causality is a widely used measure to describe the causality between two time series. It defines a stochastic process causing another process if the prediction of at the current time point, , is improved when taking into account the past samples of . This approach is appealing but gives rise to many questions on how to apply this definition to real data . Granger causality has been mostly applied within a linear prediction framework using a multivariate autoregressive (MVAR) model yielding methods such as directed transfer function (DTF), partial directed coherence (PDC), and directed partial correlation [9–12]. For example, Hesse et al. applied time-varying Granger causality to EEG data and found that conflict situation generates directional interactions from posterior to anterior cortical sites . Kamiński et al. applied DTF to EEG recordings of human brain during stage 2 sleep and located the main source of causal influence . Schelter et al. employed PDC to EEG recordings from a patient suffering from essential tremor . The extensions of Granger-causality-based methods, such as kernel Granger causality, generalized PDC (gPDC), and extended PDC (ePDC), have also found numerous applications in neuroscience [14–16]. However, Granger causality-based methods, especially those developed from MVAR models, are limited to capturing linear relations or require a priori knowledge about the underlying signal models . These approaches may be misleading when applied to signals that are known to have nonlinear dependencies, such as EEG data . DCM, on the other hand, can quantify nonlinear interactions by assuming a bilinear state space model. However, DCM requires a priori knowledge about the input to the system [7, 19] and is limited to a network with small size . Thus, a model-free measure detecting both linear and nonlinear relationships is desired.
Information theoretic tools [20–22], such as transfer entropy , address the issue of model dependency and have found numerous applications in neuroscience [17, 23, 24]. “Transfer entropy” (TE) proposed by Schreiber computes causality as the deviation of the observed data from the generalized Markov condition and is defined as  where and are the orders (memory) of the Markov processes and , respectively. is the joint probability of random variables , where and . Sabesan et al. employed TE to identify the direction of information flow for the intracranial EEG data and suggested that transfer entropy plays an important role in epilepsy research . Wibral et al. applied TE to magnetoencephalographic data to quantify the information flow in cortical and cerebellar networks . Vicente et al. extended the definition of TE and measured the information flow from to with a general time delay of , that is, replaced in the above equation with , and showed that TE has a better performance in detecting the effective connectivity for nonlinear interactions and signals affected by volume conduction such as real EEG/MEG recordings compared to linear methods . The performance of transfer entropy depends on the estimation of transition probabilities, which requires the selection of order or memory of the Markov processes and . “Directed transinformation” (T) introduced by Saito and Harashima  measures the information flow from the current sample of one signal to the future samples of another signal given the past samples of both signals. Hinrichs et al. used this measure to analyze causal interactions in event-related EEG-MEG experiments . However, this measure does not discriminate between totally dependent and independent processes . Recently, directed information proposed by Marko  and later reformalized by Massey, Kramer, Tatikonda, and others have attracted attention for quantifying directional dependencies [22, 28–31]. Directed information theory has been mostly aimed towards the study of communication channels with feedback. In recent years, new theoretical developments motivated the use of this measure in quantifying causality between two time series. In particular, Amblard and Michel  recently showed how directed information and Granger causality are equivalent for linear Gaussian processes and proved key relationships between existing causality measures and the directed information. Therefore, there has been a growing interest in applying this measure to applications in signal processing, neuroscience, and bioinformatics. For example, it has been successfully used to infer genomic networks  and to quantify effective connectivity between neural spike data in neuroscience [31, 33, 34]. In order to detect both linear and nonlinear relationships, in this paper, we propose directed information as a powerful measure to quantify the effective connectivity in the brain.
The theoretical advantages of DI over existing measures have been noted in literature [31, 33, 34]. However, until now the benefits of using DI for capturing the effective connectivity in the brain through neurophysiological data have not been illustrated thoroughly and formally. In addition, because of the relationship between Granger causality and directed information, in this paper, we mainly focus on the comparison between these two measures and investigate the advantage of DI over Granger-causality-based model measures. Theoretical developments only proved the equivalence between these two measures for the case that the time series are distributed as Gaussian in a linear model. However, to date, there has not been much work that compares the actual performance of DI and Granger-causality-based measures for realistic signal models, including both linear and nonlinear interactions. Moreover, most applications of DI to real data have been limited to using either parametric density models for the data or making assumptions about the time dependencies such as assuming a first-order Markov chain and have not considered the difficulties associated with estimating DI from a finite sample size . For complex systems, the computational complexity and the bias of the DI estimator increase with the length of the signal. The main contribution of this paper is to address these issues by evaluating the performance of DI and Granger-causality-based methods under a common framework without making any assumptions about the data distribution. In this paper, we first give a brief introduction to directed information and its computation based on nonparametric estimation methods. We propose a modified time-lagged directed information measure that simplifies the DI computation by reducing the order of the joint entropy terms while still quantifying the causal dependencies. We then evaluate the performance of DI for quantifying the effective connectivity for linear and nonlinear autoregressive models, linear mixing models, single source models, and dynamic chaotic oscillators in comparison to existing causality measures, in particular with Granger causality. Finally, we apply our method to EEG data to detect the effective connectivity in the brain.
2. Materials and Methods
2.1. Definitions and Notations
In this section, we will first review some common notations and information-theoretic definitions that will be used throughout this paper. Let be a random process with length and be the joint probability of random variables . will be used to define the time-delayed version of sequence , which is also equivalent to .
Given two continuous random variables and , the mutual information (MI) is defined as follows (All integrals in the paper are from to unless otherwise specified.): where is the joint probability density function (pdf) of and , and , are the marginal pdfs of and , respectively. with equality if and only if and are independent . In information theory, mutual information can be interpreted as the amount of uncertainty about that can be reduced by observation of , or the amount of information can provide about , that is, . Since with equality if and only if and are independent; that is, conditioning reduces entropy .
For any three random variables , and , if the conditional distribution of depends only on and is conditionally independent of , that is, , then , , and are said to form a Markov chain, denoted by . In this case, the conditional mutual information between and given defined as is equal to 0 .
2.2. Directed Information
Mutual information can be extended to random vectors or sequences and as , where . However, mutual information is a symmetric measure and does not reveal any directionality or causality between two random sequences. Massey addressed this issue by defining the directed information from a length sequence to  as follows: where is the entropy of the sequence causally conditioned on the sequence , and is defined as which differs from in that replaces in each term on the right-hand side of (4), that is, only the causal influence of the time series up to the current time sample on the process is considered.
An alternative definition of the directed information is proposed by Tatikonda in terms of Kullback-Leibler (KL) divergence . It shows that the difference between mutual information and directed information is the introduction of feedback in the definition of directed information [22, 30, 31]. Mutual information and directed information expressed by KL divergence are written as where is the feedback factor influenced by the feedback in the system, that is, the probability that the input at current time is influenced by the past values of both itself and . If there is no feedback, then and . In fact, , where and is defined as the feedforward factor affected by the memory of the system. If the system is memoryless, then .
2.3. Directed Information versus Granger Causality
Granger quantifies causality so that the time series causes if the variance of the prediction error for at the present time is reduced by including past measurements from . Based on Granger’s definition of causality, Geweke introduced the Geweke’s indices to quantify the causal linear dependencies under Gaussian assumptions . Amblard and Michel proved that the directed information rate and Geweke’s indices are equal for Gaussian processes  as indicated by where stands for the time-delayed sequence with being the length of the signal, is the directed information rate; that is, is the asymptotic variance of the prediction residue when predicting from the observation of , and refers to the linear feedback measure from random processes to defined by Geweke . This equality shows that asymptotically the DI rate is equivalent to the gain in information by predicting using the past values of both and compared to only using the past samples of , which is similar to the definition of Granger causality. Moreover, Amblard and Michel proved the equality of directed information and Granger’s approach for multivariate time series in the case of Gaussian distributions .
2.4. Computation of Directed Information
The definition of DI for two length sequences and can also be rewritten in terms of the total change of joint entropy or mutual information along time as follows:
From the above equations, we can observe that the computation of DI requires the estimation of joint probabilities of high-dimensional random variables over time. If and are normally distributed, the joint entropy can be estimated based on the covariance matrices. However, for EEG data, the distribution is usually not Gaussian. The nonparametric entropy and mutual information estimators, such as plug-in estimator, m-spacing estimator, and Kozachenko and Leonenko (KL) estimator, have been extensively addressed in literature [38, 39]. In this paper, directed information estimation based on mutual information is used to estimate DI directly from EEG data by using adaptive partitioning method discussed in . However, when the length of the signal increases, the computational complexity, the bias, and the variance of these estimators increase immensely with limited sample sizes. Methods that can reduce the dimension and simplify the computation of DI are needed.
In order to simplify the estimation of DI, we first clarify the connection between the definition of DI used in information theory and the definition as it applies to physical time series. In a physical recording system, if starts to influence after time points or with a delay of samples, we need to record at least time points to obtain points of the time sequence that have been affected by . The directed information rate from time series to can be defined as . We have where (11) comes from the fact that is independent of , and (12) is derived using the fact that has no effect on because of the time delay between these two time series. For two physical recordings and with length and a lag of , the last equation shows that DI rate for these two time series is equivalent to DI rate for two random processes with length that are not synchronized in time. In fact, may be indexed as when using the information theoretic indexing, which indexes the signal not according to the physical time point but based on when the receiver receives its first piece of information. Therefore, directed information rate computed by using physical time indices is equivalent to the directed information rate using information theoretic indices for two systems that interact through a time delay. Moreover, when the length of the signal is long enough, the directed information value using both indices will be equivalent.
Once the definition of directed information is extended from random vectors to two physical time series, we propose a modified time-lagged DI to simplify the computation of DI, which is an extension of time-lagged DI proposed for every two samples of and in  to general signal models. As we mentioned before, as the length of the signal increases, the computational complexity, the bias, and the variance of estimating DI increase immensely with limited sample sizes. In addition, the directed information defined for the physical system is actually a DI with a lag of samples over a time window with length . Therefore, an intuitive way to simplify the computation is to apply DI with lag over a small window. We first give the definition of time-lagged DI for two time series and with length at the th time sample for a block of two time samples with a time delay of : where , is the time lag between the two time series, and is the length of the whole time series. Therefore, the total directed information over the whole time series in terms of the time-lagged DI can be simplified as  (the details of the derivation are given in ),
The time-lagged DI is equivalent to the original definition of DI when is equal to the actual time delay of the system, the signals and follow a single-order model, and only depends on one past sample of itself, . However, these assumptions are not always true. Therefore, we propose the modified time-lagged DI to address these issues.
Consider a general Markov model, where and are time series with a lag of and , where , , is the order of the process , and is the order of the process . In this model, it is assumed that starts to influence with a delay of samples, and the order of the model is . When the length of the signal is large enough, then (15) can be further simplified as
Since , follows a Markov chain. According to Markov chain property, which means . Therefore, where the second equality is using the Markov property, and the inequality comes from the fact that conditioning reduces entropy. For a general Markov model, where and are stationary statistical processes without instantaneous interaction, such as , the modified time-lagged directed information (MDI) is defined as the upper bound of DI: where we let , to reduce the number of parameters. Note that letting does not lose any of the information flow compared to using the actual time delay, . The only drawback of letting is that the computational complexity of estimating the joint entropies increases since the length of the window to compute MDI increases and the dimensionality increases. The main reason why we let is because estimating the actual value for the delay accurately is not practical when the amount of data is limited. In a lot of similar work such as in , different values of are tested to choose the best one which is not computationally efficient either.
According to (20), modified time-lagged directed information is the upper bound of directed information, that is, MDI ≥ DI. Moreover, MDI is a more general extension of time-lagged DI introduced in our previous work and has two major advantages. First, MDI considers the influence of multiple past samples of on the DI value. Second, it takes into account models with multiple orders; that is, is influenced by different time lags of . The modified time-lagged directed information extends the length of the window from 2 to , which is closer to the actual information flow. When and are normally distributed, the computational complexity of the MDI is and the computational complexity of the original definition of DI is (using LU decomposition ). Therefore, the computation of MDI is more efficient than that of the original definition of DI.
2.5. Order Selection
For the implementation of MDI, we need to determine the maximum order of the model . Criterions such as Akaike’s final prediction error (FPE) can be used to determine the order of the signal model . However, this criterion is based on the assumption that the original signal follows a linear AR model and may lead to false estimation of the order when the underlying signal model is nonlinear. Therefore, model-free order selection methods, such as the embedding theorem , are needed. For the simplification of computation or parameter estimation, we are only interested in a limited number of variables that can be used to describe the whole system. Suppose we have a time series , and the time-delay vectors can be reconstructed as . Projecting the original system to this lower-dimensional state space-depends on the choice of and , and the optimal embedding dimension is related to the order of the model . A variety of measures such as mutual information can be used to determine . For discrete time signals, usually the best choice of is 1 . To determine , Cao criterion based on the false nearest neighbor procedure  is used to determine the local dimension. The underlying concept of nearest neighbor is that if is the embedding dimension of a system, then any two points that stay close in the -dimensional reconstructed space are still close in the -dimensional reconstructed space; otherwise, these two points are false nearest neighbors [19, 43]. The choice of , that is, the model order , is important for DI estimation. If is too small, we will lose some of the information flow from to . If it is too large, the computational complexity of MDI will be very high, causing the bias and the variance of the estimators to increase.
2.6. Normalization and Significance Test
Since and , then
Therefore, where indicating the instantaneous information exchange between processes and . For a physical system without instantaneous causality, that is, , then and . A normalized version of DI, which maps DI to the range, is used for comparing different interactions, where for a unidirectional system with no instantaneous interaction between and , and ; otherwise, if there is no causal relationship between the two signals, the values of and are very close to each other.
In order to test the null hypothesis of noncausality, the causal structure between and is destroyed. For each process with multiple trials, we shuffle the order of the trials of the time series 100 times to generate new observations , . In this way, the causality between and for each trial is destroyed, and the estimated joint probability changes . We compute the DI for each pair of data ( and ). A threshold is obtained at a significance level such that 95% of the directed information for randomized pairs of data () is less than this threshold. If the DI value of the original pairs of data is larger than this threshold, then it indicates there is significant information flow from to .
2.7. Simulated Data
To test the validity and to evaluate the performance of DI for quantifying the effective connectivity, we generate five different simulations. We use these simulation models to compare DI with classical Granger causality (GC) for quantifying causality of both linear and nonlinear autoregressive models, linear mixing models, single source models, and Lorenz systems. The Matlab toolbox developed by Seth is used to compute the GC value in the time domain. GC is also normalized to the range for comparison purposes . The performance of GC depends on the length of the signal, whereas the performance of DI relies on the number of realizations of time series. Therefore, for each simulation, the length of the generated signal for implementing GC is equal to the number of realizations for DI. The significance of DI values are evaluated by shuffling along the trials, while the significance of GC values are evaluated by shuffling along the time series.
Example 1 (Multiple Order Bivariate Linear Autoregressive Model). In this example, we evaluate the performance of DI on a general bivariate linear model,
In this bivariate AR model with a delay and order , controls the coupling strength between the signals and . The initial values of and and the noise and are all generated from a Gaussian distribution with mean 0 and standard deviation 1. All coefficients (, , , and ) are generated from Gaussian distributions with zero mean and unit variance with unstable systems being discarded. To evaluate the performance of directed information, we generate the bivariate model 4096 times with the same parameters but different initial values. is varied from 0.1 to 1 with a step size of 0.1, and ; that is, is influenced by through multiple time lags. Without loss of generality, we repeat the simulation 10 times, and average and over 10 simulations for different values. For each simulation, the threshold is evaluated by trial shuffling, and the average threshold is obtained. For GC, the length of the generated signal is chosen as 4096, which is the same as the number of realizations for DI. The GC values in two directions and the corresponding thresholds at the 5% significance level are obtained.
Example 2 (Multiple-Order Bivariate Nonlinear Autoregressive Model). In this example, we evaluate the performance of DI on a general bivariate nonlinear model
For this bivariate nonlinear AR model, the setting for the coupling strength and the generation of , , , , , , , , , , , and are the same as in Example 1. and interact nonlinearly through the sigmoid function. Parameters of this function and control the threshold level and slope of the sigmoidal curve, respectively. We set and . DI value and its threshold are averaged over 10 simulations for different . The GC values in two directions and the corresponding thresholds at 5% significance level are obtained.
Example 3 (Linear Mixing Model). In this example, we test the effectiveness of DI in inferring effective connectivity when there is linear mixing between these two signals. Linear instantaneous mixing is known to exist in human noninvasive electrophysiological measurements such as EEG or MEG. Instantaneous mixing from coupled signals onto sensor signals by the measurement process degrades signal asymmetry . Therefore, it is hard to detect the causality between the two signals. For unidirectional coupled signal pairs described in (25) to (28), we create two linear mixtures and as follows: where controls the amount of linear mixing and is varied from 0.05 to 0.45 with a step size of 0.05, and is fixed to 0.8 for both models. When , the two signals are identical. Both DI and GC are used to quantify the information flow between and in the two directions.
Example 4 (Single-Source Model). A single source is usually observed on different signals (channels) with individual channel noises , which is common in EEG signals due to the effects of volume conduction. In this case, false positive detection of effective connectivity occurs for methods such as Granger causality , which means that GC has low specificity. We generate two signals and as follows to test the specificity of DI when there is no significant information flow from one signal to the other signal. We have where is the common source generated by an autoregressive model, order , and are generated from a Gaussian distribution with mean 0 and standard deviation 1. is measured on both sensors and . is further corrupted by independent Gaussian noise with 0 mean and unit variance. controls the signal to noise ratio (SNR) in and is varied from 0.1 to 0.9 with a step size of 0.1, corresponding to SNR in the range of dB.
Example 5 (Nonlinear Dynamic System). In this example, we illustrate the applicability of DI to coupled Lorenz oscillators with a certain delay. The Lorenz oscillator is a three-dimensional dynamic system that exhibits chaotic behavior. Synchronization of two Lorenz systems has been widely investigated for the analysis of EEG data because the dynamic interactions related to the behavior of the cortex can be exemplified by these coupled systems . In the following, we examined two asymmetric coupled Lorenz oscillators () and () as follows : where each equation is a first-order differential equation. , , , and represents the time delay between two coupled components of these two oscillators, that is, and . corresponds to the coupling strength and is varied from 0.1 to 1 with a step size of 0.2. The differential equations are numerically integrated with a time step of 0.01 using Euler’s method , corresponding to a delay of 2 time samples between and . The initial conditions of these six components are randomly generated from a Gaussian distribution with zero mean and unit variance. We generate 100 samples, and the first 90 samples are discarded to eliminate the initial transients. We compute the information flow in two directions over 10 time points, and the significance of the obtained DI value is verified by trial shuffling.
2.8. Biological Data
In this paper, we examine EEG data from ten undergraduates at Michigan State University drawn from an ongoing study of relationships between the error-related negativity (ERN) and individual differences (Participants for the present analysis were drawn from samples reported on in [50, 51]) such as worry and anxiety. ERN is a brain potential response that occurs following performance errors in a speeded reaction time task . All participants retained for analysis make at least six errors for computation of stable ERNs, as in . Participants complete a letter version of the Eriksen Flanker task . Stimuli are presented on a Pentium R Dual Core computer, using Presentation software (Neurobehavioral systems, Inc.) to control the presentation and timing of stimuli, the determination of response accuracy, and the measurement of reaction times. Continuous electroencephalographic activity is recorded by 64 Ag-AgCl electrodes placed in accordance with the 10/20 system. Electrodes are fitted in a BioSemi (BioSemi, Amsterdam, The Netherlands) stretch-lycra cap. All bioelectric signals are digitized at 512 Hz using ActiView software (BioSemi). For each subject, EEG data are preprocessed by the spherical spline current source density (CSD) waveforms to sharpen event-related potential (ERP) scalp topographies and eliminate volume conduction . In addition, a bandpass filter is used to obtain signals in the theta band. In this study we focus on 33 electrodes corresponding to the frontal, central, and parietal regions of the brain. For each pair of 33 electrodes and for each subject, the effective connectivity is quantified by computing the modified time-lagged DI over 70 trials and a model order of in the theta band. The model order or the length of the time window is determined by the Cao Criterion. We also apply Granger causality to the same data and compare its performance with directed information.
Previous work indicates that there is increased synchronization associated with ERN for the theta frequency band (4–8 Hz) and ERN time window 25–75 ms after the response for error responses (ERN) in the anterior cingulate cortex (ACC), in particular between the lateral prefrontal cortex (lPFC) and medial prefrontal cortex (mPFC) . In this paper, we wish to verify these existing findings using the proposed DI measure and to further infer the directional causality underlying these dependencies.
3. Results and Discussion
In this section, we first evaluate the effectiveness of directed information on quantifying both linear and nonlinear causal relationships through simulated data and compare the performance of directed information with GC. We then apply the directed information to real EEG data to reveal the pair-wise information flow in the brain.
3.1. Simulated Data
Example 1 (Multiple-Order Bivariate Linear Autoregressive Model). In this example, the DI value in two directions averaged across 10 simulations with different is shown in Figure 1(a). The performance of GC is shown in Figure 1(b). The estimated order of the model is , which is in accordance with the simulation model. controls the coupling strength between and . We observe that is significant for all values of . On the contrary, is less than the threshold, which indicates the acceptance of the null hypothesis that there is no significant causal information flow from to . Since GC uses a linear autoregressive framework for quantifying causality; in this example, GC detects the causality relationship between and successfully; that is, the information flow from to is significant for all while it is insignificant for the opposite direction. It is also interesting to note that GC and DI exhibit similar behavior across different values of , indicating the equivalency of the two measures for linear Gaussian signal models.
Example 2 (Multiple-Order Bivariate Nonlinear Autoregressive Model). In this example, the performance of DI and GC for the nonlinear autoregressive model in (27) and (28) averaged across 10 simulations with different are evaluated as shown in Figure 2. The estimated order of the model is 5. We observe that when is less than 0.3, the coupling strength between and is weak and the DI value in both directions is not significant. As increases, increases and becomes significant. decreases with increasing and is still less than the threshold as expected. The results indicate increased unidirectional information flow from to with increasing and show that detecting the information flow in nonlinear processes is more difficult especially when the coupling strength is low. GC fails to detect the information flow from to for all . Since GC is implemented in a linear framework, the estimated order and the model itself do not match with the nonlinearity of the signal. Therefore, it cannot detect nonlinear causality.
Example 3 (Linear Mixing Model). For this example, the DI value and GC value averaged across 10 simulations with changing linear mixing coefficient for both linear and nonlinear AR models are shown in Figure 3. The estimated order of the model is 5 as before. When , the two observed mixing signals are identical, and we expect to see no significant information flow in the two directions. We observe that, for the linear AR model, directed information detects the causality between and when is smaller than 0.4. When is larger than 0.4, the causality between and is hard to detect because of the strong mixing; that is, and are almost identical, and the information flow in both directions becomes insignificant. Compared to DI, GC only detects the causality from to when the mixing is weak (), indicating that GC is more vulnerable to linear mixing. It is probably due to the fact that GC is sensitive to the mixture of signals, and the assumed signal model does not match with the original signal . For the nonlinear AR model, DI fails to detect causality when is larger than 0.1, which indicates that linear mixing of nonlinear source models makes it harder to detect effective connectivity compared to mixing of linear source models. On the other hand, GC fails to detect any causality even when since it cannot detect nonlinear interactions.
Example 4 (Single-Source Model). We use the single source model to test the specificity of DI. The DI value and GC value averaged across 100 simulations for changing for a single source model are shown in Figure 4. The estimated order of the model is 5 as before. In addition, the false positive rate using both DI and Granger causality with increasing is also calculated. We observe that the information flow in two directions using DI is less than the threshold for all values of , which indicates the acceptance of the null hypothesis that there is no significant causal information flow from to or to . Note that DI is normalized by the mutual information. For a common source model, the instantaneous information exchange between and contributes mostly to the mutual information between and . Thus, according to (23), and normalized by mutual information are close to 0 and less than the threshold from the randomized data pairs. The false positive rate of DI is 0 for all . Therefore, DI is able to discriminate between instantaneous mixing from actual causality and is very robust to noise. For GC, when is small (<0.2) or large (>0.9), the value of GC is less than or very close to the threshold in both directions thus indicating that there is no causal information flow between the two processes. However, GC fails to accept the null hypothesis when is between 0.3 to 0.9 and detects a nonexisting effective connectivity. GC reaches its maximum value when . This is due to the fact that GC is close to 0 when two processes and are independent or identical, that is, when and . Based on the definition of GC, the prediction of at the current time point will not be improved by taking into account the past samples of for these processes . Therefore, as increases from 0 to 0.5, becomes the most different from ; therefore, it can provide more new information about and the GC increases. As increases from 0.5 to 1, becomes independent of , and the GC decreases. The false positive rate of GC is not equal to 0 for all values of , which indicates that it has lower specificity compared to DI. Therefore, GC is not robust to the effect of a common source and may infer false positive effective connectivity. This simulation indicates that DI is more sensitive and discriminative about the information flow patterns in the presence of volume conduction, which means it is a more promising method to capture the effective connectivity for real EEG data.
Example 5 (Nonlinear Dynamic System). In this example, the DI values and GC values between and of two asymmetric coupled Lorenz systems are computed with coupling strength being set from 0.1 to 1. The estimated order of the model is 3. Though this is larger than the actual model order, our method will not lose any information except for the increased computational complexity. The results are shown in Figure 5. The results show that DI values from to increase with the coupling strength and are significant for all values of . In addition, there is no significant causal information flow from to . Therefore, DI can effectively detect the causality in a nonlinear dynamic system. On the contrary, GC cannot detect any significant information flow for all values. It is due to the fact that the model selected for implementing GC is not consistent with the dynamic characteristics of the system.
3.2. EEG Data
Previous work indicates that there is increased information flow associated with ERN for the theta frequency band (4–8 Hz) and ERN time window 25–75 ms for error responses compared to correct responses in particular between mPFC and lPFC regions . In addition, Cavanagh et al. have shown that there is increased synchronization for error trials between electrode pairs, such as FCz-F5 and FCz-F6, compared to the synchrony between FCz-CP3 and FCz-CP4 . The DI and GC values for each pair of electrodes averaged over 10 subjects are computed over a time window of 53 time points (100 ms). The estimated order of the model for each electrode pairs is 3. In order to control the error rates for multiple hypothesis testing for all pairs of electrodes, the method proposed by Genovese et al. is used in this paper . To implement this procedure, for two electrodes with time series and , we first shuffle the order of the trials of 100 times to generate new observations , . The value of is obtained by comparing it with DI values from randomized pairs of data , . We then obtain the threshold for all values (33 × 33 × 10) by controlling the FDR bound as 0.05. For , if the value is less than , then the directed information flow from to is significant; otherwise, it is not significant. Electrode pairs between which the information flow is significant in at least one of the ten subjects are shown in Figure 6(b). We also test the significance of Granger causality in the same way. When the FDR is controlled at 0.05, the information flow between electrode pairs is significant if the -value of DI or GC is less than 0.01. Electrode pairs that have significant causality relationship using both measures are shown in Figure 6. In Figures 6(a) and 6(c), each small circle shows the directed information and Granger causality from a particular electrode to other electrodes. In Figures 6(b) and 6(d), each small circle shows electrode pairs that have significant causality relationship. The results indicate that DI detects strong information flow from the frontal region (e.g., F5 and F6) to the frontal-central region (e.g., FC2 and FCz) corresponding to the lateral prefrontal cortex (lPFC) and medial prefrontal cortex (mPFC). In addition, the central-parietal region (e.g., CPz, CP1, and CP2) around the midline, corresponding to the motor cortex, has strong influence on the central and frontal regions (e.g., FCz and F6) since this is a speeded response task involving the motor cortex. The details of the significant electrode interactions are shown in Table 1. These results are aligned with the previous work in , which shows that error processing is controlled by the communication between the lateral prefrontal cortex and medial prefrontal cortex. When GC is applied to the same data, the information flow pattern around the midline is similar to the DI. However, the information flow from the lateral prefrontal cortex to the rest of the brain is significant. On one hand, the similar patterns of connectivity using both measures verify the validity of proposed DI computation algorithm. On the other hand, GC shows significance over a wide region of the brain especially in the lateral areas compared to DI, which may be due to GC’s low specificity to volume conduction in the form of a common source. Previous work and our simulation in Example 4 have indicated that Granger-causality-based measures may infer erroneous effective connectivity in the case of the common source as seen in EEG data [19, 46]. However, without ground truth, we cannot confirm that some links reported as significant by GC are spurious and due to volume conduction in a conclusive manner, but the results from DI agree more with the suggestions in , that most of the increase in connectivity during cognitive control, that is, ERN, should be between medial prefrontal cortex and lateral prefrontal cortex, compared to the results of GC. Therefore, DI is more sensitive and discriminative about the information flow patterns compared to GC for real neurophysiological data.
In this paper, we illustrated the advantages of a new directed information measure over Granger-causality-based measures for quantifying the effective connectivity in the brain. In order to illustrate the advantages of this measure, first, we applied directed information measure to identify the causality relationships for both linear and nonlinear AR models, linear mixing models, single source models, and Lorenz systems and compare its performance with Granger causality. Directed information is shown to be more effective in detecting the causality of different systems compared to Granger causality. We then applied the directed information measure on EEG data from a study containing the error-related negativity to infer the information flow patterns between different regions. The results showed that the directed information measure can capture the effective connectivity in the brain between the mPFC and lPFC areas as predicted by previous work.
Directed information, as a model-free measure, is able to detect both linear and nonlinear causality relationships between two signals. However, other model-free entropy-based measures would also detect effective connectivity such as transfer entropy and directed transinformation. Directed transinformation introduced by Saito measures the information flow from the current sample of one signal to the future samples of another signal given the past samples of both signals but does not discriminate between totally dependent and independent processes. Transfer entropy and directed information are very closely related to each other. Transfer entropy quantifies the information gained at each time step by measuring the deviation of the observed data from the generalized Markov condition. Therefore, the definition of transfer entropy implicitly assumes a stationary Markov process . Compared to transfer entropy, directed information quantifies the sum of information obtained over the whole time series  and does not make any assumptions about the underlying signal model. Thus, theoretically, the original definition of directed information can apply to any signal models. In real applications, in order to simplify the computation of directed information, we usually make certain assumptions about the underlying signal model such as the modified time-lagged DI proposed in this paper, which basically assumes a stationary Markov process similar to transfer entropy. In addition, Amblard and Michel proved that, for a stationary process, directed information rate can be decomposed into two parts, one of which is equivalent to the transfer entropy when in (1) and the other to the instantaneous information exchange rate . In another words, for a physical system without instantaneous interactions between its subsystems, the rate of these two measures, directed information and transfer entropy, is equivalent asymptotically as the length of the signal goes to infinity.
There are still remaining issues with the implementation of directed information. First, the performance of directed information relies on accurate estimation from limited sample sizes that introduces bias to the estimated values. This problem can be addressed by either using parametric density models or improving existing mutual information and entropy estimators. Recently, Zhao et al. proposed an universal algorithm to estimate directed information for stationary ergodic processes by using sequential probability assignment, which may be used to improve the effective connectivity results discussed in this paper . Second, the performance of directed information relies on the selection of the model order. If the order of the model is too small, it will lose the information from to . If it is too large, the computational complexity is very high. In addition to classical embedding dimension determination methods such as the Cao criterion used in this paper, Faes et al. proposed a sequential procedure to determine the embedding dimension of multivariate series . This method is based on an information-theoretic technique and shows promising performances for various signal models, which may be extended to DI computation in the future. Third, directed information does not discriminate between direct and indirect interactions among multivariate time series. However, this is not a shortcoming of DI since DI does not assume any particular signal interaction model: bivariate or multivariate. Similar to other information theoretic measures, such as mutual information, whether the particular measure can identify interactions between multiple processes depends on how the measure is applied. For example, in the case of mutual information, though the original definition is for two random processes and , it is possible to extend it to multiple processes . Similarly, we can apply DI over multiple processes using conditional directed information such as the definition given by Kramer. We address this issue in a previous paper  by using conditional directed information and develop an algorithm to infer the actual network. Similarly, GC originally is defined for two time series that a stochastic process causing another process if the prediction of at the current time point, , is improved when taking into account the past samples of . However, in application it has been extended to multiple processes through the use of multivariate AR models. Future work will focus on the comparison of these two measures in a multivariate setting.
The authors would like to thank Dr. Jason Moser from the Department of Psychology at Michigan State University for sharing his EEG data. This work was in part supported by the National Science Foundation under Grants Nos. CCF-0728984 and CAREER CCF-0746971.
- S. L. Bressler, “Large-scale cortical networks and cognition,” Brain Research Reviews, vol. 20, no. 3, pp. 288–304, 1995.
- O. Sporns, D. R. Chialvo, M. Kaiser, and C. C. Hilgetag, “Organization, development and function of complex brain networks,” Trends in Cognitive Sciences, vol. 8, no. 9, pp. 418–425, 2004.
- M. A. Koch, D. G. Norris, and M. Hund-Georgiadis, “An investigation of functional and anatomical connectivity using magnetic resonance imaging,” NeuroImage, vol. 16, no. 1, pp. 241–250, 2002.
- K. J. Friston, “Functional and effective connectivity: a review,” Brain Connectivity, vol. 1, no. 1, pp. 13–36, 2011.
- K. J. Friston, “Functional and effective connectivity in neuroimaging: a synthesis,” Human Brain Mapping, vol. 2, no. 1-2, pp. 56–78, 1994.
- E. Pereda, R. Q. Quiroga, and J. Bhattacharya, “Nonlinear multivariate analysis of neurophysiological signals,” Progress in Neurobiology, vol. 77, no. 1-2, pp. 1–37, 2005.
- K. J. Friston, L. Harrison, and W. Penny, “Dynamic causal modelling,” NeuroImage, vol. 19, no. 4, pp. 1273–1302, 2003.
- C. W. J. Granger, “Testing for causality. A personal viewpoint,” Journal of Economic Dynamics and Control, vol. 2, no. 1, pp. 329–352, 1980.
- L. A. Baccala and K. Sameshima, “Partial directed coherence: a new concept in neural structure determination,” Biological Cybernetics, vol. 84, no. 6, pp. 463–474, 2001.
- W. Hesse, E. Möller, M. Arnold, and B. Schack, “The use of time-variant EEG Granger causality for inspecting directed interdependencies of neural assemblies,” Journal of Neuroscience Methods, vol. 124, no. 1, pp. 27–44, 2003.
- M. Kamiński, M. Ding, W. A. Truccolo, and S. L. Bressler, “Evaluating causal relations in neural systems: Granger causality, directed transfer function and statistical assessment of significance,” Biological Cybernetics, vol. 85, no. 2, pp. 145–157, 2001.
- W. Mader, D. Feess, D. Saur, et al., “Investigating multivariate systems using directed partial correlation,” International Journal of Bioelectromagnetism, vol. 12, no. 1, pp. 21–25, 2001.
- B. Schelter, M. Winterhalder, M. Eichler et al., “Testing for directed influences among neural signals using partial directed coherence,” Journal of Neuroscience Methods, vol. 152, no. 1-2, pp. 210–219, 2006.
- D. Marinazzo, W. Liao, H. Chen, and S. Stramaglia, “Nonlinear connectivity by Granger causality,” NeuroImage, vol. 58, no. 2, pp. 330–338, 2010.
- L. Leistritz, T. Weiss, J. Ionov, K. J. Bär, W. H. R. Miltner, and H. Witte, “Connectivity analysis of somatosensory evoked potentials to noxious intracutaneous stimuli in patients with major depression,” Methods of Information in Medicine, vol. 49, no. 5, pp. 484–491, 2010.
- L. Faes and G. Nollo, “Extended causal modeling to assess partial directed coherence in multiple time series with significant instantaneous interactions,” Biological Cybernetics, vol. 103, no. 5, pp. 387–400, 2010.
- H. Hinrichs, T. Noesselt, and H. J. Heinze, “Directed information flow—a model free measure to analyze causal interactions in event related EEG-MEG-experiments,” Human Brain Mapping, vol. 29, no. 2, pp. 193–206, 2008.
- F. Lopes da Silva, J. P. Pijn, and P. Boeijinga, “Interdependence of EEG signals: linear versus nonlinear associations and the significance of time delays and phase shifts,” Brain Topography, vol. 2, no. 1-2, pp. 9–18, 1989.
- R. Vicente, M. Wibral, M. Lindner, and G. Pipa, “Transfer entropy-a model-free measure of effective connectivity for the neurosciences,” Journal of Computational Neuroscience, vol. 30, no. 1, pp. 45–67, 2011.
- T. Schreiber, “Measuring information transfer,” Physical Review Letters, vol. 85, no. 2, pp. 461–464, 2000.
- Y. Saito and H. Harashima, Recent Advances in EEG and EMG Data Processing, Elsevier, Amsterdam, The Netherlands, 1981.
- J. Massey, “Causality, feedback and directed information,” in Proceedings of the International Sympoium on Information Theory and Its Applications, pp. 27–30, 1990.
- M. Rubinov and O. Sporns, “Complex network measures of brain connectivity: uses and interpretations,” NeuroImage, vol. 52, no. 3, pp. 1059–1069, 2010.
- M. Lungarella and O. Sporns, “Mapping information flow in sensorimotor networks,” PLoS Computational Biology, vol. 2, no. 10, pp. 1301–1312, 2006.
- S. Sabesan, L. B. Good, K. S. Tsakalis, A. Spanias, D. M. Treiman, and L. D. Iasemidis, “Information flow and application to epileptogenic focus localization from intracranial EEG,” IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 17, no. 3, pp. 244–253, 2009.
- M. Wibral, B. Rahm, M. Rieder, M. Lindner, R. Vicente, and J. Kaiser, “Transfer entropy in magnetoencephalographic data: quantifying information flow in cortical and cerebellar networks,” Progress in Biophysics and Molecular Biology, vol. 105, no. 1-2, pp. 80–97, 2011.
- M. Al-Khassaweneh and S. Aviyente, “The relationship between two directed information measures,” IEEE Signal Processing Letters, vol. 15, pp. 801–804, 2008.
- H. Marko, “The bidirectional communication theory—a generalization of information theory,” IEEE Transactions on Communications, vol. 21, no. 12, pp. 1345–1351, 1973.
- G. Kramer, “Capacity results for the discrete memoryless network,” IEEE Transactions on Information Theory, vol. 49, no. 1, pp. 4–21, 2003.
- S. Tatikonda and S. Mitter, “The capacity of channels with feedback,” IEEE Transactions on Information Theory, vol. 55, no. 1, pp. 323–349, 2009.
- P. O. Amblard and O. J. J. Michel, “On directed information theory and Granger causality graphs,” Journal of Computational Neuroscience, vol. 30, no. 1, pp. 7–16, 2011.
- A. Rao, A. O. Hero III, D. J. States, and J. D. Engel, “Using directed information to build biologically relevant influence,” in Proceedings of the Computational Systems Bioinformatics, pp. 145–156, 2007.
- C. J. Quinn, T. P. Coleman, N. Kiyavash, and N. G. Hatsopoulos, “Estimating the directed information to infer causal relationships in ensemble neural spike train recordings,” Journal of Computational Neuroscience, vol. 30, no. 1, pp. 17–44, 2011.
- Y. Liu and S. Aviyente, “Information theoretic approach to quantify causal neural interactions from EEG,” in Proceedings of the 44th IEEE Asilomar Conference on Signals, Systems and Computers (ASILOMAR '10), pp. 1380–1384, November 2010.
- P. Mathai, N. C. Martins, and B. Shapiro, “On the detection of gene network interconnections using directed mutual information,” in Proceedings of the Information Theory and Applications Workshop (ITA '07), pp. 274–283, February 2007.
- T. M. Cover, J. A. Thomas, and J. Wiley, Elements of Information Theory, Wiley Online Library, 1991.
- J. Geweke, “Measurement of linear dependence and feedback between multiple time series,” Journal of the American Statistical Association, vol. 77, no. 378, pp. 304–313, 1982.
- E. G. Miller, “A new class of entropy estimators for multidimensional densities,” in Proceedings of the IEEE International Conference on Accoustics, Speech, and Signal Processing, vol. 3, pp. 297–300, April 2003.
- G. A. Darbellay and I. Vajda, “Estimation of the information by an adaptive partitioning of the observation space,” IEEE Transactions on Information Theory, vol. 45, no. 4, pp. 1315–1321, 1999.
- Y. Liu and S. Aviyente, “Time-lagged directed information,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '11), pp. 3864–3867, 2011.
- W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vettering, Numerical Recipes: The Art of Scientific Computing, Cambridge University Press, 3rd edition, 2007.
- B. Schelter, M. Winterhalder, and J. Timmer, Handbook of Time Series Analysis: Recent Theoretical Developments and Applications, VCH Verlagsgesellschaft mbH, 2006.
- L. Cao, “Practical method for determining the minimum embedding dimension of a scalar time series,” Physica D, vol. 110, no. 1-2, pp. 43–50, 1997.
- G. Pipa and S. Grün, “Non-parametric significance estimation of joint-spike events by shuffling and resampling,” Neurocomputing, vol. 52, pp. 31–37, 2003.
- A. K. Seth, “A MATLAB toolbox for Granger causal connectivity analysis,” Journal of Neuroscience Methods, vol. 186, no. 2, pp. 262–273, 2010.
- G. Nolte, A. Ziehe, V. V. Nikulin et al., “Robustly estimating the flow direction of information in complex physical systems,” Physical Review Letters, vol. 100, no. 23, Article ID 234101, 2008.
- M. Breakspear, “Nonlinear phase desynchronization in human electroencephalographic data,” Human Brain Mapping, vol. 15, no. 3, pp. 175–198, 2002.
- W. Michiels and H. Nijmeijer, “Synchronization of delay-coupled nonlinear oscillators: an approach based on the stability analysis of synchronized equilibria,” Chaos, vol. 19, no. 3, Article ID 033110, 2009.
- J. C. Butcher and J. Wiley, Numerical Methods for Ordinary Differential Equations, vol. 2, Wiley Online Library, 2003.
- J. S. Moser, H. S. Schroder, C. Heeter, T. P. Moran, and Y.-H. Lee, “Mind your errors: evidence for a neural mechanism linking growth mindset to adaptive post-error adjustments,” Psychological Science, vol. 22, no. 12, pp. 1484–1489, 2011.
- J. S. Moser, T. Moran, and A. Jendrusina, “Parsing relationships between dimensions of anxiety and action monitoring brain potentials in female undergraduates,” Psychophysiology, vol. 49, no. 1, pp. 3–10, 2012.
- B. A. Eriksen and C. W. Eriksen, “Effects of noise letters upon the identification of a target letter in a nonsearch task,” Perception and Psychophysics, vol. 16, no. 1, pp. 143–149, 1974.
- D. M. Olvet and G. Hajcak, “The stability of error-related brain activity with increasing trials,” Psychophysiology, vol. 46, no. 5, pp. 957–961, 2009.
- J. Kayser and C. E. Tenke, “Principal components analysis of Laplacian waveforms as a generic method for identifying ERP generator patterns: I. Evaluation with auditory oddball tasks,” Clinical Neurophysiology, vol. 117, no. 2, pp. 348–368, 2006.
- S. Aviyente, E. M. Bernat, W. S. Evans, and S. R. Sponheim, “A phase synchrony measure for quantifying dynamic functional integration in the brain,” Human Brain Mapping, vol. 32, no. 1, pp. 80–93, 2011.
- J. F. Cavanagh, M. X. Cohen, and J. J. B. Allen, “Prelude to and resolution of an error: EEG phase synchrony reveals cognitive control dynamics during action monitoring,” Journal of Neuroscience, vol. 29, no. 1, pp. 98–105, 2009.
- C. R. Genovese, N. A. Lazar, and T. Nichols, “Thresholding of statistical maps in functional neuroimaging using the false discovery rate,” NeuroImage, vol. 15, no. 4, pp. 870–878, 2002.
- J. Lizier, The local information dynamics of distributed computation in complex systems, Ph.D. dissertation, University of Sydney, 2010.
- L. Zhao, H. Permuter, Y. H. Kim, and T. Weissman, “Universal estimation of directed information,” in Proceedings of the IEEE International Symposium on Information Theory (ISIT '10), pp. 1433–1437, June 2010.
- L. Faes, G. Nollo, and A. Porta, “Information-based detection of nonlinear Granger causality in multivariate processes via a nonuniform embedding technique,” Physical Review E, vol. 83, no. 5, Article ID 051112, 2011.
- W. Zhao, E. Serpedin, and E. R. Dougherty, “Inferring connectivity of genetic regulatory networks using information-theoretic criteria,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 5, no. 2, pp. 262–274, 2008.