Table of Contents Author Guidelines Submit a Manuscript
Computational Intelligence and Neuroscience
Volume 2016, Article ID 2961727, 15 pages
http://dx.doi.org/10.1155/2016/2961727
Research Article

Analysis of Residual Dependencies of Independent Components Extracted from fMRI Data

1Dipartimento di Ingegneria dell’Informazione, University of Pisa, 56122 Pisa, Italy
2Laboratory of Clinical Biochemistry, Department of Experimental Pathology, University of Pisa Medical School, 56126 Pisa, Italy
3Fondazione Toscana Gabriele Monasterio, 56124 Pisa, Italy

Received 2 September 2015; Accepted 22 November 2015

Academic Editor: Ricardo Aler

Copyright © 2016 N. Vanello et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Independent component analysis (ICA) of functional magnetic resonance imaging (fMRI) data can be employed as an exploratory method. The lack in the ICA model of strong a priori assumptions about the signal or about the noise leads to difficult interpretations of the results. Moreover, the statistical independence of the components is only approximated. Residual dependencies among the components can reveal informative structure in the data. A major problem is related to model order selection, that is, the number of components to be extracted. Specifically, overestimation may lead to component splitting. In this work, a method based on hierarchical clustering of ICA applied to fMRI datasets is investigated. The clustering algorithm uses a metric based on the mutual information between the ICs. To estimate the similarity measure, a histogram-based technique and one based on kernel density estimation are tested on simulated datasets. Simulations results indicate that the method could be used to cluster components related to the same task and resulting from a splitting process occurring at different model orders. Different performances of the similarity measures were found and discussed. Preliminary results on real data are reported and show that the method can group task related and transiently task related components.

1. Introduction

Functional magnetic resonance imaging (fMRI) is a widespread and well-established technique for the in vivo functional exploration of the brain. fMRI data analysis methods can be roughly classified as confirmatory or hypothesis-driven methods and exploratory or data-driven methods [13]. The former are used in order to test the validity of the experimenters’ hypotheses but do not allow the detection of unexpected phenomena, that is, effects that are not modelled a priori. On the other hand, data-driven methods provide results that are based on general assumptions about signal generation but are often difficult to be fully interpreted [4].

Independent component analysis (ICA) is one of the most used exploratory methods both for task-associated neural responses and for resting state signal processing and is based on the assumption of statistical independence of the components to be extracted [4]. This method has proven its capabilities of separating physiological components of different origins (e.g., vascular and respiratory), of detecting unexpected phenomena, such as activations transiently time-locked with the stimulus, and of isolating artefact-related signal changes, such as those due to head movements [5, 6].

Nonetheless, ICA approaches present different drawbacks. In fact, the extracted components are difficult to classify since they do not have an explicit order or relationship among themselves [7]. The stochastic nature of the algorithm, the finite number of the observations, and the presence of noise influence the solution reliability and stability. For this reason several approaches have been proposed to face this issue [810]. In [8] the reliability of the solution is investigated by clustering the results obtained with multiple runs. A different approach was proposed for solving the same issue in [9] where an alignment of different maps obtained from multiple runs of the algorithm is obtained using information criteria and then the reliability of the components is estimated looking at the consistency and variability with tests. In [10], the independent components are classified according to different measures, as maps dipolarity, which is thought to be related to physiological plausibility of the component, and a consistency measure obtained through the use of surrogate decomposition.

Another issue affects the estimation of ICA model. In fact, the best number of components to be extracted, that is, the model order, is not known a priori. Underestimating the model order may cause an information loss, while overestimating it may produce spurious results or the splitting of interesting components into more components [7, 11]. In [11], the effects of varying model order are investigated in simulated datasets. Noise and signal intensity values as well as dataset dimensions were highlighted as parameters largely influencing this choice. Moreover, the authors suggested that adopting a high model order might lead to different conclusions in the case of group analysis with respect to single subject analysis. Specifically, in the case of intersubject variability of the maps, increasing model order might lead to results difficult to be interpreted, while in single subject analysis a higher model order was seen as a possibility to distinguish subnetworks characterized by a different temporal behavior. In [12], the effects of increasing model orders in a probabilistic group ICA are highlighted. While the stability of ICA decomposition was found to decrease with increasing model order, the use of higher model orders allowed detection of interesting neuroanatomical and functional components. To overcome order indeterminacy problem, various methods have been proposed [1318]. Specifically, information theoretic criteria, such as Akaike Information Criterion, Minimum Description Length, and Bayesian Information Criterion, have been applied to solve this issue in fMRI data [1316]. In [13], the classical information theoretic criteria were adapted to account for temporal and spatial correlations in fMRI data using a subsampling technique to obtain independent and identical distributed samples. In [16], the model order selection was obtained by introducing a noisy ICA model within a Bayesian framework [17]. In [18], bootstrap stability technique as applied to PCA reduction step is proposed. At this time, however, there is no general agreement about the best approach to choose ICA model order.

Apart from stability and model order indeterminacy, the interpretation of ICA results must take into account the fact that the statistical independence among the estimated ICs is only approximated by available algorithms. In fact, the presence of noise and the finite number of observations for each measurement do not allow accurate estimation of the higher order statistics [19] used to search for statistical independence. Thus, residual dependence between the extracted components can still be found and can be used to reveal some structure in the dataset. A topographic approach was suggested for ICA [20], where the ICA model was modified to take into account a topographic order among the extracted components: the distance between two elements in the topographic maps is related to the residual dependency. In [21], a comparison of ICA approaches that exploit clustering techniques to estimate the model is performed. Specifically, the topographic approaches are compared with standard ones, as well as with an approach defining a tree-like dependency structure among the components [22]. This work demonstrated that the information inherent in the dependencies among the components can separate artifacts and task related components as well as detecting interesting relationships among components. In a work by Ma et al. [23], the residual dependencies among components were estimated using a similarity measure based on Mutual Information (MI) [24] between the components estimated using a kernel density estimation (KDE) approach [25]. Afterwards, the components are classified using a hierarchical clustering technique [26]. This latter work showed that this approach could group interesting components, as applied to group fMRI data acquired during rest, across different algorithms and model orders. Specifically, it is claimed that the higher dimensionality of clustered components at higher orders is related to the merging of split components.

In this work, we propose an approach similar to the one developed by Ma et al. [23] which differentiates for a number of methodological aspects. Firstly, we explored two different methods for MI estimation, specifically a histogram-based technique and the KDE approach. Moreover, we adopted a different MI-based measure that is a metric in a mathematical sense and a different criterion for the clustering of the algorithm results. Finally, in [23], the robustness of the approach against model order indeterminacy was deduced from the analysis results of real resting state fMRI data. Differently, here we performed an analysis on the ICs estimated from simulated dataset with increasing model orders. By the analysis of the results from simulated and real dataset, we investigated the possibility of improving the grouping of the components originated from a splitting process due to model order overestimation.

2. Theory and Methodology

Datasets from fMRI are formed by time sequences of images or volumes, acquired while a subject is performing some sensory/motor or cognitive task at rest. In this work we focused on single subject fMRI data acquired during the execution of a block design task, alternating in time two different conditions. Synthetic data were simulated accordingly.

2.1. Simulated Data

A synthetic brain dataset was used for our preliminary studies of the clustering method [27]. Signal increase in response to neural activation was simulated convolving the time course describing the task with a typical hemodynamic response function: we chose the three-parameter gamma variate function defined as [28], where is a constant proportional to the amplitude of the signal increase. The task description was modeled as a square wave alternating in time: 15 seconds ON with 15 seconds OFF conditions. The total time length of each simulated dataset was three minutes, for a total number of scans equal to 60 (simulated TR = 3 s). The activated regions were created modulating the baseline intensities of a group of voxels, selected using a mask, with the time series previously described. Activated regions are supposed to be limited in a region of connected voxels and were created using a mask formed by region smoothed with a Gaussian kernel to simulate the spatial point spread function due to the vasculature. The Gaussian kernel was a bidimensional one with 3 mm full width half maximum (FWHM) parameter. All the simulations that we present in this paper were obtained by defining two nonoverlapping activated regions of about 2 centimeters of diameter. In the following, these activated regions will be indicated as region of interest (ROI) number 1 and number 2. The signal change was about 2 percent as compared to the baseline level in the center of the activated regions. Different time delays between the activation time courses of the two regions were simulated: 1.25, 2.5, and 5 seconds.

The noise in the images was taken to be i.i.d. Gaussian distributed with zero mean and variance . For each delay value, different noise levels were simulated. The simulated noise standard deviation was 0.33%, 0.66%, 1%, and 1.33% of the mean image value at the baseline level. The contrast to noise ratio, defined as , where is the signal change following an activation, was equal to, approximately, 6, 3, 2, and 1.5 (resp.): this was the maximum value at the center of the activated regions.

2.2. Real Data

Brain activity was measured in a 25-year-old right-handed healthy female. The subject signed a written informed consent prior to the enrolment into the study under a protocol approved by the Ethics Committee of the University of Pisa, Italy. The scanner used was a 1.5 Tesla, GE Signa Cv/i. An anatomical image was acquired with a 3D SPGR sequence. The functional scans were gradient echo-EPI with TR = 3 sec, TE = 40 msec, and FA = 90 degrees with bandwidth of 62.5 kHz. Twenty axial slices, covering the entire brain, were acquired with slice thickness of 5 mm, 240 mm FOV, and in plane 64 × 64 spatial matrix, and the number of time scans was 63, for a total acquisition time of 189 s. The subject head was restrained with foam in order to minimize head movements. The subject performed a simple finger tapping sequence with her right hand fingers: the task was a block design paradigm alternating 15 sec ON and 15 seconds OFF conditions. The number of scans was 60 for a total run length of three minutes. The images time series were interpolated to correct for slice timing effects and volume registered to a reference scan to reduce movement related effects. The images were then spatially transformed to the Talairach-Tournoux Atlas reference system [29]. These preprocessing steps were performed using AFNI [30].

2.3. ICA Model

In spatial ICA we can write the observed data as , where is the th image or volume in the sequence and is a spatial index for each volume element (voxel). The observed data can be written as a linear mixing of spatial ICs : where . Both and the mixing matrix are unknown. In this model and are seen as random variables and is an index for the observations of each random variable.

The ICA problem consists in finding an unmixing matrix so that the estimated independent components can be written as . Under the hypothesis that the mixing matrix is invertible, , each estimated component can be written as a linear combination of the observed variables , with the th column of and . Each can be seen as a spatial map, individuating a value for every voxel. The th spatial fixed map is time modulated by the corresponding time course, given by the th column of . The components, whose associated time courses highly correlate with the paradigm, are considered as task related components or consistently task related (CTR). The components, whose activation is related only partially with the paradigm, are called transiently task related (TTR) components [5]. Several approaches to estimate the ICA have been described (see [6] and references therein) and for the reasons highlighted in the introduction they all lead to components that are only approximately statistically independent. One way to estimate the ICs is a method based on the maximization of the non-Gaussianity of the [31].

A fast fixed point algorithm [32] can be used to find the weights such that non-Gaussianity of is maximized. The fastICA algorithm can exploit different nonlinear functions to approximate negentropy whose maximization leads to non-Gaussianity maximization.

In this work the ICA decomposition was obtained using the fastICA algorithm with tanh as nonlinearity [33]. Different model orders were applied, specifically 5, 10, 15, and 20. After the extraction step, each independent map was transformed into map statistics to find the voxels contributing significantly to the corresponding component. Given an independent map , the map can be computed as , where is the mean of the values of and are their standard deviation.

2.4. Proposed Method

We hypothesized that the components resulting from a splitting process due to a model order overestimation show a higher residual dependence with respect to other components. Under this hypothesis, the residual dependencies among the components could be explored to detect split components. We proposed to estimate the residual dependencies using pairwise distance measure between two components, and , based upon the definition of mutual information. A hierarchical clustering approach was then employed to classify and visualize the similarities, in terms of distance measure, between the extracted ICs. The clustering results can be visualized by a dendrogram that highlights the merging of the components due to the similarity criterion.

2.4.1. Similarity Measure

The similarity measure between two components and is defined as follows:where is the joint entropy and is the mutual information between two sources. The choice to use rather than is based on the fact that the latter is not a distance metric in the mathematical sense [34]. In this work histogram-based technique and a kernel density estimation approach to compute were compared.

(1) Histogram-Based Technique. The probability that a variable value, that is, independent component, lies in the th interval can be estimated as the frequency of occurrence, so that we can write , where is the total number of observations for and is the number of times belongs to the th interval. This probability in the following will be referred to as .

The probability that the variable lies in the th interval while the variable lies in the th interval is given by , where is the number of times the couple belongs to the bidimensional bin . This quantity can be written as .

We estimated the joint entropy as while the Mutual Information was computed as where and are the number of states, or bins, for variable and , respectively.

The number of bins was chosen as [35], where is the number of observations, that is, the voxels in the image after masking out of brain voxels. A rank ordering operation was used before performing the histogram-based technique; that is, the data values were replaced by their ranks before the histogram is built. This operation facilitates the computation of mutual information as the distributions of the components are transformed into uniform distributions. Moreover, since we were interested in similarities between two distributions, this approach allows the results not to be strongly affected by the marginal distributions of the components. The correlations between two components are preserved and hence the validity of the results.

(2) Kernel Density Estimation. Kernel density estimation (KDE) can be used to estimate mutual information [25]. This method consists in estimating the probability density function (pdf) of a variable with a linear combination of kernel functions such thatwhere is the th value of the variable and is a probability density function itself.

A Gaussian kernel can be used to write the pdf asThe parameter is a smoothing parameter, called bandwidth: small values lead to taking into account finer details that can be not interesting or originate from noise; larger values can hide interesting features in the distributions. For Gaussian distributed variables, with variance , the values that minimize the mean square error in estimating the density have been found to be [36].

In the bidimensional case, the bidimensional kernel function is given bywith . The bandwidth is related to the underlying distribution as in the monodimensional case. For Gaussian variables an optimal value has been found to be with and is the average marginal standard deviation [37].

Given the probability density functions , , and the mutual information and the entropy between two variables and can be estimated asThese quantities can be estimated with standard procedures for numerical integration. We adopted adaptive Simpson quadrature approach.

2.4.2. Clustering

The hierarchical clustering approach is based on the Ward method [38]. This method consists in merging every possible cluster pair and choosing the one which minimizes the information loss. To estimate this quantity, the error sum-of-squares (ESS) is used: where is the mean of cluster and are the data points. The distance between two clusters is defined byAt each stage, the Ward method merges two groups such that their will be minimized. The merging of the components can be visualized with a dendrogram: the abscissa of the dendrogram indicates the components, while the ordinate level, at which two components or groups of components are merged, is related to the change in the error sum-of-squares after joining groups.

2.5. Method Validation

The performance of the proposed method can be evaluated on simulated datasets by verifying that the dendrogram correctly merges the components that were split by the ICA algorithm. To identify the components maps that were related to the simulated activations, receiver operating characteristics (ROC) curves [39] were used and applied to the obtained ICA maps. ROC curves are plots of true positive detection fraction against false positive detection fraction obtained varying the threshold level of the maps. In the case of simulated dataset, the true activated areas are known and serve as ground truth measure, so that it is possible to estimate true positives by looking at the activated voxels, that is, over threshold, in one independent component and by checking whether they correspond to the known activated regions. An area under curve (AUC) estimated from the ROC curve ranging from 0.7 to 0.8 is considered a result showing a fair accuracy of the test performed, whereas an AUC between 0.6 and 0.7 is considered a poor accuracy index [40]. An independent component whose AUC is significant for both ROIs is a component that merges both the two activated regions.

Since we focused on fMRI data acquired during a task execution, the classification of the results can take into account the corresponding time course of each component. Specifically, interesting CTR and TTR components can be highlighted.

3. Results

3.1. Simulated Datasets

Results from the application of the ICA algorithm on simulated datasets are summarized in Tables 1, 2, and 3. Each table is related to the result obtained from the datasets with time delays between the activations time courses of 1.25, 2.5, and 5 seconds, respectively. For each table different noise levels results are summarized. The two activated regions are indicated by ROIs #1 and #2. In each column, the ICs whose maps result in an AUC greater than 0.6 are shown. Specifically the IC index is shown, in brackets, along with the corresponding AUC.

Table 1: Evaluation of ICA of the dataset with activations in ROI #1 and ROI #2 with time delay of 1.25 seconds at different noise levels and model orders. The AUC is shown near the IC index (in brackets).
Table 2: Evaluation of ICA of the dataset with activations in ROI #1 and ROI #2 with time delay of 2.5 seconds at different noise levels and model orders. The AUC is shown near the IC index (in brackets).
Table 3: Evaluation of ICA of the dataset with activations in ROI #1 and ROI #2 with time delay of 5 seconds at different noise levels and model orders. The AUC is shown near the IC index (in brackets).

In the case of the shortest time delay between the activations time courses, with noise standard deviation greater than 0.33%, the two ROIs are not distinguished by the ICA model (see Table 1): so the same IC index appears in both columns. The same results are obtained for a time delay of 2.5 seconds at higher noise levels (see Table 2). For time delay equal to 5 seconds, with the chosen noise levels, the ICA could always distinguish the two activated regions (see Table 3). The model order 5 in most cases allows the separation of each activation in single IC maps. In some cases, using this model order the component splitting occurs. For higher model orders, the splitting process occurs at all noise levels for time delay of 1.25 and 2.5 seconds, while for time delay of 5 s one of the two activated regions is often described by one IC.

In Figure 1(a), the IC maps related to the activated regions, obtained from a simulated dataset with added noise standard deviation and time delay between the two simulated activations equal to 2.5 s, are shown. The results were obtained for model order . The IC maps, in red-yellow, are transformed in z-scores maps, thresholded choosing || larger than or equal to 2, and superimposed on the anatomy. The two ROIs are described by different ICs and each ROI is described by more than one component. In Figure 1(b), four IC maps extracted from simulated datasets with noise level and time delay between simulated activation equal to 1.25 s are shown. In this case the ICA could not separate the two activated regions and the CTR ICs, obtained for model order , are related to both ROIs.

Figure 1: IC maps extracted from simulated datasets. The IC maps (in red-yellow) are transformed in -scores maps, thresholded, and superimposed on the anatomy. Only the ICs related to activated ROI #1 and ROI #2 are shown. (a) Noise level and time delay is 2.5 s; model order . (b) Model order , noise level , and time delay is 1.25 s. In this case, ICA could not separate the two activated regions in different components. Specifically, the activated regions are described by four IC maps. Each map described both ROI #1 and ROI #2. (c) The IC map number 4 estimated in the case of noise level and time delay of 1.25 seconds with model order = 15. To better show the intensity map distribution this IC is shown twice using two different thresholds applied to the -score statistics: the spatial distribution of the map is flatter with respect to the components related to the ROIs.

An IC that seems to represent the intensity distribution of all pixels in the images is estimated at the different model orders. The IC has a sub-Gaussian distribution and might originate from the small changes in the baseline mean level, after removing the average value from each image in the sequence (centering operation), as described in Section 2.3. This phenomenon depends on the fact that imposed activations in both ROIs affect the mean level of the image synchronously. As can be seen in Figure 1(c), where a component with the above-described characteristics is shown, the map has some complementary distribution with respect to the activated regions.

3.1.1. Histogram-Based Technique

The histogram-based technique was applied to ICs extracted from simulated datasets and the pairwise distances were estimated. Since they are representative, only the results for noise level equal to 0.66% and time delay equal to 2.5 seconds are shown in Figure 2. All other cases will be discussed. The hierarchical clustering algorithms successfully merged the components that were found related to the same activated regions. Specifically, in all cases that are listed in Tables 13, when the ICA algorithm succeeds in separating the two activated regions, that is, there are no ICs that are related to both ROIs, the clustering algorithm can merge the components that are related to the different ROIs. In Figure 3(a), the results for noise level equal to 1% and time delay of 1.25 seconds are shown: as it can be seen from Table 1, the two activation regions are not distinguished by the ICA model and IC components describing ROI #1 and ROI #2 are merged together. The corresponding IC maps are shown in Figure 1(b).

Figure 2: Dendrogram obtained from clustering the pairwise mutual distances between extracted ICs from simulated datasets in the case of noise level and time delay of 2.5 seconds. The algorithm for estimating the pairwise distances is the histogram-based approach with rank ordering. The results obtained with model order = 5 (a), 10 (b), 15 (c), and 20 (d) are shown.
Figure 3: Dendrograms obtained from clustering of the pairwise mutual distances, applied to distances estimated with histogram-based approach (a) and KDE approach (b) between extracted ICs in the case of noise level and time delay of 1.25 seconds with model order = 10. The relative IC maps are shown in Figure 1.

The ICs related to different activations were often grouped together for model order equal to 5. In all other cases, when ICA detects two separated areas, the two components are grouped in different homogeneous clusters. In Figure 2(a), the components describing the ROI #2 that were separated in two IC components were merged together and then merged again with ROI #1 component. In each case, the components corresponding to the same activation regions are merged together.

3.1.2. Kernel Density Estimation

The analyses described in the previous subsubsection were repeated using the kernel density estimation based method. The results are similar to those obtained for the previous histogram-based approach. Specifically, when the ICA algorithm succeeds in separating the two activated regions the KDE approach can merge the corresponding components in different clusters. As for the histogram-based technique, the components that describe both ROIs simultaneously, occurring when the IC composition could not separate them, are merged together. At the same time, low model orders can result in a merging of the activated regions, albeit found separated by the ICA algorithm. The results pertaining to noise level equal to 0.66% and time delay equal to 2.5 seconds are shown in Figure 4.

Figure 4: Dendrogram obtained from clustering of the pairwise mutual distances between extracted ICs from simulated datasets in the case of noise level and time delay of 2.5 seconds. The algorithm for estimating the pairwise distances is the kernel-based one. The results obtained with model order = 5 (a), 10 (b), 15 (c), and 20 (d) are shown.
3.1.3. Comparison of Histogram-Based and Kernel Density Estimation Approaches

Some differences in the results were found by the two approaches. One difference pertains the ordinates in the dendrograms. The ordinates are related to the increase in within group error sum-of-squares after joining two clusters. In the case of dendrograms corresponding to KDE method the heights of the clusters, that is, the Ward distances, are lower than those obtained from histogram-based technique, meaning a smaller increase in the within-cluster sums of squares after joining the groups in the former case. This fact shows that the distances estimated from the KDE method are on average more homogeneous.

On the other side, the KDE approach was more prone than the histogram-based approach to the clustering of the sub-Gaussian component described in Section 3.1 with a component describing one of the activated regions. This phenomenon can be seen in Figures 4(c) and 4(a). The component #4 is the one shown in Figure 1(c). As a comparison, in Figures 2(c) and 2(a), the results of the histogram-based approach applied to the same dataset are shown.

3.2. Real Datasets

In the case of real data, it was not possible to estimate an AUC performance parameter, since we did not know a priori the spatial distribution of the activated regions. It was possible to detect interesting components by looking at the correlation coefficient between the associated time course for each map and the expected hemodynamic time course, given the experimental paradigm. Another criterion that could be used to classify the components in real datasets was related to previous knowledge about the regions involved for the specific task [41]. Moreover, a comparison with the results obtained with a hypothesis driven analysis technique, such as the general linear model (GLM), [1, 42] can be used to detect the regions that are consistently task related and should be included in ICA results.

3.2.1. GLM Results

The results of the GLM applied to the real dataset will be shown. The expected hemodynamic response, used as a regressor of interest, was estimated as the boxcar time series describing the paradigm, convolved with hemodynamic response function [28]. The preprocessing steps for linear analysis are the same as those for ICA. The errors are considered to be normally distributed. The best linear unbiased estimator (BLUE) [43] for the parameter of interest is used, and the null hypothesis of zero activation was tested with test. The results of the GLM applied to real data are shown in Figure 5(e). A statistic map estimated for the coefficient of the regressor of interest of an axial slice at  mm in Talairach-Tournoux Atlas [29] coordinates is superimposed on a T1 weighted anatomical image. The dataset we took into account has an approximated CNR equal to 3, as estimated on the selected slice. The threshold was chosen at , corresponding to an uncorrected value lower than 0.01. Significant activations are found in the ipsilateral and contralateral primary motor cortex and in the supplementary motor area (SMA). Right and left precuneus and posterior parietal areas were additionally recruited.

Figure 5: (a)–(d) The IC maps superimposed upon an anatomical T1 weighted image and thresholded with for different model orders are shown. The orientation is RAI. A color bar for the correspondence between color and value is shown. (a) Consistently task related (CTR, ) IC for real dataset with model order equal to 5. (b) ICs maps extracted with model order equal to 10. Component #2 (TTR, ) pertains the SMA and Brodmann area 5 activity. The component number 7 (CTR, ) is related to right and left precuneus, in the primary sensory and motor areas while the component #9 (CTR, ) is related to contralateral primary motor area (the orientation is RAI), SMA, and right precuneus. (c) Resulting ICs from real dataset with model order equal to 15. The component #10 (CTR, ) is mainly related to the contralateral motor area. The activity in the SMA and Brodmann area 5 is mainly described by component #12 (TTR, ). The component #15 (CTR, ) describes the ipsilateral and contralateral sensory areas and right precuneus. (d) CTR ICs for real dataset with model order equal to 20. The component #7 (TTR, ) is related to SMA activity and Brodmann area 5. The component #12 (CTR, ) is mainly related to the contralateral motor area, SMA, and Brodmann area 5 as well. The activity in the precuneus is mainly described by component #13 (CTR, ). The activity of contralateral and ipsilateral sensory areas is described by map #14 (CTR, ). (e) GLM analysis results of the real fMRI dataset. The value of test about the parameter of interest is shown. A color bar related to values is shown.
3.2.2. ICA Results

The ICA model was applied to this real dataset with a model order equal to 5, 10, 15, and 20. With a model order equal to 5, one component was detected to be consistently task related (). In Figure 5(a), the corresponding map (IC #1), thresholded at , is shown to be superimposed at an anatomical T1 weighted mask. All the functional areas that were found to be activated using the GLM approach (Figure 5(e)) are found to be described by this IC map. Moreover, a small activation in left superior parietal lobule is described.

Using model order equal to 10, two consistently task related components and a TTR component were found by the ICA. The results are shown in Figure 5(b): the activity in the SMA is depicted mainly by IC #2 (TTR, ). This component shows activation also in Brodmann area 5, not previously found. The component IC #7 (CTR, ) is related to the activity in the precuneus and in the primary sensory and motor areas, while IC #9 (CTR, ) shows activation in the contralateral primary motor area and minor activations in the SMA, Brodmann area 5, and right precuneus. The regions are those found by the ICA with model order = 5 but these seem to be split into several components with model order = 10. On the other side, model order equal to 5 did not detect the activity in Brodmann area 5.

The IC analysis results for model order equal to 15 are shown in Figure 5(c). Two consistently task related areas and one transiently task related area were found. Specifically, component number 15 (CTR, ) describes the ipsilateral and contralateral sensory areas and right precuneus and component number 10 (CTR, ) that is mainly related to the contralateral motor area. The activity in the SMA and Brodmann area 5 is mainly described by a transiently task related component number 12 (TTR, ) shown in the center image.

The IC analysis results for model order equal to 20 are shown in Figure 5(d). Four CTR maps are found: the activity in the right and left precuneus is well described by CTR component number 13 (CTR, ). Activity in SMA is described by CTR component number 12 (CTR, ) and transiently task related component number 7 (TTR, ). While activity in ipsilateral and contralateral motor and sensory areas is described by maps 12 and 14 (CTR, ).

3.2.3. Clustering with Histogram-Based Technique

Figure 6(a) depicts the hierarchical clustering algorithm results, applied to distances estimated using the histogram-based technique. In graph (A), the clustering results for model order equal to 5 are shown: the clustering operation in this case was not easy to interpret since only one CTR component was found by ICA (IC #1). For model order equal to 10 (Figure 6(a), graph (B)) the interesting components (ICs #2, #7, and #9) are merged together by the hierarchical clustering algorithm before any other components are merged with these ICs.

Figure 6: Dendrogram obtained from clustering of the pairwise mutual distances between extracted ICs from real data. The distances were estimated through a histogram-based technique after rank ordering (a) and KDE technique (b). The clustering operation was repeated for different IC model orders: (A) 5, (B) 10, (C) 15, and (D) 20. The interesting CTR components are described in Section 3.2.2. See text for a discussion.

Results for model order equal to 15 are shown in graph (C): components number 12 and 15 are merged together. The CTR component number 10 is merged with components number 3 and 6. While component 3 seems to be related to activity in SMA, Brodmann area 5, and right precuneus, component number 6 seems to be related to vascular related signal changes (see Figure 7(a)).

Figure 7: (a) Component maps #3 and #6 obtained from real data, with model order = 15. (b) IC maps extracted from real fMRI data with ICA and model order equal to 20. These maps were merged with CTR components by the clustering algorithm with kernel density estimation. On the left component #3 is shown, in the center component #8 is shown, and on the right component #15 is shown.

Results for model order equal to 20 are shown in graph (D) of Figure 6(a): only components number 12 and 13 are merged together. It was not possible to identify significant clusters with this model order.

3.2.4. Clustering with KDE Technique

The results of clustering operation obtained on the distances estimated with kernel density approach are shown in Figure 6(b). Graph (A) describes the dendrogram after ICA with model order equal to 5: as for histogram-based technique (see previous subsubsection) there is only one CTR component and no significant clustering of this component is evident. In graph (B), the results for model order equal to 10 are shown: the task related components, #2, #7, and #9, are merged together by the algorithm. The results for model order equal to 15 are shown in graph (C): components #10, #12, and #15 are merged together. Component number 10 is merged with component number 3 whose spatial distribution is shown in Figure 7(a).

In graph (D) of Figure 6(b), the clustering of components after ICA with model order equal to 20 is shown. A cluster of components number #7, #12, and #14 was found. At higher levels components #3, #8, and #15 (Figure 7(b)) and the CTR component #13 were merged. The component number 3 depicts activity in Brodmann area 5 and right precuneus. Component number 15 shows significant activation in right precuneus and left precentral gyrus. Component number 8 shows activation not previously detected in the right middle frontal gyrus and minor activations in Brodmann area 5.

4. Discussion

The method proposed here aimed at studying the residual dependencies between the ICs to reveal some informative structure in brain functional data acquired during fMRI protocol. In this work, our aim was twofold: to verify whether this approach could identify and merge the ICs derived from a splitting process caused by an overestimation of ICA model order and to compare two different strategies for the estimation of mutual information-based metric for the analysis of ICs dependencies. Previous work [23] demonstrated that this method can group anatomically and functionally related components from real fMRI datasets acquired during rest. This behavior was found consistently across different model orders. Here, using simulated task related fMRI activations we specifically addressed the issue related to the merging of split components at different model orders with different time relationships between the activated regions and at different contrast to noise ratios.

Moreover, in [23] the mutual information measure adopted to explore the residual dependencies was estimated using the KDE approach. Here, we compared the performances of the latter approach and a histogram-based approach. The histogram-based technique is the classical way to estimate entropy and mutual information [44] and requires that the estimated sources values be partitioned into bins or intervals. An adaptive partitioning operation is used so that the results are not strongly affected by the individual marginal distributions of the components. To achieve this, a rank ordering operation of the elements of each IC is applied [45]. The second method for estimating MI is based on kernel density estimation (KDE) [25]. The KDE method shows a better convergence rate of the MI estimate to the underlying real density than the histogram-based technique and is not sensitive to the choice of the origin. On the other hand, the KDE results are dependent on the choice of kernel functions parameters that are not known a priori.

The ICA of the simulated datasets presented in this paper clearly demonstrated that model order inaccuracy may cause splitting of the same activation region into more components. In this case, the proposed clustering approach highlights the components that have been split. In fact, the clustering algorithm succeeded in grouping together components belonging to the same region of activation found by ICA. Specifically, those components are found to merge earlier in the dendrogram with respect to other components. To classify the independent components that were related to the task we adopted the ROC curves. Specifically, the area under the ROC curve (AUC) was introduced as an index to determine the ROIs to which each component is related. The low values of the AUC parameter were due to the fact that the same activated region is decomposed in different ICs. In some cases the ICA algorithm fails in separating the components. This phenomenon depends upon the time delay between the activation time series and upon the contrast to noise ratio. Even in the cases when the ICA cannot separate the components, as it can be found from dendrogram analysis, the clustering operation correctly associates these components to the same cluster. However, at this time, we cannot claim that this approach could offer the possibility of choosing the right model order of ICA analysis. Nonetheless, exploring the earlier merging of some components may represent the way to reveal brain areas that are related to the same phenomenon of interest. The clustering algorithm with both methods for estimating the MI was found to merge the ICs related to different areas for model order equal to 5. For higher model orders this was never observed. This observation might confirm the need of using higher model order with respect to lower ones [46].

The performances of the hierarchical clustering on simulated datasets, by estimating the pairwise distances with the two approaches, the histogram-based technique and the kernel density estimation one, are different. One difference is related to the homogeneity of the clusters that can be seen looking at the ordinates of the clusters’ height. Specifically, in the KDE method, the clusters are more homogeneous with respect to those found by the histogram-based technique. This can be seen by looking at the heights of the clusters that were lower for the former approach. However, with regard to simulated datasets, the KDE approach was more prone than the histogram-based approach to merge a specific sub-Gaussian IC to one of the ICs related to the activated regions. This component may be a result of the mathematical process used to synthesize our data which caused small changes in the baseline mean level of the image voxels. Thus this component might arise as a diffuse activation when the variables centering operation needed for ICA is performed. Nonetheless, the KDE based approach showed the merging of this sub-Gaussian component more frequently than the histogram-based approach, even at higher model orders. Interestingly in [11, 12] larger diffuse, low kurtosis components were found to merge to other components at low model orders, while they were separated at higher orders. This could be further investigated to highlight whether this issue of the KDE approach is relevant when ICA is applied to real fMRI data and whether the choice of a different criterion for the kernel size could alleviate it.

In the case of the real dataset here studied, both methods seem to provide good results for model order equal to ten. For higher order models, the kernel density estimation based method gives better results, merging interesting components both for model order equal to 15 and for model order equal to 20, while the histogram-based technique does not always group interesting components. The ICA could identify activation in areas not found by the GLM as the activity in Brodmann area 5 and left superior parietal lobule. It is interesting to note that Brodmann area 5 was found by ICA with model orders greater than 5: in all cases the hierarchical clustering stage could merge this TTR component with the other CTR.

In this work we did not take into account the stochastic nature of the ICA algorithm and of the noise that could lead to different results on different runs. Nonetheless, consistent results were here obtained across the different 48 simulated datasets obtained by varying the above-mentioned parameters. To face this issue, the results obtained after different executions of the ICA algorithm for several times on the same combinations of noise, time delays, and model order could be done.

Other methodological differences distinguish our work from [23]. In fact, from the methodological point of view, the MI-based metric and the clustering procedures adopted here were different. Future work could evaluate how such different choices influence the final results.

One important issue that was not studied here is the choice of threshold or the dendrogram level used to determine the clusters of interest: while this value could be individuated for simulated datasets, a systematic approach while changing the data volume dimension, the size and distributions of the activation, and the noise level is missing for real data. Moreover, a criterion for the choice of the threshold would require a study with a larger number of participants as well as tasks. Other approaches could be used. In [23], the iterative use of statistical tests was applied to the mean between-clusters and within-cluster distances as a stopping procedure for the dendrogram partitioning.

In this work, we have chosen to apply this method to results obtained from the fastICA algorithm that was proven to work efficiently within the context of fMRI data analysis [47]. However, the proposed method can be applied to results from other ICA algorithms, since it is completely independent from the algorithm employed. With regard to this point, the approach proposed here differs from the one described in [21], in which the ICA algorithms were modified to take into account a topographical order or a dependence structure among the components. Moreover, our goal was not to test the reliability or the stability of the ICA algorithms as in [810]. Specifically, the proposed approach could be applied to the maps resulting from the above-cited methods, with the aim of acquiring different information about the relationship among the components. As an example, it could be applied to the centrotypes of the clusters detected by the ICASSO method described in [8].

Another future development of the study presented here could be the analysis and the comparison of the performances of the histogram and KDE based approaches on components extracted at group level, where the intersubject variability represents a relevant issue [11]. Moreover, the evaluation of the proposed approach on simulated data is focused on a specific scenario characterized by two nonoverlapping areas with similar albeit delayed activation time series. Since the fMRI studies have reached a high level of complexity and often are focused on brain activity during rest, different tests should be further developed in the future.

Although limited, the test results presented here indicate that this approach could give fruitful information regarding the statistical similarity of the obtained independent components and cope with the issue of component splitting.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

The authors would like to thank R. W. Cox for useful criticism and suggestions.

References

  1. K. J. Friston, A. P. Holmes, K. J. Worsley, J.-P. Poline, C. D. Frith, and R. S. J. Frackowiak, “Statistical parametric maps in functional imaging: a general linear approach,” Human Brain Mapping, vol. 2, no. 4, pp. 189–210, 1994. View at Publisher · View at Google Scholar · View at Scopus
  2. K. J. Friston, W. Penny, C. Phillips, S. Kiebel, G. Hinton, and J. Ashburner, “Classical and Bayesian inference in neuroimaging: theory,” NeuroImage, vol. 16, no. 2, pp. 465–483, 2002. View at Publisher · View at Google Scholar · View at Scopus
  3. G. Marrelec and V. Perlbarg, “Contribution of exploratory methods to the investigation of extended large-scale brain networks in functional MRI: methodologies, results, and challenges,” International Journal of Biomedical Imaging, vol. 2008, Article ID 218519, 14 pages, 2008. View at Publisher · View at Google Scholar · View at Scopus
  4. M. J. McKeown and T. J. Sejnowski, “Independent component analysis of fMRI data: examining the assumptions,” Human Brain Mapping, vol. 6, no. 5-6, pp. 368–372, 1998. View at Publisher · View at Google Scholar · View at Scopus
  5. M. J. McKeown, S. Makeig, G. G. Brown et al., “Analysis of fMRI data by blind separation into independent spatial components,” Human Brain Mapping, vol. 6, no. 3, pp. 160–188, 1998. View at Publisher · View at Google Scholar · View at Scopus
  6. C. F. Beckmann, “Modelling with independent components,” NeuroImage, vol. 62, no. 2, pp. 891–901, 2012. View at Publisher · View at Google Scholar · View at Scopus
  7. A. Hyvarinen, J. Karhunen, and E. Oja, Independent Component Analysis, Wiley-Interscience, 1st edition, 2001.
  8. J. Himberg, A. Hyvärinen, and F. Esposito, “Validating the independent components of neuroimaging time series via clustering and visualization,” NeuroImage, vol. 22, no. 3, pp. 1214–1222, 2004. View at Publisher · View at Google Scholar · View at Scopus
  9. W. Du, S. Ma, G.-S. Fu, V. D. Calhoun, and T. Adali, “A novel approach for assessing reliability of ICA for FMRI analysis,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '14), pp. 2084–2088, Florence, Italy, May 2014. View at Publisher · View at Google Scholar · View at Scopus
  10. F. Artoni, D. Menicucci, A. Delorme, S. Makeig, and S. Micera, “RELICA: a method for estimating the reliability of independent components,” NeuroImage, vol. 103, pp. 391–400, 2014. View at Publisher · View at Google Scholar · View at Scopus
  11. E. A. Allen, E. B. Erhardt, Y. Wei, T. Eichele, and V. D. Calhoun, “Capturing inter-subject variability with group independent component analysis of fMRI data: a simulation study,” NeuroImage, vol. 59, no. 4, pp. 4141–4159, 2012. View at Publisher · View at Google Scholar · View at Scopus
  12. A. Abou-Elseoud, T. Starck, J. Remes, J. Nikkinen, O. Tervonen, and V. Kiviniemi, “The effect of model order selection in group PICA,” Human Brain Mapping, vol. 31, no. 8, pp. 1207–1216, 2010. View at Publisher · View at Google Scholar · View at Scopus
  13. Y.-O. Li, T. Adali, and V. D. Calhoun, “Estimating the number of independent components for functional magnetic resonance imaging data,” Human Brain Mapping, vol. 28, no. 11, pp. 1251–1266, 2007. View at Publisher · View at Google Scholar · View at Scopus
  14. V. D. Calhoun, T. Adali, G. D. Pearlson, and J. J. Pekar, “A method for making group inferences from functional MRI data using independent component analysis,” Human Brain Mapping, vol. 14, no. 3, pp. 140–151, 2001. View at Publisher · View at Google Scholar · View at Scopus
  15. V. D. Calhoun, T. Adali, G. D. Pearlson, and J. J. Pekar, “Erratum: a method for making group inferences from functional MRI data using independent component analysis (Human Brain Mapping (2001) 14 (140-151)),” Human Brain Mapping, vol. 16, no. 2, p. 131, 2002. View at Google Scholar
  16. C. F. Beckmann and S. M. Smith, “Probabilistic independent component analysis for functional magnetic resonance imaging,” IEEE Transactions on Medical Imaging, vol. 23, no. 2, pp. 137–152, 2004. View at Publisher · View at Google Scholar · View at Scopus
  17. T. P. Minka, “Automatic choice of dimensionality for PCA,” in Advances in Neural Information Processing Systems (NIPS), MIT Press, 2000. View at Google Scholar
  18. W. Majeed and M. J. Avison, “Robust data driven model order estimation for independent component analysis of fMRI data with low contrast to noise,” PLoS ONE, vol. 9, no. 4, Article ID e94943, 2014. View at Publisher · View at Google Scholar · View at Scopus
  19. C. L. Nikias and J. M. Mendel, “Signal processing with higher-order spectra,” IEEE Signal Processing Magazine, vol. 10, no. 3, pp. 10–37, 1993. View at Publisher · View at Google Scholar
  20. A. Hyvärinen, P. O. Hoyer, and M. Inki, “Topographic independent component analysis,” Neural Computation, vol. 13, no. 7, pp. 1527–1558, 2001. View at Publisher · View at Google Scholar · View at Scopus
  21. A. Meyer-Bäse, M. K. Hurdal, O. Lange, and H. Ritter, “Clustering of dependent components: a new paradigm for fMRI signal detection,” EURASIP Journal on Advances in Signal Processing, vol. 2005, no. 19, pp. 3089–3102, 2005. View at Publisher · View at Google Scholar · View at Scopus
  22. F. R. Bach and M. I. Jordan, “Beyond independent components: trees and clusters,” Journal of Machine Learning Research, vol. 4, no. 7-8, pp. 1205–1233, 2004. View at Publisher · View at Google Scholar · View at MathSciNet
  23. S. Ma, N. M. Correa, X.-L. Li, T. Eichele, V. D. Calhoun, and T. Adali, “Automatic identification of functional clusters in fMRI data using spatial dependence,” IEEE Transactions on Biomedical Engineering, vol. 58, no. 12, pp. 3406–3417, 2011. View at Publisher · View at Google Scholar · View at Scopus
  24. C. E. Shannon, “A mathematical theory of communication,” The Bell System Technical Journal, vol. 27, no. 3, pp. 379–423, 1948. View at Publisher · View at Google Scholar · View at MathSciNet
  25. Y.-I. Moon, B. Rajagopalan, and U. Lall, “Estimation of mutual information using kernel density estimators,” Physical Review E, vol. 52, no. 3, pp. 2318–2321, 1995. View at Publisher · View at Google Scholar · View at Scopus
  26. A. K. Jain, M. N. Murty, and P. J. Flynn, “Data clustering: a review,” ACM Computing Surveys, vol. 31, no. 3, pp. 264–323, 1999. View at Publisher · View at Google Scholar · View at Scopus
  27. C. A. Cocosco, V. Kollokian, R. K. S. Kwan, G. B. Pike, and A. C. Evans, “BrainWeb: online interface to a 3D MRI simulated brain database,” NeuroImage, vol. 5, no. 4, p. 425, 1997. View at Google Scholar
  28. M. S. Cohen, “Parametric analysis of fMRI data using linear systems methods,” NeuroImage, vol. 6, no. 2, pp. 93–103, 1997. View at Publisher · View at Google Scholar · View at Scopus
  29. J. Talairach and P. Tournoux, Co-Planar Stereotaxic Atlas of the Human Brain: 3-D Proportional System: An Approach to Cerebral Imaging (Thieme Classics), Thieme, 1988.
  30. R. W. Cox, “AFNI: software for analysis and visualization of functional magnetic resonance neuroimages,” Computers and Biomedical Research, vol. 29, no. 3, pp. 162–173, 1996. View at Publisher · View at Google Scholar · View at Scopus
  31. N. Delfosse and P. Loubaton, “Adaptive blind separation of independent sources: a deflation approach,” Signal Processing, vol. 45, no. 1, pp. 59–83, 1995. View at Publisher · View at Google Scholar · View at Scopus
  32. A. Hyvärinen and E. Oja, “A fast fixed-point algorithm for independent component analysis,” Neural Computation, vol. 9, no. 7, pp. 1483–1492, 1997. View at Publisher · View at Google Scholar · View at Scopus
  33. N. Correa, T. Adali, and V. D. Calhoun, “Performance of blind source separation algorithms for fMRI analysis using a group ICA method,” Magnetic Resonance Imaging, vol. 25, no. 5, pp. 684–694, 2007. View at Publisher · View at Google Scholar · View at Scopus
  34. T. M. Cover and J. A. Thomas, Elements of Information Theory, Wiley-Interscience, New York, NY, USA, 1991. View at Publisher · View at Google Scholar · View at MathSciNet
  35. G. W. Snedecor and W. G. Cochran, Statistical Methods, Iowa State University Press, Ames, Iowa, USA, 8th edition, 1989. View at MathSciNet
  36. R. Steuer, J. Kurths, C. O. Daub, J. Weise, and J. Selbig, “The mutual information: detecting and evaluating dependencies between variables,” Bioinformatics, vol. 18, supplement 2, pp. S231–S240, 2002. View at Publisher · View at Google Scholar · View at Scopus
  37. B. W. Silverman, Density Estimation for Statistics and Data Analysis, Chapman & Hall/CRC, 1st edition, 1986.
  38. J. H. Ward, “Hierarchical grouping to optimize an objective function,” Journal of the American Statistical Association, vol. 58, no. 301, pp. 236–244, 1963. View at Publisher · View at Google Scholar · View at MathSciNet
  39. P. Skudlarski, R. T. Constable, and J. C. Gore, “ROC analysis of statistical methods used in functional MRI: individual subjects,” NeuroImage, vol. 9, no. 3, pp. 311–329, 1999. View at Publisher · View at Google Scholar · View at Scopus
  40. D. Bamber, “The area above the ordinal dominance graph and the area below the receiver operating characteristic graph,” Journal of Mathematical Psychology, vol. 12, no. 4, pp. 387–415, 1975. View at Google Scholar · View at MathSciNet
  41. T. A. Hammeke, F. Z. Yetkin, W. M. Mueller et al., “Functional magnetic resonance imaging of somatosensory stimulation,” Neurosurgery, vol. 35, no. 4, pp. 677–681, 1994. View at Publisher · View at Google Scholar · View at Scopus
  42. K. J. Friston, A. P. Holmes, J.-B. Poline et al., “Analysis of fMRI time-series revisited,” NeuroImage, vol. 2, no. 1, pp. 45–53, 1995. View at Publisher · View at Google Scholar · View at Scopus
  43. C. E. McCulloch and S. R. Searle, Generalized, Linear, and Mixed Models, Wiley Series in Probability and Statistics, Wiley-Interscience, 1st edition, 2001. View at MathSciNet
  44. D. W. Scott, Multivariate Density Estimation: Theory, Practice, and Visualization, Wiley, 1st edition, 1992. View at Publisher · View at Google Scholar · View at MathSciNet
  45. G. A. Darbellay and I. Vajda, “Estimation of the information by an adaptive partitioning of the observation space,” IEEE Transactions on Information Theory, vol. 45, no. 4, pp. 1315–1321, 1999. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  46. V. Kiviniemi, T. Starck, J. Remes et al., “Functional segmentation of the brain cortex using high model order group PICA,” Human Brain Mapping, vol. 30, no. 12, pp. 3865–3886, 2009. View at Publisher · View at Google Scholar · View at Scopus
  47. F. Esposito, E. Formisano, E. Seifritz et al., “Spatial independent component analysis of functional MRI time-series: to what extent do results depend on the algorithm used?” Human Brain Mapping, vol. 16, no. 3, pp. 146–157, 2002. View at Publisher · View at Google Scholar · View at Scopus