Research Article | Open Access
Hao Guo, Mengna Qin, Junjie Chen, Yong Xu, Jie Xiang, "Machine-Learning Classifier for Patients with Major Depressive Disorder: Multifeature Approach Based on a High-Order Minimum Spanning Tree Functional Brain Network", Computational and Mathematical Methods in Medicine, vol. 2017, Article ID 4820935, 14 pages, 2017. https://doi.org/10.1155/2017/4820935
Machine-Learning Classifier for Patients with Major Depressive Disorder: Multifeature Approach Based on a High-Order Minimum Spanning Tree Functional Brain Network
High-order functional connectivity networks are rich in time information that can reflect dynamic changes in functional connectivity between brain regions. Accordingly, such networks are widely used to classify brain diseases. However, traditional methods for processing high-order functional connectivity networks generally include the clustering method, which reduces data dimensionality. As a result, such networks cannot be effectively interpreted in the context of neurology. Additionally, due to the large scale of high-order functional connectivity networks, it can be computationally very expensive to use complex network or graph theory to calculate certain topological properties. Here, we propose a novel method of generating a high-order minimum spanning tree functional connectivity network. This method increases the neurological significance of the high-order functional connectivity network, reduces network computing consumption, and produces a network scale that is conducive to subsequent network analysis. To ensure the quality of the topological information in the network structure, we used frequent subgraph mining technology to capture the discriminative subnetworks as features and combined this with quantifiable local network features. Then we applied a multikernel learning technique to the corresponding selected features to obtain the final classification results. We evaluated our proposed method using a data set containing 38 patients with major depressive disorder and 28 healthy controls. The experimental results showed a classification accuracy of up to 97.54%.
Resting-state functional magnetic resonance imaging (rs-fMRI) using blood oxygenation level-dependent (BOLD) signals as neurophysiological indicators can detect spontaneous low-frequency brain activity and has been successfully applied to the diagnosis of neuropsychiatric diseases such as schizophrenia [1–4], Alzheimer’s disease [5–7], epilepsy [8–10], attention deficit hyperactivity disorder (ADHD) , and stroke [12, 13]. Resting functional brain network analysis helps clarify the mechanisms of neuropsychiatric disorders and has the potential to provide relevant imaging markers that may offer new perspectives for the diagnosis and evaluation of clinical brain diseases . In traditional brain network analysis, it is assumed that the correlation between different brain regions does not change with time during rs-fMRI scanning. Networks constructed using methods based on this assumption are called low-order networks .
However, this assumption may lead researchers to overlook the dynamic interaction patterns between brain regions during the entire scan, which are essentially time-varying. Indeed, several recent studies have indicated that functional connectivity analyses can be rich in dynamic temporal information [15, 16]. High-order functional connectivity networks contain abundant dynamic time information, so this method has been proposed and applied in the diagnosis of brain diseases [14, 17].
The most common method for constructing a high-order functional network is the dynamic sliding window method, in which the whole rs-fMRI time series is divided into several time windows . A low-order functional connectivity network is built in each time window, and then all the low-order networks are stacked. A clustering algorithm is performed to divide all relevant time series into several clusters. The average time series of each cluster is then taken as a new node, and the Pearson correlation coefficient is calculated between each node pair as the weight of connectivity .
In this method, clustering is employed to decrease the associated computational costs, and classification accuracy is greatly influenced by the randomness of the selection of initial clustering centers and the number of clusters. However, because the time series of all connectivities within each cluster are averaged, the network loses neurological interpretability.
In the present study, we used the minimum spanning tree method  to reduce the computational cost while preserving the core framework of networks. A classic approach in graph theory, this unbiased method greatly simplifies the network structure while preserving its core framework, thus avoiding the influences of network sparseness and other parameters on network structure. It also guarantees the network’s neurological interpretability and has been widely used in previous studies [20–22].
Furthermore, the traditional feature extraction method in minimum spanning tree networks uses a quantifiable network with local features for classification of brain diseases, such as degree, clustering coefficient, minimum path length, and eccentricity [23, 24]. However, a clear shortcoming of this method was the chance that some of the useful topology information in the network (including connection patterns in the sample itself and the common connection patterns between the samples) would be lost, resulting in reduced classifier performance. Frequent subgraph mining technology was proposed to mine discriminative subgraph pattern features for machine-learning classification of brain diseases [25, 26]. Subgraph pattern features could account for the connection pattern information between multiple brain regions, but it was not sensitive to changes in single brain regions . Therefore, both methods can lead to loss of sample information.
Here, we propose a novel feature extraction method that combines quantifiable local network features with subgraph pattern features. Specifically, we computed degree, eccentricity, and betweenness centrality of each brain region as local network features and extracted the subgraph features using a frequent subgraph mining method for a group of healthy controls (HC) and a group of people with major depressive disorder (MDD). Then a kernel function for each type of feature was constructed, namely, a vector kernel (local network features) and a graph kernel (subgraph features). Finally, the two kernel matrices were combined and a multikernel support vector machine was constructed as a classifier. The proposed method achieves better classification performance than traditional methods that use only a single type of feature.
2. Materials and Methods
2.1. Proposed Framework
Figure 1 shows the flowchart of the proposed method, which includes four main steps: (1) data acquisition and preprocessing; (2) network construction, in which a high-order functional connectivity network is constructed first, followed by construction of a minimum spanning tree network; (3) feature extraction and selection, in which two types of feature are extracted and selected (the first is used to calculate quantifiable local network features (degree, betweenness centrality, and eccentricity) and uses the Kolmogorov–Smirnov test for feature selection, and the second is used to mine frequent subgraphs from the HC and MDD groups and selects the most discriminative subnetworks as the subgraph patterns); (4) construction of a classification model, in which the kernel matrix of the two types of feature is calculated. The multiple-kernel support vector machine (SVM) is adopted to combine the two heterogeneous kernels, enabling the distinction of individuals with MDD from healthy controls.
2.2. Data Acquisition and Preprocessing
The study was carried out in accordance with the recommendations of the medical ethics committee of Shanxi Province (reference number: 2012013). All subjects provided written informed consent in accordance with the Declaration of Helsinki. Twenty-eight healthy subjects and thirty-eight people with MDD underwent rs-fMRI in a 3T scanner (Siemens Trio 3-Tesla scanner, Siemens, Erlangen, Germany). Participant demographic information is shown in Table 1.
|Data are minimum–maximum (mean ± standard deviation). HAMD, 24-item Hamilton scale; two-sample, two-tailed -test; two-tailed Pearson’s chi-squared test.|
Data collection was completed at the First Hospital of Shanxi Medical University. Radiologists familiar with fMRI performed all scans. During each scan, the participant was asked to relax with their eyes closed and not think about anything in particular but to stay awake and avoid falling asleep. Each scan consisted of 248 contiguous echo-planar imaging (EPI) functional volumes (33 axial slices, repetition time (TR) = 2000 ms, echo time (TE) = 30 ms, thickness/skip = 4/0 mm, field of view (FOV) = 192 × 192 mm, matrix = 64 × 64 mm, and flip angle = 90°). The first 10 volumes in the time series were discarded to account for magnetization stabilization. See Supplemental Text S1 for detailed scanning parameters.
Data preprocessing was performed in SPM8 (http://www.fil.ion.ucl.ac.uk/spm/) with slice-timing and head-movement corrections. Two samples containing a translation of more than 3.0 mm and rotation of more than 3.0° were excluded from the final analysis of 66 samples. Functional images were normalized using the 12 parameters from the affine transformation and the cosine-based nonlinear transformation from the normalization of the anatomic image to the Montreal Neurological Institute (MNI) space. Additional normalization of the functional data sets to the SPM8 EPI template was then performed, and the data were resampled to a voxel size of mm using a sinc interpolation. No smoothing kernel was applied to limit spurious local connectivity between voxels. Finally, we performed linear detrending and band-pass filtering (0.01–0.10 Hz) to reduce the effects of low-frequency drift and high-frequency physiological noise. Then, for each subject, the brain space of the fMRI images was parcellated into 90 regions of interest (ROIs) (45 in each hemisphere) based on the automated anatomical labeling (AAL) atlas , and each region was defined as a node in the network. Each regional mean time series was regressed against the average cerebral spinal fluid and white-matter signals as well as the six parameters from motion correction. The residuals of these regressions constituted the set of regional mean time series used for undirected graph analysis.
2.3. Construction of the High-Order Minimum Spanning Tree Network
2.3.1. High-Order Functional Connectivity Network
A high-order functional connectivity network was constructed using a flowchart with the following steps (Figure 2): (1) partition the entire rs-fMRI time series into multiple overlapping segments of subseries by adopting a fixed-length sliding window; (2) construct temporal low-order functional connectivity networks in each time window; (3) stack together all low-order functional connectivity networks for all subjects; (4) construct a high-order functional connectivity network for each subject by taking the low-order functional connectivity as the new nodes and the pairwise Pearson correlation coefficient between each pair of nodes as the path weight.
To enable construction of a low-order functional connectivity network in each time window, we divided the whole time series into a number of overlapping subseries segments using the sliding time-window method. Specifically, if the length of the sliding window is and the step size between two successive windows is , let denote the th segment of the subseries extracted from . The total number of segments generated by this approach is given byThe length of our sliding window was 90 and the step length was 1. For an illustration of the sliding window, see Supplemental Figure S1.
For the th subject, the th segment in the subseries for all ROIs can be expressed in matrix form as where is the total number of ROIs. Then, the entry for the th temporal functional connectivity matrix for the th subject can be obtained by the Pearson correlation between the th and th ROIs. The temporal functional connectivity networks for the th subject can be established by taking as nodes and as the weights of new edges, as per the following equation:
In this way, it is possible to construct dynamic temporal functional connectivity networks for each subject. For each ROI pair for the th subject, we can concatenate to obtain a correlation time series:We can then stack together all the dynamic temporal functional connectivity networks for each subject, as per (4).
The main goal of this article is to reveal the intrinsic relationship between the correlation time series and the dynamic temporal information contained within it. We calculated the Pearson correlation coefficient between each pair of correlation time series for each subject as follows:Thus, the construction of a high-order functional connectivity network is achieved by taking as new nodes and as the weights of new edges and then connecting nodes and . The new high-order functional connectivity network can be represented asTherefore, can be said to represent the high-order correlation, and the corresponding network represents the high-order functional connectivity network. The high-order correlation indicates the linear correlation strength between two correlation time series and reflects the interaction between up to four brain regions. Compared with the traditional network, the high-order functional connectivity network not only takes into account the time-varying characteristics of functional connectivity but also represents the more complex and abstract interaction patterns among brain regions.
2.3.2. Minimum Spanning Tree
To further reduce the complexity of the high-order functional connectivity network, we constructed a minimum spanning tree. This is a weighted subnetwork (fully connected network) that connects all the nodes in the network without forming loops and has the minimum total weight of all possible spanning trees . We constructed the minimum spanning tree based on the weighted network. Since we were interested in determining the strongest connection in the network, we used Kruskal’s algorithm to obtain the strongest connection weights . This algorithm first sorts the edges into descending weight order and then starts the construction of the minimum spanning tree from the largest-weight edge, adding the next largest-weight edge until all nodes are connected in an acyclic subnetwork consisting of edges. When the addition of an edge forms a loop, this edge is ignored. For more information regarding Kruskal’s algorithm, see Supplemental Text S2.
2.4. Feature Extraction and Selection
After completion of the network, we extracted features of two different types: quantifiable local network features of the minimum spanning tree and subgraph patterns from frequent subgraph mining. We selected quantifiable local network features of the minimum spanning tree using the Kolmogorov-Smirnov test. For the connected patterns from frequent subgraph mining, we used discriminative scores to select the most discriminative subgraphs.
2.4.1. Local Network Features and Selection Methods
We selected the local network properties of the minimum spanning tree (degree, betweenness centrality, and eccentricity) as features. We calculated the three properties of each node in the high-order minimum spanning tree network. Table 2 gives the definition and formula of these three properties. We used multilinear regression analysis to assess the confounding effect of age, sex, and educational attainment on each network attribute. The independent variable was the mean of each network attribute (except for the degree, owing to its nature) and the dependent variables were age, sex, and educational attainment. The results showed no significant correlations between betweenness centrality, eccentricity, and corresponding variables (see Supplemental Table S1 for results).
|, connection between and j; , shortest path from u to v; , number of shortest paths between h and j; , number of shortest paths between h and which passthrough i.|
We used the Kolmogorov-Smirnov test  to select the quantifiable local network features of the minimum spanning tree (). The results were then corrected using the Benjamini-Hochberg false positive rate () .
2.4.2. Frequent Subgraph and Discriminative Evaluation
(1) Frequent Subgraph. In this paper, subgraph pattern extraction was mainly based on frequent subgraph mining. The frequent subnetwork refers to the connected patterns that appear most often in the network . The purpose of frequent subnetwork mining is to uncover the most frequent connected patterns (i.e., subnetworks) in the whole network . We applied this algorithm to the HC and MDD groups. In the field of data mining, a large number of frequent subgraph mining methods have been proposed [45, 46], including a priori-based graph mining  and the frequent subgraph discovery algorithm . Here, we used the well-known gSpan algorithm  to extract the frequent subnetworks from the functional connectivity network. Because of its high efficiency in graph traversal and subgraph mining, the gSpan algorithm has been widely applied in many research fields, including neural imaging [25–27].
The gSpan algorithm works as follows . First, the gSpan constructs a new lexicographic order among graphs and maps each graph to a unique minimum depth-first search (DFS) code as its canonical label. Then, based on the lexicographic order, gSpan uses the DFS strategy to efficiently mine frequently connected subgraph patterns. In the present study, we termed the hierarchical search space of frequent subgraphs the “DFS code tree,” where each node in the tree represents a DFS code (i.e., subgraph). The th level subgraph is generated from the th level subgraph (i.e., parent) by adding one frequent edge. Finally, all subgraphs with nonminimal DFS codes are pruned to avoid redundant candidate generations. In subgraph mining, the number of subgraphs is mainly controlled by frequency. Given a set of graphs, , the frequency of a subgraph is defined asThe DFS lexicographic order used in frequent subgraph mining and the gSpan algorithm are described in detail in Supplemental Text S3.
(2) Discriminative Evaluation. The discriminative subnetwork can be used as a feature for classification , but it is worth noting that gSpan is only used for mining the frequent subgraph, which, by itself, has no discriminative power. For information on the discriminative capabilities of different subgraphs, see Supplemental Figure S2. However, some of the frequent subnetworks may have less discriminative information for classification. To address this problem, we selected the most discriminative subnetworks from the frequent subnetworks using subgraph discriminative scores (which express subgraph frequency differences) . This strategy is called frequent-scoring feature selection. In the present study, the method involved choosing the same number of frequent subgraphs from the HC and MDD groups, calculating and sorting the discriminative scores of frequent subgraphs, and selecting the top subnetworks with higher discriminative scores. Thus, discriminative subnetworks are selected. For the given graphs and refers to the set of frequent subgraphs for all positive samples and refers to the set of all frequent subgraph features for negative samples. The discriminative scores of subgraph can be calculated asThe discriminative score of a subgraph pattern is simply defined as the difference between its positive frequency and negative frequency. A larger score reflects a larger difference between the patterns in the two groups. indicates that the subgraph exists in all graphs for the HC group and that there is no such pattern in any graph for the MDD group. indicates that the subgraph exists in all graphs for the MDD group and that there is no such pattern in any graph for the HC group.
2.5. Construction of Classification Model
The classification model chosen in this paper is a multikernel SVM. Recent studies on multikernel learning have shown that the integration of multiple kernels can significantly improve classification and enhance interpretability of results . Generally, the integration of the kernel is achieved by linear combination of multiple kernels:where is a basic kernel built for subjects and , is the number of kernel matrices required, and is the nonnegative weighting parameter.
A graph kernel can be regarded as a group of similarities between a pair of subjects. The brain network data is mapped from the original network space to the feature space, and the similarity between the two brain networks is further measured by comparing their topology. In this study, we used the Weisfeiler-Lehman subtree, based on the Weisfeiler-Lehman isomorphism test , to measure the topological similarity between paired connectivity networks. This type of graph kernel can effectively capture topological information from graphs and improve performance. Given two graphs, the basic process of the Weisfeiler-Lehman test is as follows: if the two graphs are unlabeled (i.e., the nodes of the graph have not been assigned labels), each node is first labeled with the number of edges that are connected to that node. Then, at each iteration step, the label of each node is updated based on its previous label and the labels of its neighbors. That is, the sorted set of updated node labels for each node is compressed such that it contains new and shorter labels. This process iterates until the node label sets are identical or the number of iterations reaches its predefined maximum value. For a detailed description of the Weisfeiler-Lehman isomorphism test and pseudocode, see Supplemental Text S4.
As this study involves two different types of kernel (vector-based kernels and graph kernels), a normalization step must be performed individually before combining them. This normalization step can be accomplished using the following formula:Note that, unlike the previous multikernel learning method, in which the weighting parameters are jointly optimized together with other classifier parameters, in this study, the optimal weighting parameters are determined via a grid search of the training data. Once the optimal weighting parameters are obtained, the multikernel learning-based classifier can be naturally embedded into the conventional single-kernel classifier framework. In this paper, we selected the SVM as the classifier framework.
As described above, we used multikernel learning methods to perform classification. As different types of kernels represent different properties of the network, we combined multiple features through multikernel learning. Specifically, the vector-based kernel describes the correlation between pairwise brain regions according to degree, betweenness centrality, and eccentricity, and the graph-based kernel describes the topological information contained in the whole network.
We performed two types of feature extraction on the constructed network. The first involved the quantifiable local network features, namely, degree, betweenness centrality, and eccentricity. The second involved the extraction of discriminative subgraph patterns from the HC and MDD groups.
3.1. Abnormal Functional Connectivities
The high-order functional connectivity network was , so there were 4004 edges in the high-order minimum spanning tree network. After constructing the network, we analyzed three kinds of traditional quantifiable network properties. We selected the high-order functional connectivities of at least two network properties with (false discovery rate corrected). We obtained 40 abnormal functional connectivities in total, encompassing a total of 42 abnormal regions (Table 3). All 40 significant abnormal functional connectivities and frequency-corresponding nodes are shown in Supplemental Figure S3. These significant regions were concentrated in the limbic-cortical networks (left anterior cingulate and paracingulate gyri; bilateral median cingulate and paracingulate gyri; right posterior cingulate gyrus; bilateral caudate nucleus; bilateral lenticular nucleus; bilateral putamen; bilateral thalamus; bilateral hippocampus; bilateral parahippocampal gyrus; and bilateral amygdala), frontal lobe (bilateral precentral gyrus; bilateral dorsolateral superior frontal gyrus; bilateral superior frontal gyrus, orbital part; right middle frontal gyrus; bilateral middle frontal gyrus, orbital part; bilateral inferior frontal gyrus, triangular part; and bilateral inferior frontal gyrus, opercular part), temporal lobe (right temporal pole: middle temporal gyrus; left Heschl gyrus), and cuneus (bilateral cuneus; bilateral lingual gyrus; bilateral precuneus; right postcentral gyrus; left calcarine fissure; and surrounding cortex).
|Values in bold indicate significance (). For all brain region abbreviations, see Supplemental Table S2.|
We selected the top 10 brain regions with the most significant differences in terms of frequency (Table 4). These mainly included the bilateral temporal pole: middle temporal gyrus; left superior frontal gyrus, orbital part; left thalamus; right lenticular nucleus; putamen; left lingual gyrus; right cuneus; left posterior cingulate gyrus; and right dorsolateral superior frontal gyrus.
|For all brain region abbreviations, see Supplemental Table S2.|
3.2. Frequent Subgraph Patterns
We analyzed the frequent discriminative subnetworks. We mined two sets of frequent subnetworks from the functional connectivity networks of the MDD and HC groups, with respective frequencies and . Specifically, we mined 4057 subgraphs from the HC group and 4078 from the MDD group. For statistical information regarding the number of edges in the subgraphs, see Supplemental Table S3. We calculated the discriminative scores of the frequent subgraphs and found 16 subgraphs for the HC group and 37 for the MDD group. To ensure that the features were balanced, we selected 16 discriminative subnetworks to assess the subgraph patterns from the two groups.
To analyze the connected patterns, connections in the 16 subgraph connected patterns for each of the HC and MDD groups were merged with the subgraph in Supplemental Figure S4. By analyzing the subgraphs of the HC and MDD groups, we found some nodes that were significantly different between the two groups. These significantly different nodes were mainly concentrated in the bilateral lenticular nucleus: the putamen; bilateral lingual gyrus; bilateral amygdala; bilateral thalamus; bilateral median cingulate and paracingulate gyri; right posterior cingulate gyrus; bilateral cuneus; left anterior cingulate and paracingulate gyri; right superior frontal gyrus, orbital part; right middle frontal gyrus; right temporal pole; middle temporal gyrus; left precentral gyrus; right lenticular nucleus; pallidum; and so forth.
We also analyzed these significantly different brain regions and ranked them according to the frequency in which they appeared in the HC and MDD groups. The significantly different brain regions and the frequency-corresponding nodes are given in Supplemental Figure S5. We selected the top 10 regions as the most discriminative (Table 5).
|For all brain region abbreviations, see Supplemental Table S2.|
3.3. Classification Results
We evaluated the classification performance of the proposed method by measuring classification accuracy, sensitivity, specificity, and area under the curve (Table 6). Table 6 also compares the classification performance of the partial correlation functional connectivity network, Pearson functional connectivity network, high-order functional connectivity network, and frequent subgraph mining methods. The results indicate that our proposed method achieves good results in terms of classification accuracy, sensitivity, specificity, and area under the curve.
|FC, functional connectivity; MST, minimum spanning tree; SAD, social anxiety disorder; MCI, mild cognitive impairment; ADHD, attention deficit hyperactivity disorder; MDD, major depressive disorder; AUC, area under receiver operating characteristic (ROC) curve.|
Specifically, to compare the method proposed in this paper with those used previously, we constructed partial and Pearson correlation networks and a high-order functional connectivity network without minimum spanning tree analysis (see Supplemental Text S5 for the detailed information of other contrast networks). In addition, we used a high-order minimum spanning tree network for assessing quantifiable local network features and subgraph patterns as features. Our experimental results showed that the proposed method of classification performed significantly better than the partial correlation network, Pearson correlation network, and high-order functional connectivity network and also better than the method in which only the quantifiable local network features and subgraph pattern features were assessed. This illustrates the potential for the integration of the two different types of features to significantly improve classification performance. We used the Relief method  to calculate the average weights of subgraph pattern features, the minimum spanning tree of quantifiable local network features, and both types of features combined (Figure 3). The average weight of the subgraph pattern features was 550.31, that of the minimum spanning tree of the quantifiable local network features was 915.42, and that for both feature types together was 945.16. Figure 4 shows the receiver operating characteristic curve of the proposed method, the partial correlation network, the Pearson correlation network, the higher-order functional connectivity network, and the approach with only subgraph patterns as features and quantifiable local network features. These results indicate that our proposed method clearly improved classification performance.
4.1. Abnormal Brain Regions
We extracted quantifiable local network features and frequent subgraph features to explore brain regions with significantly abnormal connectivities between the HC and MDD groups. By calculating the quantifiable local network features, we were able to obtain 40 significant abnormal connectivities, involving 42 brain regions. Then, we selected the top 10 most frequently implicated brain regions, as those were the most significantly different between the two groups. Consistent with previous studies, these regions included the bilateral temporal pole: the middle temporal gyrus; the left superior frontal gyrus, orbital part; the right lateral dorsolateral frontal gyrus; the left thalamus; the right putamen; the left lingual gyrus; the right cuneus; and the left posterior cingulate gyrus.
The present results are consistent with our previous findings. Ma et al.  adopted voxel-based morphometry to investigate brain regions with gray-matter abnormality in patients with treatment-resistant depression and in those with treatment-responsive depression. They found that patients in both groups showed clear gray-matter abnormalities in the right temporal pole of the temporal gyrus, specifically the middle temporal gyrus. Qiu et al.  examined cortical thickness and surface area in first-episode, treatment-naïve, mid-life MDD and observed a significant increase in gray-matter volume in the left superior frontal gyrus, left thalamus, and right cuneus. Sacchet et al.  obtained whole-brain T1-weighted images in their own HC and MDD groups and evaluated gray-matter volumes in the basal ganglia (specifically the caudate nucleus, lentiform pallidum, and putamen). They reported that gray-matter volumes in the bilateral lenticular nucleus and putamen were significantly different between patients with depression and healthy controls. Jung et al.  used voxel-based morphometry to detect structural changes in healthy subjects and patients with depression who underwent 8 weeks of antidepressant treatment. The results showed significantly different gray-matter volume in the left lingual gyrus between the two participant groups. Fang et al.  measured spontaneous whole-brain homodynamic responses using amplitude of low-frequency fluctuation (ALFF) and fractional ALFF (fALFF) and found that the ALFF and fALFF decreased in the left posterior cingulate gyrus, the right cuneus, and the superior frontal gyrus after depression treatment. Cotter et al. also found abnormalities in the dorsal prefrontal cortex of patients with MDD .
In the present study, frequent subgraph mining revealed a total of 32 discriminative patterns (16 in the HC group and 16 in the MDD group). We found 19 common brain regions in the 32 connected patterns in the two groups. According to the frequency of each region in the connected patterns, we selected the 10 most discriminative brain regions. These included the bilateral lenticular nucleus putamen, bilateral lingual gyrus, bilateral amygdala, left median cingulate paracingulate gyri, right posterior cingulate gyrus, and bilateral thalamus. Anand et al.  studied the differences between limbic-cortical activities and connections in patients with MDD versus HCs. They found significant differences between patients and controls in the bilateral anterior cingulate cortex, bilateral amygdala, and bilateral thalamus. The top 10 most frequent regions in our study included the amygdala, which is part of the limbic system and is involved in the formation of emotional behavior, spontaneous activity, and endocrine integration processes. Previous studies [60–62] have indicated that the amygdala plays a significant role in the pathogenesis of depression. Veer et al.  used independent component analysis to assess rs-fMRI data from 19 medication-free patients with a recent diagnosis of MDD (within 6 months prior to inclusion) and no comorbidity and 19 age- and gender-matched controls. They found decreased activation in the bilateral amygdala, which is associated with emotional behavior, the frontal lobe, which is associated with attention and working memory, and the lingual gyrus, which is related to visual processing. The other discriminative brain regions that we identified via frequent subgraph mining are also consistent with previous results, such as the bilateral lenticular putamen , the left median cingulate and paracingulate gyri , and the right posterior cingulate gyrus .
Three of the brain regions with the most significant differences were obtained by two analysis methods: quantifiable local network features and discriminative subgraph patterns. These included the right lenticular nucleus (putamen), left lingual gyrus, and left thalamus. The right lenticular nucleus (putamen) and the left thalamus are key regions of the limbic-cortical circuit and the default network. Meng et al.  suggested that reward-related dysfunction of the right putamen within a striatum-centered limbic-cortical circuit may inhibit learning related to appreciating and enjoying positive life experiences, which is critical for depression recovery. As the thalamus makes a critical connection between the amygdala and the prefrontal cortex, it is well positioned for involvement in MDD pathophysiology. Similarly, the right lingual gyrus is a key region of the visual network. Jung et al. reported that the volume of the lingual gyrus is associated with neuropsychological features of depression . The lingual gyrus plays an important role in visual processing. Therefore, the results of this study may be helpful in the search for biomarkers of MDD.
4.2. Classification Result Analysis
To study dynamic changes in functional connectivities between brain regions, Chen et al. used sliding time windows to construct a high-order functional connectivity network that can be used for classification . Their method has a high accuracy for diagnosing mild cognitive impairment (MCI). To show that features obtained from subgraph patterns can better reflect the topological information among brain regions, Du et al. adopted frequent subgraph mining technology to mine frequent subnetworks in fMRI data from people with ADHD. They used a frequent-scoring feature selection method to choose discriminative subnetworks and kernel principal component analysis to extract features, before using LIBSVM (library for support vector machines) for classification . Wang et al. also used frequent subgraph mining techniques to mine discriminative subnetworks that were based on fMRI data from people with MCI . They combined traditional quantifiable properties with local clustering coefficients as features and then used a multikernel SVM for classification. And Fei et al. used frequent subgraph mining techniques combined with a discriminative subnetwork mining algorithm . They then used a graph kernel-based SVM for classification. Their results showed that frequent subgraph patterns are highly accurate as features of classification.
Table 6 compares the accuracy, specificity, sensitivity, and area under the curve of the methods used in the present study. Different methods can produce different results using the same data set, and similar methods can produce different results with different data sets. Therefore, we constructed a partial correlation network, Pearson correlation network, and high-order functional connectivity network with the same data set. Our method for constructing a high-order minimum spanning tree network is superior to the other networks (Table 6). Compared with the traditional method, the high-order minimum spanning tree network can reveal stronger and more complex interactions between brain regions and thus may significantly improve the diagnostic accuracy rate in patients with MDD. Likewise, construction of a high-order minimum spanning tree functional connectivity network may result in better extraction of information regarding the interactions between brain regions in original rs-fMRI time series.
We independently classified quantifiable local network features and subgraph patterns as features on the same data set. Regardless of classification accuracy, sensitivity, specificity, or area under the curve, the classification result produced by the method proposed in this study performed better than an analysis with only quantifiable local network features or with only the subgraph pattern as features (Table 6). We used the Relief method to calculate the average weights for subgraph pattern features, quantifiable local network features, and both types of feature combined. As a result, our proposed method obtained the highest average weights. The Relief algorithm is a feature-weighting algorithm in which weights are continuously adjusted to show correlation between features, until the feature with the largest weight can be identified. Because this algorithm has high efficiency and can be used to make an accurate selection of discriminative features, it has been widely used in many fields, including biomedicine . Given the results from the classification and feature analysis using the Relief method, the combination of quantifiable local network features and subgraph patterns as features appears to effectively reflect the information contained in a single brain region, while simultaneously reflecting the topological information contained in multiple brain regions. Combining these two different types of feature is likely to substantially improve diagnostic accuracy for patients with MDD.
4.3. Influence of Frequency on Graph Features
In this experiment, we mined frequent subnetworks from the functional connectivity network, which was constructed using data from the HC and MDD groups. This construction involved the selection of frequency, which can control the number of selected graph features. However, the high-order functional connectivity network indicates that there exists a temporal correlation between different low-order, dynamic functional connectivity networks. Thus, the size of the network can reach . Even if the intercept associated with sparsity is very small (0.1 or 0.05), the size of the network can reach tens of thousands of edges, and the data for each subject will also have tens of thousands of edges. Hence, the number of subgraphs will be larger when mining frequent subgraphs, which is not conducive to the selection or analysis of subgraph features. Therefore, we constructed the minimum spanning tree network after constructing the high-order functional connectivity network. This method guarantees the integrity of the topological information associated with the high-order functional connectivity network and reduces the complex scale of the network. However, each subject in the minimum spanning tree network has only 4004 edges, which account for only 0.02% of high-order functional connectivity. Thus, frequent subgraph mining frequency must not be too big; if it is, subgraph pattern mining will not be possible. In contrast, if the frequency selection is too small, the subgraph patterns may be too large, increasing the chance of abandoning a large number of discriminative subnetwork patterns in the process of frequent subgraph mining. In the present study, we selected the frequencies of the HC and MDD groups as 0.29 and 0.21, respectively. We left the minimum spanning tree quantifiable local network features unchanged and only changed the features of the subgraph patterns. The different frequencies of the HC and MDD groups were used for classification. The classification results were optimal when the frequencies were 0.29 and 0.21 for the HC and MDD groups, respectively (Table 7).
|HC, healthy control group; MDD, major depressive disorder group; AUC, area under receiver operating characteristic (ROC) curve.|
4.4. Influence of Optimal Weighted Parameters on Classification
Multikernel SVMs are widely adopted in neuroimaging classification . Optimizing the weighting parameter is very important in such classification, and optimal parameter selection will affect the classification results. We tested optimal parameters from 0 to 1 with a step size of 0.1. Figure 5 shows the classification accuracy of different parameters. Classification accuracy was 94%–98% when different optimal parameters were used and was highest (97.54%) when the optimal parameter was 0.4.
The high-order functional connectivity network is relatively large, making it computationally expensive to use certain elements of complex network or graph theory to calculate topological properties. In the construction of a network, previous classification methods are based on local network features, so some useful network topology information may be lost. To address this, we proposed and tested the high-order minimum spanning tree to reduce computational consumption. We combined quantifiable local network features with discriminative subgraph patterns as features and then used a multikernel SVM for classification. The results showed that the high-order minimum spanning tree functional connectivity network could reflect dynamic changes in functional connectivities between brain regions. Additionally, the high-order network takes into account time-varying characteristics, such that the functional connectivity can reflect stronger and more complex interactions among more brain regions. The consistency of the results from the two different types of feature, that is, quantifiable local network features and frequent discriminative subgraph patterns, indicates that the detected significant differences between brain regions were consistent. More importantly, compared with traditional methods, the proposed method appears to offer better classification performance and thus may greatly improve the accuracy of MDD diagnosis. In future work, we plan to explore the impact of these functional connectivities and the relationship between the various ROIs, with the aim of further improving classification performance and better explaining pathology.
This study was approved by the medical ethics committee of Shanxi Province, and the approval certification number is 2012013.
All subjects have given written informed consent in accordance with the Declaration of Helsinki.
The sponsors had no role in the design or execution of the study; the collection, management, analysis, or interpretation of the data; or the preparation, review, or approval of the manuscript. All authors have read through the manuscript and approved it for publication. Hao Guo had full access to all data in the study and takes responsibility for its integrity and the accuracy of data analysis.
Conflicts of Interest
The authors declare that there are no conflicts of interest.
This study was supported by research grants from the National Natural Science Foundation of China (61373101, 61472270, 61402318, and 61672374), the Natural Science Foundation of Shanxi Province (201601D021073), and the Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi (2016139).
Supplementary 1. Supplemental Text S1: Image acquisition.
Supplementary 2. Supplemental Text S2: Frequent subgraph mining algorithm.
Supplementary 3. Supplemental Text S3: Kruskal’s algorithm.
Supplementary 4. Supplemental Text S4: Weisfeiler-Lehman algorithm.
Supplementary 5. Supplemental Text S5: The methods and results of other contrast networks.
Supplementary 6. Supplemental Table S1: Results of multiple linear regression analysis between network properties and confounding variables.
Supplementary 7. Supplemental Table S2: All regions of interest (abbreviations and full names).
Supplementary 8. Supplemental Table S3: Number of frequent subgraph edges.
Supplementary 9. Supplemental Figure S1: Sliding window.
Supplementary 10. Supplemental Figure S2: Discrimination of different subgraphs.
Supplementary 11. Supplemental Figure S3: Minimum spanning tree functional connectivities and degree of the corresponding node.
Supplementary 12. Supplemental Figure S4: Subgraphs and connected patterns in HC and MDD groups.
Supplementary 13. Supplemental Figure S5: Discriminative brain regions and corresponding degree.
- Y. Liu, M. Liang, Y. Zhou et al., “Disrupted small-world networks in schizophrenia,” Brain, vol. 131, no. 4, pp. 945–961, 2008.
- M.-E. Lynall, D. S. Bassett, R. Kerwin et al., “Functional connectivity and brain networks in schizophrenia,” The Journal of Neuroscience, vol. 30, no. 28, pp. 9477–9487, 2010.
- S. Micheloyannis, E. Pachou, C. J. Stam et al., “Small-world networks and disturbed functional connectivity in schizophrenia,” Schizophrenia Research, vol. 87, no. 1–3, pp. 60–66, 2006.
- M. Rubinov, S. A. Knock, C. J. Stam et al., “Small-world properties of nonlinear brain activity in schizophrenia,” Human Brain Mapping, vol. 30, no. 2, pp. 403–416, 2009.
- K. Supekar, V. Menon, D. Rubin, M. Musen, and M. D. Greicius, “Network analysis of intrinsic functional brain connectivity in Alzheimer's disease,” PLoS Computational Biology, vol. 4, no. 6, Article ID e1000100, 2008.
- Y. He, Z. Chen, and A. Evans, “Structural insights into aberrant topological patterns of large-scale cortical networks in Alzheimer's disease,” The Journal of Neuroscience, vol. 28, no. 18, pp. 4756–4766, 2008.
- C. J. Stam, “Use of magnetoencephalography (MEG) to study functional brain networks in neurodegenerative disorders,” Journal of the Neurological Sciences, vol. 289, no. 1-2, pp. 128–134, 2010.
- D. E. Van et al., “Long-term effects of temporal lobe epilepsy on local neural networks: a graph theoretical analysis of corticography recordings,” Plos One, vol. 4, no. 11, 2009.
- S. Pieper et al., “Network-level analysis of cortical thickness of the epileptic brain,” Neuroimage, vol. 52, no. 4, pp. 1302–1313, 2010.
- M.-T. Horstmann, S. Bialonski, N. Noennig et al., “State dependent properties of epileptic brain networks: Comparative graph-theoretical analyses of simultaneously recorded EEG and MEG,” Clinical Neurophysiology, vol. 121, no. 2, pp. 172–185, 2010.
- L. Wang, C. Zhu, Y. He et al., “Altered small-world brain functional networks in children with attention-deficit/hyperactivity disorder,” Human Brain Mapping, vol. 30, no. 2, pp. 638–649, 2009.
- F. D. V. Fallani, L. Astolfi, F. Cincotti et al., “Evaluation of the brain network organization from EEG signals: a preliminary evidence in stroke patient,” Anatomical Record Advances in Integrative Anatomy Evolutionary Biology, vol. 292, no. 12, pp. 2023–2031, 2009.
- J. Wang, L. Wang, and Y. Zang, “Parcellation-dependent small-world brain functional networks: a resting-state fMRi study,” Human Brain Mapping, vol. 30, no. 5, pp. 1511–1523, 2009.
- X. Chen, H. Zhang, Y. Gao, C.-Y. Wee, G. Li, and D. Shen, “High-order resting-state functional connectivity network for MCI classification,” Human Brain Mapping, vol. 37, no. 9, pp. 3282–3296, 2016.
- E. Damaraju, E. A. Allen, A. Belger et al., “Dynamic functional connectivity analysis reveals transient states of dysconnectivity in schizophrenia,” NeuroImage: Clinical, vol. 5, pp. 298–308, 2014.
- E. A. Allen, E. Damaraju, S. M. Plis, E. B. Erhardt, T. Eichele, and V. D. Calhoun, “Tracking whole-brain connectivity dynamics in the resting state,” Cerebral Cortex, vol. 24, no. 3, pp. 663–676, 2014.
- H. Zhang, X. Chen, F. Shi et al., “Topographical Information-Based High-Order Functional Connectivity and Its Application in Abnormality Detection for Mild Cognitive Impairment,” Journal of Alzheimer's Disease, vol. 54, no. 3, pp. 1095–1112, 2016.
- R. M. Hutchison, T. Womelsdorf, E. A. Allen et al., “Dynamic functional connectivity: promise, issues, and interpretations,” NeuroImage, vol. 80, pp. 360–378, 2013.
- C. Vikas., “Minimum Spanning Tree Algorithm,” International Journal of Computer Applications, vol. 1, no. 8, pp. 39–45, 2010.
- U. Lee, S. Kim, and K.-Y. Jung, “Classification of epilepsy types through global network analysis of scalp electroencephalograms,” Physical Review E: Statistical, Nonlinear, and Soft Matter Physics, vol. 73, no. 4, Article ID 041920, 2006.
- V. D. Edwin et al., “Epilepsy surgery outcome and functional network alterations in longitudinal MEG: a minimum spanning tree analysis,” Neuroimage, vol. 86, no. 1, pp. 354–363, 2014.
- P. Tewarie, A. Hillebrand, M. M. Schoonheim et al., “Functional brain network analysis using minimum spanning trees in Multiple Sclerosis: an MEG source-space study,” NeuroImage, vol. 88, pp. 308–318, 2014.
- C. McDiarmid, T. Johnson, and H. S. Stone, “On finding a minimum spanning tree in a network with random weights,” Random Structures & Algorithms, vol. 10, no. 1-2, pp. 187–204, 1997.
- P. Tewarie, E. Van Dellen, A. Hillebrand, and C. J. Stam, “The minimum spanning tree: an unbiased method for brain network analysis,” NeuroImage, vol. 104, pp. 177–188, 2015.
- F. Fei, B. Jie, and D. Zhang, “Frequent and discriminative subnetwork mining for mild cognitive impairment classification,” Brain Connectivity, vol. 4, no. 5, pp. 347–360, 2014.
- J. Du, L. Wang, B. Jie, and D. Zhang, “Network-based classification of ADHD patients using discriminative subnetwork selection and graph kernel PCA,” Computerized Medical Imaging and Graphics, vol. 52, pp. 82–88, 2016.
- L. Wang, F. Fei, B. Jie, and D. Zhang, “Combining multiple network features for mild cognitive impairment classification,” in Proceedings of the 14th IEEE International Conference on Data Mining Workshops, ICDMW, pp. 996–1003, 2014.
- Y. Li, S. Dou, E. Wang et al., “Evaluation of brain default network fMRI of insomnia with depression patients at resting state,” Life Science Journal, vol. 11, no. 8, pp. 794–801, 2014.
- I. M. Veer, C. F. Beckmann, M.-J. van Tol et al., “Whole brain resting-state analysis reveals decreased functional connectivity in major depression,” Frontiers in Systems Neuroscience, vol. 4, article 41, 2009.
- M. D. Greicius, B. H. Flores, and V. Menon, “Resting-state functional connectivity in major depressionn: abnormally increased contributions from subgenual cingulate cortex and thalamus,” Biological Psychiatry, vol. 62, no. 5, pp. 429–437, 2007.
- A. Anand, Y. Li, Y. Wang et al., “Activity and connectivity of brain mood regulating circuit in depression: a functional magnetic resonance study,” Biological Psychiatry, vol. 57, no. 10, pp. 1079–1088, 2005.
- H. Tao, S. Guo, T. Ge et al., “Depression uncouples brain hate circuit,” Molecular Psychiatry, vol. 18, no. 1, pp. 101–111, 2013.
- X. Wang, Y. Ren, Y. Yang, W. Zhang, and N. N. Xiong, “A weighted discriminative dictionary learning method for depression disorder classification using fMRI data,” in Proceedings of the IEEE International Conferences on Big Data and Cloud Computing, pp. 618–623, IEEE, October 2016.
- Y. Yue, Y. Yuan, Z. Hou, W. Jiang, F. Bai, and Z. Zhang, “2129 – Abnormal functional connectivity of amygdala in late onset depression was associated with cognitive deficits, but not with depressive severity,” European Psychiatry, vol. 28, p. 1, 2013.
- Y. I. Sheline, J. L. Price, Z. Yan, and M. A. Mintun, “Resting-state functional MRI in depression unmasks increased connectivity between networks via the dorsal nexus,” Proceedings of the National Acadamy of Sciences of the United States of America, vol. 107, no. 24, pp. 11020–11025, 2010.
- H. Guo, X. Cao, Z. Liu, and J. Chen, “Abnormal functional brain network metrics for machine learning classifier in depression patients identification,” Research Journal of Applied Sciences, Engineering & Technology, vol. 5, no. 10, pp. 3015–3020, 2013.
- L. Qiao, H. Zhang, M. Kim, S. Teng, L. Zhang, and D. Shen, “Estimating functional brain networks by incorporating a modularity prior,” NeuroImage, vol. 141, pp. 399–407, 2016.
- M.-L. Wong, C. Dong, V. Andreev, M. Arcos-Burgos, and J. Licinio, “Prediction of susceptibility to major depression by a model of interactions of multiple functional genetic variants and environmental factors,” Molecular Psychiatry, vol. 17, no. 6, pp. 624–633, 2012.
- F. Liu, W. Guo, and J.-P. Fouche, “Multivariate classification of social anxiety disorder using whole brain functional connectivity,” Brain Structure & Function, vol. 220, no. 1, pp. 101–115, 2015.
- N. Tzouriomazoyer et al., “Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain,” Neuroimage, vol. 15, no. 1, p. 273, 2002.
- J. Kruskal, “On the shortest spanning subtree of a graph and the traveling salesman problem,” Proceedings of the American Mathematical Society, vol. 7, no. 1, pp. 48–50, 1956.
- A. M. Elfeki and J. Bahrawi, “kolmogorov-smirnov test,” International Encyclopedia of Statistical Science, vol. 10, no. 1, pp. 718–720, 2014.
- A. Roberto et al., Benjamini–Hochberg False Discovery Rate (FDR) as A Function of The P-Value, 2014.
- M. Polajnar and J. Demšar, “Small network completion using frequent subnetworks,” Intelligent Data Analysis, vol. 19, no. 1, pp. 89–108, 2014.
- W. Lin, X. Xiao, and G. Ghinita, “Large-scale frequent subgraph mining in MapReduce,” in Proceedings of the 30th IEEE International Conference on Data Engineering (ICDE '14), pp. 844–855, April 2014.
- M. Kuramochi and G. Karypis, “Frequent subgraph discovery,” in Proceedings of the 1st IEEE International Conference on Data Mining (ICDM '01), pp. 313–320, Piscataway, NJ, USA, December 2001.
- A. Inokuchi, T. Washio, and H. Motoda, “An apriori-based algorithm for mining frequent substructures from graph data,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics): Preface, vol. 1910, pp. 13–23, 2000.
- M. Kuramochi and G. Karypis, “An efficient algorithm for discovering frequent subgraphs,” IEEE Transactions on Knowledge and Data Engineering, vol. 16, no. 9, pp. 1038–1051, 2004.
- S. F. Rosario and K. Thangadurai, RELIEF: Feature Selection Approach, 2015.
- X. Yan and J. Han, “gSpan: graph-based substructure pattern mining,” in Proceedings of the 2nd IEEE International Conference on Data Mining (ICDM '02), pp. 721–724, Maebashi City, Japan, December 2002.
- X. Kong et al., Discriminative Feature Selection for Uncertain Graph Classification, 2013, Discriminative Feature Selection for Uncertain Graph Classification.
- G. R. G. Lanckriet et al., “Learning the kernel matrix with semi-definite programming,” in Proceedings of the Nineteenth International Conference, 2002.
- N. Shervashidze, P. Schweitzer, E. J. van Leeuwen, K. Mehlhorn, and K. M. Borgwardt, “Weisfeiler-Lehman graph kernels,” Journal of Machine Learning Research (JMLR), vol. 12, no. 3, pp. 2539–2561, 2011.
- C. Ma, J. Ding, J. Li et al., “Resting-state functional connectivity bias of middle temporal Gyrus and Caudate with altered Gray matter volume in major depression,” PLoS ONE, vol. 7, no. 9, Article ID e45263, 2012.
- L. Qiu, S. Lui, W. Kuang et al., “Regional increases of cortical thickness in untreated, first-episode major depressive disorder,” Translational Psychiatry, vol. 4, article e378, 2014.
- M. D. Sacchet, M. C. Camacho, E. E. Livermore, E. A. Thomas, and I. H. Gotlib, “Accelerated aging of the putamen in patients with major depressive disorder,” Journal of Psychiatry & Neuroscience, vol. 42, no. 3, pp. 164–171, 2017.
- J. Jung, J. Kang, E. Won et al., “Impact of lingual gyrus volume on antidepressant response and neurocognitive functions in Major Depressive Disorder: A voxel-based morphometry study,” Journal of Affective Disorders, vol. 169, pp. 179–187, 2014.
- F. Junfang, W. Qian, and W. Bin, “Amplitude of low-frequency oscillations in major depression as revealed by resting state functional magnetic resonance imaging,” Journal of Clinical Radiology, 2015.
- D. Cotter, D. Mackay, G. Chana, C. Beasley, S. Landau, and I. P. Everall, “Reduced neuronal size and glial cell density in area 9 of the dorsolateral prefrontal cortex in subjects with major depressive disorder,” Cerebral Cortex, vol. 12, no. 4, pp. 386–394, 2002.
- Y.-T. Chen, M.-W. Huang, I.-C. Hung, H.-Y. Lane, and C.-J. Hou, “Right and left amygdalae activation in patients with major depression receiving antidepressant treatment, as revealed by fMRI,” Behavioral and Brain Functions, vol. 10, article 101, 2014.
- H. J. Rachel, Differences between Healthy Controls and remitted Major Depression for left amygdala seed, 2014.
- M. Ye, T. Yang, P. Qing, X. Lei, J. Qiu, and G. Liu, “Changes of functional brain networks in major depressive disorder: A graph theoretical analysis of resting-state fMRI,” PLoS ONE, vol. 10, no. 9, Article ID e0133775, 2015.
- C. Meng, F. Brandl, M. Tahmasian et al., “Aberrant topology of striatum's connectivity is associated with the number of episodes in depression,” Brain, vol. 137, no. 2, pp. 598–609, 2014.
- X. Liu, J. Liu, Z. Feng, X. Xu, and J. Tang, “Mass classification in mammogram with semi-supervised relief based feature selection,” in Proceedings of the 5th International Conference on Graphic and Image Processing, ICGIP 2013, China, October 2013.
Copyright © 2017 Hao Guo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.