Translational Research in Complementary and Alternative Medicine
View this Special IssueResearch Article  Open Access
Tianlu Chen, Yu Cao, Yinan Zhang, Jiajian Liu, Yuqian Bao, Congrong Wang, Weiping Jia, Aihua Zhao, "Random Forest in Clinical Metabolomics for Phenotypic Discrimination and Biomarker Selection", EvidenceBased Complementary and Alternative Medicine, vol. 2013, Article ID 298183, 11 pages, 2013. https://doi.org/10.1155/2013/298183
Random Forest in Clinical Metabolomics for Phenotypic Discrimination and Biomarker Selection
Abstract
Metabolomic data analysis becomes increasingly challenging when dealing with clinical samples with diverse demographic and genetic backgrounds and various pathological conditions or treatments. Although many classification tools, such as projection to latent structures (PLS), support vector machine (SVM), linear discriminant analysis (LDA), and random forest (RF), have been successfully used in metabolomics, their performance including strengths and limitations in clinical data analysis has not been clear to researchers due to the lack of systematic evaluation of these tools. In this paper we comparatively evaluated the four classifiers, PLS, SVM, LDA, and RF, in the analysis of clinical metabolomic data derived from gas chromatography mass spectrometry platform of healthy subjects and patients diagnosed with colorectal cancer, where crossvalidation, plot, receiver operating characteristic curve, variable reduction, and Pearson correlation were performed. RF outperforms the other three classifiers in the given clinical data sets, highlighting its comparative advantages as a suitable classification and biomarker selection tool for clinical metabolomic data analysis.
1. Introduction
Metabolomics [1] or metabonomics [2, 3] is an emergingomics approach using nuclear magnetic resonance (NMR) spectroscopy or gas chromatography/liquid chromatographymass spectrometry (GCMS or LCMS) technologies. It constitutes a field of science that deals with the measurement of metabolite variations in a biological compartment for the study of the physiological processes in response to xenobiotic interventions that is complementary to organspecific biochemical and histological findings. Through the analysis of one or several kinds of biofluids including serum, urine, saliva, and tissue samples, the global and dynamic alterations in metabolism can be deciphered [4]. Therefore, metabolomics has been increasingly used in many applications such as identifying metabolite markers for clinical diagnosis and prognosis [5], monitoring the chemicalinduced toxicity [6], exploring the potential mechanism of diverse diseases [7], and assessing therapeutic effects of treatment modalities [8, 9]. Univariate and/or multivariate statistical methods are routinely used in metabolomics studies, aiming at successful classification of samples with metabolic phenotypic variations and identification of potential biomarkers while minimizing the technical variations.
To date, the most widely used classification methods in metabolomic data processing include principal component analysis (PCA), projection to latent structures (PLS) analysis, support vector machine (SVM), Linear discriminant analysis (LDA), and univariate statistical analysis such as Student's ttest and analysis of variance (ANOVA) test [10, 11]. We recently applied some of these methods in combination to identify metabolitebased biomarkers in hepatocellular carcinoma [5], gastric cardia cancer [12], knee osteoarthritis [13], oral cancer [14], and schizophrenia [7]. Nevertheless, more effective and robust bioinformatics tools are in critical need for metabolomic data analysis especially when dealing with clinical samples with large individual variability due to diverse demographic and genetic background of patients and various pathological conditions or treatments.
A machine learning method, random forest (RF), is reported as an excellent classifier with the following advantages: simple theory, fast speed, stable and insensitive to noise, little or no overfitting, and automatic compensation mechanism on biased sample numbers of groups [15]. RF has been widely used in microarray [16–18] and single nucleotide polymorphism (SNP) [19] data analysis achieving good performance. However, in the field of clinical metabolomic data analysis, it has not got enough attention and concern. In addition, no comprehensive performance evaluation about this classifier is reported.
In this research, RF was used in the analysis of a GCMS derived clinical metabolomic dataset. Its classification and biomarker selection performances were compared with PLS, LDA, and SVM comprehensively. The score plot based on cross validation was used for classification accuracy evaluation. The crossvalidation and ROC (receiver operating characteristic) curve were carried out to test their prediction ability and stability. The plot was adopted for overfitting measurement. Variable number dependence of the 4 classifiers was explored by eliminating variables step by step. Besides these classification performances, the variable ranking and putative biomarker selection power of RF was examined as well by Pearson correlation.
2. Methods
2.1. Metabolomic Data Set
Colorectal cancer (CRC) is one of the common types of cancer and the leading causes of cancer death in the world [20]. Urinary samples of 67 CRC patients (67 preoperation samples and 63 matched postoperation ones) and 65 healthy volunteers were collected from the Cancer Hospital affiliated to Fudan University (Shanghai, China). Patients enrolled in this study were not on any medication before preoperative sample collection. The postoperative samples were collected on the 7 day after surgery. Sample collection protocol was approved by the Cancer Hospital Institutional Review Board and written consents were signed by all participants prior to the study. The healthy volunteers were selected by a routine physical examination, and any subjects with inflammatory conditions or gastrointestinal tract disorders were excluded. Other background information such as diet and alcohol consumption was considered during sample selection to minimize the dietinduced metabolic variations. All the samples were collected in the morning before breakfast, immediately centrifuged, and stored at −80°C until analysis. Clinical characteristics of all the samples in this study are provided in Table 1. All the samples were chemically derivatized and subsequently analyzed by GCMS following our previously published procedures [21].

The acquired MS data were pretreated and processed according to our previously published protocols [5, 7]. A total of 187 variables (areas of peaks, denoting concentrations of metabolites), 35 metabolites were obtained from the spectral data analysis. Normalization (to the total intensity to compensate for the overall variability during sample extraction, injection, detection, and disparity of urine volume), mean centering, and unit variance scaling of the data sets were performed prior to statistical analysis. Finally, the data set contains 187 variables and 195 samples. Two cases: (a) Normal versus CRC patients (preoperative) and (b) Preoperative versus postoperative patients were involved for analysis.
2.2. Random Forest
Random forest (RF), developed by Breiman [22], is a combination of treestructured predictors (decision trees). Each tree is constructed via a tree classification algorithm and casts a unit vote for the most popular class based on a bootstrap sampling (random sampling with replacement) of the data. The simplest random forest with random features is formed by selecting randomly, at each node, a small group of input variables to split on. The size of the group is fixed throughout the process of growing the forest. Each tree is grown by using the CART (classification and regression tree) methodology without pruning. The tree number of the forest in this study is set to be 200, the number of input variables tried for each node is the square root of the number of total variables, and the minimum size of the terminal nodes is set to be 2. The “score” of RF is the scaled sum of votes derived from the trained trees for outofbag samples.
RF includes two methods for measuring the importance of a variable or how much it contributes to predictive accuracy. The default method is the Gini score (the method of this study). For any variable, the measure is the increase in prediction error if the values of that variable are permuted across the outofbag observations. This measure is computed for every tree, then averaged over the entire ensemble, and divided by the standard deviation over the entire ensemble. Therefore, the larger the Gini score is (ranges from 1 to 100), the more important a variable is.
Please refer to the appendices for the introduction of other classifiers (PLS, SVM, and LDA).
2.3. Evaluation of Classification Performance
The classification performance of RF as well as PLS, LDA, and SVM can be evaluated and compared using several approaches: crossvalidation, plot, ROC, and reduction of variable number.
2.3.1. CrossValidation: Prediction Ability and Stability
Two types of crossvalidations: kfold and hold out were employed to estimate the prediction ability with low bias and low variance. (1) In the kfold crossvalidation, the training set was first divided into k subsets (the folds) of approximately equal size. Sequentially each subset was tested using the classifier trained on the remaining k−1 subsets, where k was set to be 7 and 10 in this study. (2) Holdout crossvalidation is similar to kfold crossvalidation except for the repeatedly (100 times) random selection of the two mutually exclusive training and testing (holdout) subsets in accordance with a given ratio. This method was used with an understanding that the more instances left for the holdout set, the higher the bias of the estimate would be. On the other hand, fewer holdout set instances mean that the confidence interval for the accuracy would be wider. Besides accuracy and prediction ability, the repeated holdout crossvalidation was used to test the stability of a classifier. The holdout ratios were set as 10%, 15%, and 33%, respectively, on all the classifiers.
2.3.2. R^{2}/Q^{2} Plot—Overfitting
Consider
In the equations, represents total number of samples, is the predicted class (0 or 1) of the sample when all the samples are used for model building, is the actual class, is the average of the predicted class, and is the predicted class when all the samples except the sample are used for model building (leave one out crossvalidation).
The criteria for classifier validity are as follows. (1) All the and values on the permuted data set are lower than those on the actual data set. (2) The regression line of (line joining the actual point to the centroid of the cluster of permuted values) has a negative intercept value on the vertical axis.
2.3.3. Receiver Operating Characteristic (ROC): Diagnosis Potential
ROC analysis is a classic method from signal detection theory and is now commonly used in clinical research [23]. ROC of a classifier shows its performance as a tradeoff between specificity and sensitivity. Sensitivity is defined as the proportion of subjects with disease whose tests is positive, and calculated by the formula, TruePositive/(TruePositive+FalseNegative). Specificity, on the other hand, is defined as the proportion of subjects without disease whose tests is negative, and calculated in the formula, TrueNegative/(TrueNegative+FalsePositive). Typically, 1specificity is plotted on the axis and sensitivity is plotted on the vertical axis. All the predictive behavior of a classifier can be represented by the points in the ROC curve independent of class distributions or error cost [23]. The area under the ROC curve (AUC) is a statistic summary of its diagnostic potential.
2.3.4. Variable Number Dependence
Generally, too many irrelevant variables are liable to result in overfitting decisions, whereas differences between groups cannot be extracted and depicted completely if crucial variables are not concerned [24]. Variable number dependence is therefore a necessary factor for classifier performance evaluation.
To avoid bias, it is advisable to rank and eliminate variables one by one. Initially, the whole dataset is taken when a classifier is computed. Then, a list of variables in descending order relative to classification importance is established and the variable in the end is eliminated for subsequent analysis. This process is repeated until only one variable is left for classifier building. The last few variables are of great potential to be biomarkers for separating the groups.
2.4. Evaluation of Biomarker Selection Performance
Prediction ability and stability, overfitting, diagnosis potential, and variable number dependence are important aspects for a classifier. Variable ranking and biomarker selection is of the same importance in metabolomics study.
For RF, variables are ranked by Gini score, a measurement of average accuracy of all trees containing a particular variable [22]. For PLS, the conventional VIP (variable importance in projection) value is used for variable ranking. For LDA, the coefficients of variables in the discriminant function indicate their importance. As to SVM, the importance of variables is evaluated by the SVM recursive feature elimination (SVMRFE) algorithm [25].
As each classifier possesses its own algorithm for variable importance ranking with its own strength and weakness, the Pearson correlation coefficient of every two ranks was used to evaluate their consistency and the rank of ttest (by ascending order of variable values) was taken as an unbiased reference. The consistency comparison was conducted on two levels: ranks of all variables and ranks of identified metabolites.
All the metabolites were identified and verified by public libraries such as HMDB, KEGG, and/or reference standards available in our laboratory.
All the classifiers andevaluation methods were carried out using Matlab toolbox (Version 2009a, Mathworks).
3. Results and Discussion
3.1. Classification Performance
RF as well as PLS, LDA, and SVM were applied on the dataset for the two comparative cases (Figures 1(a) and 1(b)). In Figure 1(a), red and blue dots represent the normal and CRC patients, respectively. axis is the sample index and axis is the corresponding “score” of every classifier. RF achieved the best separation with no missclassified samples (the accuracy is 100%). The performance of SVM, LDA, and PLS was good, with descending accuracy of 90.9%, 87.1%, and 82.6%, respectively. Similarly, Figure 1(b) shows the separation between CRC preoperative patients (red dots) and CRC postoperative patients (blue dots). RF yielded 95.4%, a higher classification accuracy than that of LDA, SVM, and PLS, which achieved 90.8%, 80.8%, and 80.8%, respectively.
(a)
(b)
3.2. Prediction Ability, Stability, Overfitting, and Diagnostic Ability Evaluation
The accuracy of classification is crucial for a classifier, while other classification behaviors such as prediction ability, stability, degree of overfitting and diagnostic ability are of equal significance as well.
The holdout crossvalidation results (33% holdout samples, 100 times) of RF (purple), PLS (blue), LDA (red), and SVM (green) on the two cases are presented as box plots (Figure 2). The axis denotes the error rate (the smaller, the better). The purple box of RF is always the lowest and shortest in both cases. As to the other three classifiers, their performances are similar showing no significant difference. These results were validated further by more crossvalidation results listed in Table 2. As expected, the average error rates and standard deviations (S.D.s) of (1) holdout crossvalidations on 10% and 15% samples; and (2) 7and 10fold crossvalidations by PLS, SVM, and LDA are at almost the same level and are all greater than that of RF. Therefore, RF is the one with highest prediction ability and best stability among all the classifiers.

(a)
(b)
Figures 3(a) and 3(b) display the correlation between the actual variable and the permuted variable (axis) versus the values (axis) of RF (purple), PLS (blue), LDA (red), and SVM (green) on the two cases, respectively. A dot denotes and a circle represents . The straight line and dash dot line are the regression lines (lines linking the value from the actual data set and the 100 permuted ones) of and , respectively. Obviously, the values of RF on the permuted data sets are all under zero and are well lower than those on the actual data set. Consequently, RF does not over fit the data and so will give out reliable result on new samples. As to SVM, its values on the permuted data sets are slightly lower than or close to that on actual data set, although most of its values on the permuted data sets are under zero and lower than those on the actual data set. Hence, SVM is overfitted and false positive result is prone to appear. This probably caused by its dependence on kernel functions and support vectors.
(a)
(b)
The ROC curve coupled with its area under the curve (AUC) is a common method used to estimate the diagnosis potential of a classifier in clinical applications. A larger AUC indicates higher prediction ability. The ROC curves and AUC values of all the classifiers in the two cases are plotted in Figure 4. RF outperforms the others once more with the greatest AUC values ().
3.3. Variable Number Dependence
Figures 5(a) and 5(b) show the classification error rates (axis) against the number of variables (axis) involved in the two cases, respectively. It can be seen that with the decrease of variable number used for classifier building, all the curves keep stable initially, and then rise gradually. Further reduction of variables degrades the classifier performance heavily because of the shortage of useful information. The point (or short section) where the curve begins to rise correlates to the optimum number of variables for classifier building. Additionally, RF usually needs fewer variables to achieve the same error rate as the other three classifiers. In case (B), for example, when the error rate is restricted to be less than 0.18 (the red line), RF needs minimum 10 variables whereas PLS, LDA, and SVM require about 150, 45, and 125 ones at least.
(a) Normal versus CRC
(b) Pre versus post
3.4. Putative Biomarker Selection
Variable number dependence section is to evaluate whether and how much the performance of RF depends on the number of variables involved. This section is to evaluate its capability on important variable (putative biomarker) selection. The Pearson correlation matrixes of ranks from every two classifiers (including tTest) based on all variables (AB) and identified metabolites (CD) in the two cases are listed in Table 3. On the whole, RF, PLS and tTest have good consistency with each other (high Pearson correlation coefficients) regardless of whether all variables (Figure 6(a)) or identified metabolites (Figure 6(b)) are involved.
 
variable rank by the value of test. ^{ b}variable rank by the VIP value of PLS. ^{ c}variable rank by the Gini value of RF. ^{ d}variable rank by the SVMREF. ^{ e}variable rank by the LDA coefficient. ^{ f}Pearson correlation coefficient of PLS and test variable rank lists for differentiating Normal and CRC. 
(a) Person correlation
(b) Person correlation
Interestingly, in Table 3, the highest and second highest correlation coefficients are 0.794 for PLS and tTest (case A) and 0.756 for RF and PLS (case B) indicating the consistency and mutual complementarity of classifiers. All the important metabolites selected by both tTest () and PLS () could be filtered by RF () as well. Consistent with previous metabolomics study, dysregulated metabolic pathways, such as glycolysis, TCA cycle, urea cycle, pyrimidine metabolism, polyamine metabolism as well as gut microbialhost cometabolism were observed [26]. Additionally, more significant metabolites in the above pathways could be found by RF (case A) providing complementary information for CRC study. Figure 7 presents some box plots of intensity in corresponding groups.
(a)
(b)
4. Conclusion
In this study, RF was applied successfully in metabolomic data analysis for clinical phenotype discrimination and biomarker selection. Its various performances were evaluated and compared with the other three classifiers PLS, SVM, and LDA by two types of crossvalidations, plot, ROC curve, variable elimination, and Pearson correlation. RF demonstrated the best overall performance including accuracy, prediction ability, overfitting, diagnosis potential, stability, and putative biomarker selection, highlighting its potential as an excellent classification and biomarker selection tool for clinical metabolomic data analysis. PLS outperforms the others in variable ranking and putative biomarker selection whereas its classification ability is not satisfactory enough. LDA demonstrates reasonably good performance in classification but its biomarker selection ability is open to question. SVM may be slightly overfitting regardless of its good classification accuracy and diagnosis potential.
The combinational usage of multiple methods, RF, tTest, and PLS, for example, may provide more comprehensive information for a “global” understanding of the metabolomics or other “omics” data.
Appendices
A. Projection to Latent Structures (PLS)
The basic object of PLS is to find the linear (or polynomial) relationship between the superior variable (a vector indicating sample groups) and the dataset (variables). The modeling consists of simultaneous projections of both the and spaces on low dimensional hyper planes. The coordinates of the points (projection of and ) on these hyper planes constitute the elements of the matrix ( scores), ( scores), (loadings), and (weights). The analysis has the following objectives:(a)To well approximate the and spaces,(b)To maximize the correlation between and . That is, to maximize the covariance between the sample positions (from different groups) on the hyper planes.
The PLS model accomplishing these objectives can be expressed as
The model will iteratively compute one component at a time, that is: one vector derived from scores , scores , weights (or ), and loadings . The component extraction process will stop when the predictive accuracy obtained in 7fold crossvalidation ( value) begin to descend. All the components are calculated in descending order of importance. The “score” of PLS is the score of the first component which contains most information of dataset.
The formula to calculate (variable importance) can be expressed as where is the number of components extracted and is the number of variables in dataset . is the correlation between and expressed by component . is the explanation of by component . is the explanation of by all the components. The Sum of squares of all is equal to the number of variables in the model hence the average is equal to 1.
The derived from the first component is used for variable ranking. Variables with larger , larger than 1 in particular, are the most relevant for explaining (classification).
B. Support Vector Machine (SVM)
The key to the success of SVM is the kernel function which maps the data from the original space into a high dimensional (possibly infinite dimensional) feature space. By constructing a linear boundary in the feature space, the SVM produces nonlinear boundaries in the original space. Given a training sample, a maximummargin hyper plane splits a given training sample in such a way that the distance from the closest cases (support vectors) to the hyper plane () is maximized where is the weight matrix, is the dataset, and is a constant term. The “score” of SVM is computed by .
SVM Recursive Feature elimination (SVMRFE) is a wrapper approach which uses the norm of the weights to rank the variables (the larger the norm of is, the more important the corresponding variable is). Initially whole of the data is taken and a classifier is computed. The norm of is computed for each of the features and the feature with the smallest norm is eliminated. This process is repeated till all the features are ranked.
Linear kernel was used for SVM classification and feature selection. This kernel was chosen to reduce the computational complexity and eliminate the need for retuning kernel parameters for every new subset of variables. Another important advantage of choosing a linear kernel is that the norm of the weight can be directly used as a ranking criterion; however this is not possible in other kernels such as the radial basis function kernel or a sigmoid kernel.
C. LDA
LDA adopts a linear combination of variables () as a predictor to characterize or separate two or more classes of objects or events.
Different with PLS which look for linear combinations of variables to best explain both the data set and the superior variable , the criterion of is to maximize the ratio () of the variance between the classes to the variance within the classes: where is the variance between or within groups, is the mean of variables from corresponding group, and is the coefficients of variables.
The “score” of is the value of testing samples and the coefficients are used for variable ranking.
Abbreviations
GCMS:  Gas chromatography mass spectrometry 
RF:  Random forest 
LDA:  Linear discriminant analysis 
SVM:  Support vector machine 
PCA:  Principal component analysis 
PLS:  Projection to latent structures 
ROC:  Receiver operating characteristic 
CRC:  Colorectal cancer 
NMR:  Nuclear magnetic resonance 
MS:  Mass spectrometry. 
Acknowledgment
This work was financially supported by the National Basic Research Program of China (2007CB914700), National Natural Science Foundation of China Program (81170760), and the Natural Science Foundation of Shanghai, China (10ZR1414800).
References
 O. Fiehn, “Metabolomics—the link between genotypes and phenotypes,” Plant Molecular Biology, vol. 48, no. 12, pp. 155–171, 2002. View at: Publisher Site  Google Scholar
 J. K. Nicholson, J. Connelly, J. C. Lindon, and E. Holmes, “Metabonomics: a platform for studying drug toxicity and gene function,” Nature Reviews Drug Discovery, vol. 1, no. 2, pp. 153–161, 2002. View at: Publisher Site  Google Scholar
 J. K. Nicholson, J. C. Lindon, and E. Holmes, “‘Metabonomics’: understanding the metabolic responses of living systems to pathophysiological stimuli via multivariate statistical analysis of biological NMR spectroscopic data,” Xenobiotica, vol. 29, no. 11, pp. 1181–1189, 1999. View at: Google Scholar
 J. Schnabel, “Targeting tumour metabolism,” Nature Reviews Drug Discovery, vol. 9, pp. 503–504, 2010. View at: Publisher Site  Google Scholar
 T. L. Chen, G. X. Xie, X. Y. Wang et al., “Serum and urinemetabolite profiling reveals potential biomarkers of human hepatocellular carcinoma,” Molecular & Cellular Proteomics, vol. 10, pp. 1–13, 2011. View at: Google Scholar
 G. X. Xie, X. J. Zheng, X. Qi et al., “Metabonomic evaluation of melamineinduced acute renal toxicity in rats,” Journal of Proteome Research, vol. 9, no. 1, pp. 125–133, 2010. View at: Publisher Site  Google Scholar
 J. Yang, T. Chen, L. Sun et al., “Potential metabolite markers of schizophrenia,” Molecular Psychiatry, vol. 18, no. 1, pp. 67–78, 2013. View at: Google Scholar
 Y. Q. Bao, T. Zhao, X. Y. Wang et al., “Metabonomic variations in the drugtreated type 2 diabetes mellitus patients and healthy volunteers,” Journal of Proteome Research, vol. 8, no. 4, pp. 1623–1630, 2009. View at: Publisher Site  Google Scholar
 X. Wang, J. Lin, T. Chen, M. Zhou, M. Su, and W. Jia, “Metabolic profiling reveals the protective effect of diammonium glycyrrhizinate on acute hepatic injury induced by carbon tetrachloride,” Metabolomics, vol. 7, no. 2, pp. 226–236, 2010. View at: Publisher Site  Google Scholar
 J. Trygg, E. Holmes, and T. Lundstedt, “Chemometrics in metabonomics,” Journal of Proteome Research, vol. 6, no. 2, pp. 469–479, 2007. View at: Publisher Site  Google Scholar
 H. W. Cho, S. B. Kim, M. K. Jeong et al., “Discovery of metabolite features for the modelling and analysis of highresolution NMR spectra,” International Journal of Data Mining and Bioinformatics, vol. 2, no. 2, pp. 176–192, 2008. View at: Publisher Site  Google Scholar
 Z. Cai, J. Zhao, X. Wang et al., “A combined proteomics and metabolomics profiling of gastric cardia cancer reveals characteristic dysregulations in glucose metabolism,” Molecular & Cellular Proteomics, vol. 9, pp. 2617–2628, 2010. View at: Publisher Site  Google Scholar
 X. Li, S. Yang, Y. Qiu et al., “Urinary metabolomics as a potentially novel diagnostic and stratification tool for knee osteoarthritis,” Metabolomics, vol. 6, no. 1, pp. 109–118, 2010. View at: Publisher Site  Google Scholar
 J. Wei, G. X. Xie, Z. T. Zhou et al., “Salivary metabolite signatures of oral cancer and leukoplakia,” International Journal of Cancer, vol. 129, no. 9, pp. 2207–2217, 2011. View at: Google Scholar
 D. Amaratunga, J. Cabrera, and Y. S. Lee, “Enriched random forests,” Bioinformatics, vol. 24, no. 18, pp. 2010–2014, 2008. View at: Publisher Site  Google Scholar
 A. Statnikov, L. Wang, and C. F. Aliferis, “A comprehensive comparison of random forests and support vector machines for microarraybased cancer classification,” BMC Bioinformatics, vol. 9, pp. 319–328, 2008. View at: Publisher Site  Google Scholar
 X. Y. Wu, Z. Y. Wu, and K. Li, “Identification of differential gene expression for microarray data using recursive random forest,” Chinese Medical Journal, vol. 121, no. 24, pp. 2492–2496, 2008. View at: Google Scholar
 A. Acharjeea, B. Kloosterman, R. C. H. D. Vos et al., “Data integration and network reconstruction with ~omics data using Random Forest regression in potato,” Analytica Chimica Acta, vol. 705, no. 12, pp. 56–63, 2011. View at: Publisher Site  Google Scholar
 R. Jiang, W. W. Tang, X. B. Wu, and W. H. Fu, “A random forest approach to the detection of epistatic interactions in casecontrol studies,” BMC Bioinformatics, vol. 10, supplement 1, pp. 65–76, 2009. View at: Publisher Site  Google Scholar
 A. Jemal, R. Siegel, J. Xu, and E. Ward, “Cancer statistics, 2010,” CA Cancer Journal for Clinicians, vol. 60, no. 5, pp. 277–300, 2010. View at: Publisher Site  Google Scholar
 Y. Qiu, M. Su, Y. Liu et al., “Application of ethyl chloroformate derivatization for gas chromatographymass spectrometry based metabonomic profiling,” Analytica Chimica Acta, vol. 583, no. 2, pp. 277–283, 2007. View at: Publisher Site  Google Scholar
 L. Breiman, “Random forests,” Machine Learning, vol. 45, no. 1, pp. 5–32, 2001. View at: Publisher Site  Google Scholar
 A. P. Bradley, “The use of the area under the ROC curve in the evaluation of machine learning algorithms,” Pattern Recognition, vol. 30, no. 7, pp. 1145–1159, 1997. View at: Publisher Site  Google Scholar
 K. Duan, S. S. Keerthi, and A. N. Poo, “Evaluation of simple performance measures for tuning SVM hyperparameters,” Neurocomputing, vol. 51, pp. 41–59, 2003. View at: Publisher Site  Google Scholar
 I. Guyon, J. Weston, S. Barnhill, and V. Vapnik, “Gene selection for cancer classification using support vector machines,” Machine Learning, vol. 46, no. 1–3, pp. 389–422, 2002. View at: Publisher Site  Google Scholar
 Y. P. Qiu, G. X. Cai, M. M. Su et al., “Serum metabolite profiling of human colorectal cancer using GC−TOFMS and UPLC−QTOFMS,” Journal of Proteome Research, vol. 8, no. 10, pp. 4844–4850, 2009. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2013 Tianlu Chen et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.