Research Article  Open Access
Bin Gan, ChunHou Zheng, Jun Zhang, HongQiang Wang, "Sparse Representation for Tumor Classification Based on Feature Extraction Using Latent LowRank Representation", BioMed Research International, vol. 2014, Article ID 420856, 7 pages, 2014. https://doi.org/10.1155/2014/420856
Sparse Representation for Tumor Classification Based on Feature Extraction Using Latent LowRank Representation
Abstract
Accurate tumor classification is crucial to the proper treatment of cancer. To now, sparse representation (SR) has shown its great performance for tumor classification. This paper conceives a new SRbased method for tumor classification by using gene expression data. In the proposed method, we firstly use latent lowrank representation for extracting salient features and removing noise from the original samples data. Then we use sparse representation classifier (SRC) to build tumor classification model. The experimental results on several realworld data sets show that our method is more efficient and more effective than the previous classification methods including SVM, SRC, and LASSO.
1. Introduction
Tumor is a solid lesion caused by the abnormal growth of cells. A timely accurate treatment is very important clinically. The premise of an accurate treatment is an exact diagnosis due to the heterogeneity of cancer. That is, we need to classify them accurately before treating tumors. Current methods for classifying cancer malignancies mostly rely on a variety of morphological, clinical, or molecular variables. Despite recent progresses, there are still many uncertainties in diagnosis. The advent of DNA microarray and RNA_seq [1] makes it possible to analyze tumor samples and classify them based on gene expression profiles. Moreover, we can get the expression data of tens of thousands of genes through DNA microarray or RNAseq simultaneously.
Many methods for molecular data classification or clustering based on gene expression data have appeared in this area [2–14]. Huang and Zheng used independent component analysis [5] to extract features; Gao and Church introduced sparse nonnegative matrix factorization for feature extraction [4]; Zheng et al. proposed metasamplebased sparse representation [7], and Furey et al. used support vector machines [8] to classify the gene expression data. All these methods have achieved impressive classification performances.
Recently published sparse representation classification (SRC) is also a powerful tool for processing gene expression data. SRC method was inspired by many theories such as Basis pursuing [15], compressive sensing for signal reconstruction [16], and least absolute shrinkage. It has already been widely used in face recognition [17] and texture classification [18]. In SRC method, test samples can be only represented as a sparse linear combination of the training samples from the same class. Furthermore, an imposed regularized least square optimization is used to calculate an SR coefficient vector with only a few significant coefficients. In theory, a test sample can be well represented by only using the training samples from the same class. However, there is too much noise in gene expression data, which causes that the discriminative features are not obvious and the test samples can also be represented by some training samples from different classes. This will decrease the classification accuracy. To reduce noise [19–21] and get salient features [20] for tumor classification, in this paper, we introduce latent lowrank representation to preprocess gene expression data. By combining it with SRC algorithm, we propose a new method for tumor classification.
Latent lowrank representation (LatLRR) is a kind of theory which can be used to extract principal and salient features from original data. LatLRR is the improved version of LRR. The two methods can be resolved by the inexact augmented Lagrange multiplier (ALM) optimization. In [19–22], LRR has been successfully used for the recovery of subspace structure, subspace segmentation, feature extraction, outlier detection, and so forth. In [23], the author introduced LRR theory for face recognition in order to remove noise and achieved an impressive result. Based on these successful applications, in this paper, we introduce LatLRR into sparse representation classifier for tumor classification. Firstly, we use LatLRR to remove noise from original data and extract salient features. Then based on the new extracted salient features, we design sparse representation classifier to classify new test samples. We referred to the proposed method as SRCbased latent lowrank representation (SRC LatLRR).
The rest of the paper is organized as follows. Section 2 describes our proposed SRCLatLRR method in detail. We firstly review SRC and latent lowrank representation methods in Sections 2.1 and 2.2, respectively. Then we present our method in detail in Section 2.3. Section 2.4 specifies our experimental setting. In Section 3, we evaluate our method using several publicly available gene expression data sets. Section 4 concludes the paper and outlines our future work.
The abbreviations used in this paper are summarized in the Abbreviations section.
2. Methods
2.1. Sparse Representation Classification
Sparse representation classification is a supervised classification. Let denote a training sample matrix with samples and genes. As we know, each DNA microarray chip usually contains thousands of genes; the number of genes is much larger than tumor samples; that is, .
Let be the th sample of and the samples are divided into object classes. Assuming that there are samples belonging to th class and making up , the whole data set can be reexpressed as . Suppose that a new testing sample belongs to th class. Based on the theory of sparse representation, would lie in the linear span of the training samples ; that is, where is a scalar and .
Supposing a linear representation coefficient vector , can be also rewritten as Ideally, if the training samples are sufficient and the training samples sets that belong to different class are disjoint each other, then we have that is, in , only the entries corresponding to the same class as are nonzero.
From the above analysis, it can be seen that we can classify the test sample according to . So the key problem is how to calculate in (2). As in [7], would be sparse if the number of object classes is large; this is what sparse representation implies. According to the theory of compressive sensing [16, 24–26] and SR, can be achieved by solving the following minimization problem:
This problem can be solved by standard linear programming methods [15]. But (4) has no exact solutions since . Then a generalized version of (4) can be conceived: where is a scalar regularization. This function can balance the degree of noise by using . In this study, we solve this function by the truncated Newton interiorpoint method [27].
2.2. Latent LowRank Representation
Latent lowrank representation is an extension of lowrank representation. Consider an observed data matrix , where each column vector is a sample, and a dictionary , where is also a sample. can be linearly represented by the dictionary. That is, where is a coefficient matrix and each is the representation of . Equation (6) means that each column vector of can be represented by a linear combination of the bases in . In (6), the dictionary should be overcomplete enough to represent any observed data matrix . But meanwhile, this causes multiple feasible solutions of to (6). To achieve the optimal solution, low rankness criterion is introduced to (6):
Here, the optimal solution is the socalled lowestrank representation of data with respect to the dictionary . Unfortunately, function (7) can not be easy to solve because of the discrete nature of the rank function. By matrix completion method [28–30], we replace solving lowrank problem with dealing with nuclear norm [31]; then problem (7) can be rerepresented as where means the nuclear norm of matrix , that is, the sum of the singular values of matrix .
Strictly speaking, the dictionary should be overcomplete and noiseless. But this kind of dictionary is difficult to get. In practice, we usually use observed data matrix itself as the dictionary [19, 21, 32]. Finally we have the following convex optimization problem:
To solve this equation, two conditions need to be met. Firstly, the data sampling should be sufficient. Secondly, the data sampling should also contain sufficient noiseless data to achieve robust capability. In fact, the first one can be easily met but the second one not. Because gene expression data are usually noisy, in reality, function (9) may be invalid and not robust.
To solve the problem in (9), we introduce the following LRR problem [20]: where is the observed data matrix and the is the unobserved data, that is, the hidden data. We use the concatenation of and as a dictionary. The optimal result of (10) is , where and correspond to and , respectively.
By solving (10), the two problems above can be solved well. Then our next mission is to recover the affinity matrix by using only in the absence of the hidden data . The method is called latent lowrank representation (LatLRR), which is an improvement of LRR.
Supposing we have two matrices and , then by solving (10) we have the following equations: where and can be obtained through computing the skinny singular value decomposition of , and . Namely, and .
Depending on function (11), we have Let ; then we have the following simple function: If and come from the same collection of lowrank subspaces, then both and should be of lowrank, so we can achieve
Just as in [28–30], we also change the above rank minimization problem to the nuclear norm. Then we have the following convex optimization problem:
Here, we replace , , and with , , and , respectively, for ease of representation. In (15), is the noiseless observed data. By considering there may exist corrupted data or noise in , we also need to introduce a denoising model about (15); then we have where is a scalar and is the norm of sparse noise matrix . If , the problem (16) will be equivalent to (15), that is, no noise in the observed data . In (16), the optimal solutions , , and represent the principal features, salient features, and noise, respectively.
To solve the LatLRR problem listed in (16), we introduce the augmented Lagrange multiplier (ALM) [33] method and revise (16) as follows to meet the requirement of ALM algorithm:
This problem can be solved by ALM method which minimizes the following augmented Lagrange function: where and denote the trace and Frobenius norm of a matrix, respectively. is a penalty parameter. More details about (18) can be found in [33].
2.3. Sparse Representation Classification Based on LatLRR
Since LatLRR can extract the salient features and remove noise from original data sets, in this study, before using observed data for classification, we firstly use LatLRR to suppress noise and get the salient features. Then we use the denoised data for tumor classification; that is, we factorize the observed data into
Here, we only use for data classification. For a test sample , we can calculate its SR by the following function: where the parameter can be determined experimentally and is a coefficient vector. Assuming the test sample belongs to one of target classes, the training data set is sufficient. When classifying , we introduce , where is a square matrix obtained through LatLRR method when extracting the salient features.
Ideally, can be linearly represented by the samples from the same class in . Namely, the representation vector should be sparse and the nonzero entries are associated with the columns of from the same class. This will lead us to classify the test samples. However, noise and modeling errors will also introduce some nonzero entries to which correspond to the columns of from the multiple classes [17]. To solve this problem, we classify based on how well it can be reconstructed by using the coefficients from each class as in [17].
Using the result of (20), we construct as the characteristic function which selects the coefficients associated with the th class in the coefficient vector . By only using th class coefficients to reconstruct the test sample as , we can classify into the minimum residual class between and ; that is, Our classification algorithm can be summarized as follows. Input. Observed data for classes; test sample . Step 1. Normalize the columns of . Step 2. Extract the salient features of and remove to some extent noise to get data defined in (19). Step 3. Solve the optimization problem defined in (20). Step 4. Compute the residuals . Output. .
Our method can be seen as the combination of SRC [17] and latent lowrank representation for feature extraction [20], so we named it as SRCLatLRR. In SRC, the test sample is represented as a sparse linear combination of the training samples from the same class. In LatLRR, noise is removed to some extent and salient features are simultaneously extracted from the training samples. So the introduction of LatLRR can improve the classification accuracy of SRC in a way.
2.4. Evaluation of the Performance
To evaluate our proposed method, we compare our method with SRC [17, 34], LASSO [35], and SVM [8, 36, 37]. SVM has been proved to be one of the best classifiers for classifying data in the area of “high dimensionality and small sample size” [36, 37]. We do binary classification and multiclass classification experiments in Sections 3.1 and 3.2, respectively. During the experiment, the best results of SRC, LASSO, and SVM are also used to compare with those of our method, which were achieved by choosing appropriate parameters experimentally. As the number of tumor sample is too small, we use stratified 10fold cross validation in all our experiments. In the multiclass classification experiments, we do not use LASSO method because it is designed only for binary class classification problems [35]. As we know, dimensionality reduction can improve the classification performance and computing speed, so we reduce data dimensionality using betweencategory to withincategory sums of squares methods in our experiments.
3. Experimental Results
3.1. TwoClass Classification Problem
In this subsection, three twoclass microarray data sets are used to evaluate our method: colon cancer [38], prostate cancer [39], and diffuse large Bcell lymphoma [40].
The colon data set contains 62 samples consisting of 40 tumor and 22 normal. The prostate data set contains prostate tumors and normal prostate samples, each consisting of the expression levels of 12600 genes. For the DLBCL data set, the gene expression values were measured by highdensity oligonucleotide microarrays. An overview of the three data sets is given in Table 1.

The classification results by using SVM, LASSO, SRC, and the proposed SRCLatLRR are listed in Table 2. From Table 2, we can see that our method SRCLatLRR performs well on all the three data sets. Even the performance of SRCLatLRR is not better than SRC on the prostate cancer data set, but it is better than SVM and LASSO. In summary, SRC has an advantage for the prostate cancer and DLBCL data sets, but SRCLatLRR is the best classifier for the colon cancer and DLBCL data sets.

To further evaluate our method, in this experiment, we also introduced BW feature selection in our method to classify these three data sets. The results are listed in Table 3, and the number of genes selected is given in the parenthesis behind data set. From Table 3, we can see that after feature selection, our proposed classification method outperforms the other three classification methods, and it can even achieve an accuracy of 100% for the DLBCL data set.

3.2. Multiclass Classification Problem
In this subsection, we use four multiclass data sets to further check the classification performance of SRCLatLRR. The four data sets are lung cancer [41], leukemia [42], 11_tumors [43], and 9_tumors [44].
In lung cancer data set, there are four classes of lung cancer and normal class. This data set contains 203 samples. For leukemia data set, all the samples are classified into acute myelogenous leukemia, acute lymphoblastic leukemia, or mixedlineage leukemia. The data set includes 72 samples with 11225 genes. For 11_tumors, there are 11 classes of samples, which are ovary, bladder/ureter, breast, colorectal, gastroesophagus, kidney, liver, prostate, pancreas, adeno lung, and squamous lung. This data set includes 174 samples. For the 9_tumors data set, there are 60 samples with 5726 genes. These 9 types of tumors are nonsmallcell lung, colon, breast, ovarian, leukemia, renal, melanoma, prostate, and central nervous system. The detailed descriptions about these four data sets are listed in Table 4. All the four data sets were produced by oligonucleotide microarrays and the analysis tool Affymetrix GENECHIP [36].

The experimental results are listed in Table 5. From these results, we can see that the proposed method SRCLatLRR does not have a clear advantage over SVM and SRC. The reason may be that in these data sets, the training samples of each class are very few so that the sample space is not complete.

We then introduced BW feature selection before applying our method. The obtained results are listed in Table 6. From the results we can see that the proposed method classified leukemia well. For the other data sets, it has no clear advantage. But it performed better than SRC for all the four data sets.

3.3. The Choice of the Balanced Parameter
In this section, we use the data sets described in Section 3.1 to check how in (16) affect the classification performance. We show the accuracies and the removed noise level by our method at different values of in Figures 1, 2, and 3 for the colon, prostate, and DLBCL data sets, respectively. From (16), we know that the lower the is, the bigger the noise level is removed. For these three figures we use to represent the level of the removed noise. From these three figures we can see that the noise that we remove from the original data can not be too much, or it will reduce the accuracy. The reason is that if is set to be too small, useful information may be also removed besides noise. On the contrary, if is too big, the noise that was removed is too little, and we still can not get a good classification result. The experiment suggests that for colon data sets, is the best choice and and for the prostate and DLBCL data sets, respectively.
4. Conclusions
For gene expression data, cancer diagnosis is one of the most important clinical applications. In this paper, we have proposed a new SRbased method for tumor classification which uses the noiseless salient features extracted from the original samples to classify a test sample. We compared our method with several stateoftheart methods including SVM, LASSO, and SRC on seven data sets. The results of experiments show that the proposed method is better than SVM, LASSO, and SRC in a way. These demonstrate that SRCLatLRR is effective and efficient for tumor classification. We also introduced gene selection into our method. The results show that gene selection can improve the classification accuracy to some extent.
During the study we also found that, for the optimal result of LatLRR on the observed samples, represents the affinity matrix of samples [21]. In theory, the affinity matrix can be used to cluster samples. In future, we will extend it to investigate the property of sample clusters.
Abbreviations
SR:  Sparse representation 
SRC:  Sparse representation classification 
LRR:  Lowrank representation 
LatLRR:  Latent lowrank representation 
ALM:  Augmented Lagrange multiplier 
SVM:  Support vector machines 
LASSO:  Least absolute shrinkage and selection operator 
BW:  Betweencategories to withincategory. 
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work was supported by the National Science Foundation of China under Grant nos. 61272339 and 61271098, and 61374181, the Natural Science Foundation of Anhui Province under Grant no. 1308085MF85, and the Key Project of Anhui Educational Committee, under Grant no. KJ2012A005.
References
 A. Cánovas, G. Rincon, A. IslasTrejo, S. Wickramasinghe, and J. F. Medrano, “SNP discovery in the bovine milk transcriptome using RNASeq technology,” Mammalian Genome, vol. 21, no. 1112, pp. 592–598, 2010. View at: Publisher Site  Google Scholar
 T. R. Golub, D. K. Slonim, P. Tamayo et al., “Molecular classification of cancer: class discovery and class prediction by gene expression monitoring,” Science, vol. 286, no. 5439, pp. 531–527, 1999. View at: Publisher Site  Google Scholar
 J. P. Brunet, P. Tamayo, T. R. Golub, and J. P. Mesirov, “Metagenes and molecular pattern discovery using matrix factorization,” Proceedings of the National Academy of Sciences of the United States of America, vol. 101, no. 12, pp. 4164–4169, 2004. View at: Publisher Site  Google Scholar
 Y. Gao and G. Church, “Improving molecular cancer class discovery through sparse nonnegative matrix factorization,” Bioinformatics, vol. 21, no. 21, pp. 3970–3975, 2005. View at: Publisher Site  Google Scholar
 D. S. Huang and C. H. Zheng, “Independent component analysisbased penalized discriminant method for tumor classification using gene expression data,” Bioinformatics, vol. 22, no. 15, pp. 1855–1862, 2006. View at: Publisher Site  Google Scholar
 T. K. Paul and H. Iba, “Prediction of cancer Class with majority voting genetic programming classifier using gene expression data,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 6, no. 2, pp. 353–367, 2009. View at: Publisher Site  Google Scholar
 C. H. Zheng, L. Zhang, T. Y. Ng, C. K. Shiu, and D. S. Huang, “Metasamplebased sparse representation for tumor classification,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 8, no. 5, pp. 1273–1282, 2011. View at: Publisher Site  Google Scholar
 T. S. Furey, N. Cristianini, N. Duffy, D. W. Bednarski, M. Schummer, and D. Haussler, “Support vector machine classification and validation of cancer tissue samples using microarray expression data,” Bioinformatics, vol. 16, no. 10, pp. 906–914, 2000. View at: Google Scholar
 Y. S. Lee, A. Krishnan, Q. Zhu, and O. G. Troyanskaya, “Ontologyaware classification of tissue and celltype signals in gene expression profiles across platforms and technologies,” Bioinformatics, vol. 29, no. 23, pp. 3036–3044, 2013. View at: Google Scholar
 M. Tanic, E. Andress, S. M. RodriguezPinilla et al., “MicroRNAbased molecular classification of nonBRCA1/2 hereditary breast tumours,” British Journal of Cancer, vol. 109, no. 10, pp. 2724–2734, 2013. View at: Google Scholar
 J. H. Huang, H. L. Xie, J. Yan, H. M. Lu, Q. S. Xu, and Y. Z. Liang, “Rsing random forest to classify Tcell epitopes based on amino acid properties and molecular features,” Analytica Chimica Acta, vol. 804, pp. 70–75, 2013. View at: Google Scholar
 L. Nanni, S. Brahnam, S. Ghidoni, E. Menegatti, and T. Barrier, “Acomparison of methods for extracting information from the cooccurrence matrix for subcellular classification,” Expert Systems with Applications, vol. 40, no. 18, pp. 7457–7467, 2013. View at: Google Scholar
 G. R. Lioyd, L. M. Almond, N. Stone et al., “Utilising nonconsensus pathology measurements to improve the diagnosis of oesophageal cancer using a raman spectroscopic probe,” The Analyst, vol. 139, no. 2, pp. 381–388, 2014. View at: Publisher Site  Google Scholar
 G. Braz, S. V. da Rocha, M. Gattass, A. C. Silva, and A. C. de Paiva, “A mass classification using spatial diversity approaches in mammography images for false positive reduction,” Expert Systems with Applications, vol. 40, no. 18, pp. 7534–7543, 2013. View at: Google Scholar
 S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition by basis pursuit,” SIAM Review, vol. 43, no. 1, pp. 129–159, 2001. View at: Publisher Site  Google Scholar
 D. L. Donoho, “Compressed sensing,” IEEE Transactions on Information Theory, vol. 52, no. 4, pp. 1289–1306, 2006. View at: Publisher Site  Google Scholar
 J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust face recognition via sparse representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 2, pp. 210–227, 2009. View at: Publisher Site  Google Scholar
 J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, “Supervised dictionary learning,” in Proceedings of the 22nd Annual Conference on Neural Information Processing Systems (NIPS '09), 2009. View at: Google Scholar
 G. C. Liu, Z. C. Lin, S. C. Yan, J. Sun, Y. Yu, and Y. Ma, “Robust recovery of subspace structures by lowrank representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 1, pp. 171–184, 2013. View at: Google Scholar
 G. Liu and S. C. Yan, “Latent lowrank representation for subspace segmentation and feature extraction,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV '11), pp. 1615–1622, Barcelona, Spain, November 2011. View at: Publisher Site  Google Scholar
 G. C. Liu, Z. C. Lin, and Y. Yu, “Robust subspace segmentation by lowrank representation,” in Proceedings of the 27th International Conference on Machine Learning (ICML '10), pp. 663–670, 2010. View at: Google Scholar
 G. C. Liu, H. Xu, and S. C. Yan, “Exact subspace segmentation and outlier detection by lowrank representation,” JMLR: Workshop and Conference Proceedings, vol. 22, pp. 703–711, 2012. View at: Google Scholar
 L. Ma, C. H. Wang, B. H. Xiao, and W. Zhou, “Sparse representation for face recognition based on discriminative lowrank dictionary learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '12), pp. 2586–2593, June 2012. View at: Google Scholar
 E. J. Candès, J. Romberg, and T. Tao, “Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information,” IEEE Transactions on Information Theory, vol. 52, no. 2, pp. 489–509, 2006. View at: Publisher Site  Google Scholar
 E. J. Candes and T. Tao, “Nearoptimal signal recovery from random projections: universal encoding strategies?” IEEE Transactions on Information Theory, vol. 52, no. 12, pp. 5406–5425, 2006. View at: Publisher Site  Google Scholar
 R. Tibshirani, “Regression shrinkage and selection via the lasso: a retrospective,” Journal of the Royal Statistical Society B, vol. 73, no. 3, pp. 273–282, 2011. View at: Publisher Site  Google Scholar
 S. J. Kim, K. Koh, M. Lustig, S. Boyd, and D. Gorinevsky, “An interiorpoint method for largescale l_{1}regularized least squares,” IEEE Journal on Selected Topics in Signal Processing, vol. 1, no. 4, pp. 606–617, 2007. View at: Publisher Site  Google Scholar
 E. J. Candes and Y. Plan, “Matrix completion with noise,” Proceedings of the IEEE, vol. 98, no. 6, pp. 925–936, 2010. View at: Publisher Site  Google Scholar
 E. J. Candès and B. Recht, “Exact matrix completion via convex optimization,” Foundations of Computational Mathematics, vol. 9, no. 6, pp. 717–772, 2009. View at: Publisher Site  Google Scholar
 R. H. Keshavan, A. Montanari, and S. Oh, “Matrix completion from noisy entries,” Journal of Machine Learning Research, vol. 11, pp. 2057–2078, 2010. View at: Google Scholar
 M. Fazel, Matrix rank minimization with applications [Ph.D. thesis], Stanford University, Stanford, Calif, USA, 2002.
 Y. Ni, J. Sun, X. Yuan, S. Yan, and L. F. Cheong, “Robust lowrank subspace segmentation with semidefinite guarantees,” in Proceedings of the 10th IEEE International Conference on Data Mining Workshops (ICDMW '10), pp. 1179–1188, Sydney, Australia, December 2010. View at: Publisher Site  Google Scholar
 Z. Lin, M. Chen, L. Wu, and Y. Ma, “The augmented lagrange multiplier method for exact recovery of corrupted lowrank matrices,” Tech. Rep. UILUENG092215, 2009. View at: Google Scholar
 X. Hang and F. X. Wu, “Sparse representation for classification of tumors using gene expression data,” Journal of Biomedicine and Biotechnology, vol. 2009, Article ID 403689, 6 pages, 2009. View at: Publisher Site  Google Scholar
 D. Ghosh and A. M. Chinnaiyan, “Classification and selection of biomarkers in genomic data using LASSO,” Journal of Biomedicine and Biotechnology, vol. 2005, no. 2, pp. 147–154, 2005. View at: Publisher Site  Google Scholar
 A. Statnikov, C. F. Aliferis, I. Tsamardinos, D. Hardin, and S. Levy, “A comprehensive evaluation of multicategory classification methods for microarray gene expression cancer diagnosis,” Bioinformatics, vol. 21, no. 5, pp. 631–643, 2005. View at: Publisher Site  Google Scholar
 N. Pochet, F. de Smet, J. A. K. Suykens, and B. L. R. de Moor, “Systematic benchmarking of microarray data classification: assessing the role of nonlinearity and dimensionality reduction,” Bioinformatics, vol. 20, no. 17, pp. 3185–3195, 2004. View at: Publisher Site  Google Scholar
 U. Alon, N. Barka, D. A. Notterman et al., “Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays,” Proceedings of the National Academy of Sciences of the United States of America, vol. 96, no. 12, pp. 6745–6750, 1999. View at: Publisher Site  Google Scholar
 D. Singh, P. G. Febbo, K. Ross et al., “Gene expression correlates of clinical prostate cancer behavior,” Cancer Cell, vol. 1, no. 2, pp. 203–209, 2002. View at: Publisher Site  Google Scholar
 M. A. Shipp, K. N. Ross, P. Tamayo et al., “Diffuse large Bcell lymphoma outcome prediction by geneexpression profiling and supervised machine learning,” Nature Medicine, vol. 8, no. 1, pp. 68–74, 2002. View at: Publisher Site  Google Scholar
 A. Bhattacharjee, W. G. Richards, J. Staunton et al., “Classification of human lung carcinomas by mRNA expression profiling reveals distinct adenocarcinoma subclasses,” Proceedings of the National Academy of Sciences of the United States of America, vol. 98, no. 24, pp. 13790–13795, 2001. View at: Publisher Site  Google Scholar
 S. A. Armstrong, J. E. Staunton, L. B. Silverman et al., “MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia,” Nature Genetics, vol. 30, no. 1, pp. 41–47, 2002. View at: Publisher Site  Google Scholar
 A. I. Su, J. B. Welsh, L. M. Sapinoso et al., “Molecular classification of human carcinomas by use of gene expression signatures,” Cancer Research, vol. 61, no. 20, pp. 7388–7393, 2001. View at: Google Scholar
 J. E. Staunton, D. K. Slonim, H. A. Coller et al., “Chemosensitivity prediction by transcriptional profiling,” Proceedings of the National Academy of Sciences of the United States of America, vol. 98, no. 19, pp. 10787–10792, 2001. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2014 Bin Gan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.