Table of Contents Author Guidelines Submit a Manuscript
BioMed Research International
Volume 2017 (2017), Article ID 1096028, 8 pages
https://doi.org/10.1155/2017/1096028
Research Article

Gene Feature Extraction Based on Nonnegative Dual Graph Regularized Latent Low-Rank Representation

School of Electrical Engineering and Automation, Jiangxi University of Science and Technology, Ganzhou 341000, China

Correspondence should be addressed to Guoliang Yang; moc.621@03gnailgy and Zhengwei Hu; moc.361@3991iewgnehzuh

Received 21 January 2017; Accepted 13 March 2017; Published 30 March 2017

Academic Editor: Gang Liu

Copyright © 2017 Guoliang Yang and Zhengwei Hu. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Aiming at the problem of gene expression profile’s high redundancy and heavy noise, a new feature extraction model based on nonnegative dual graph regularized latent low-rank representation (NNDGLLRR) is presented on the basis of latent low-rank representation (Lat-LRR). By introducing dual graph manifold regularized constraint, the NNDGLLRR can keep the internal spatial structure of the original data effectively and improve the final clustering accuracy while segmenting the subspace. The introduction of nonnegative constraints makes the computation with some sparsity, which enhances the robustness of the algorithm. Different from Lat-LRR, a new solution model is adopted to simplify the computational complexity. The experimental results show that the proposed algorithm has good feature extraction performance for the heavy redundancy and noise gene expression profile, which, compared with LRR and Lat-LRR, can achieve better clustering accuracy.

1. Introduction

With the accelerated pace of modern life, the high incidence of cancer has brought great challenges to human health. How to detect, prevent, and treat cancer effectively has become an international hotspot of medical research. Gene expression profile is a specific cDNA sequence data of cells, which can describe cells’ current physiological function and state. Researches show that tumor cells and normal cells could be identified effectively by analyzing and processing the original gene expression data. However, the scale of the gene expression profile is huge and complex due to the diversity and specificity of the cells; therefore the traditional methods of data analysis and processing have been unable to adapt to these extremely large-scale data.

Gene expression profile extracting includes two kinds of methods: linear and nonlinear. Early linear transformation methods include principal component analysis [13] (PCA), linear discriminant analysis [46] (LDA), and independent component analysis [7, 8] (ICA). The main methods of nonlinear transformation include kernel method [9], neural network [10, 11], manifold learning [12, 13], and sparse representation [14, 15]. In recent years, LRR [1618] and neural networks have been widely used in feature extraction and classification of gene expression profile. Reference [19] used NMF for gene feature extraction and achieved more satisfactory results. Ref. [20] proposed a gene expression profile classification means based on ontology perception. Ref. [21] proposed a subcellular cooccurrence matrix feature extraction method. Ref. [22] proposed a gene expression profile classification method by neural network hybrid back-propagation. Ref. [23] proposed a supervised way of tumor prediction with multiview.

The size of the gene expression profile is large, and there are interrelationships between the samples. The internal spatial structure of the data may be destroyed in the process of linear transformation. In this paper, a model of feature extraction based on NNDGLLRR is proposed on the basis of Lat-LRR, which with low-rank sparse constraint can remove the redundant components of gene expression and suppress the noise. Nonnegative constraints make the calculation with a certain degree of sparsity, in line with the practical significance of the data, and enhance the robustness of the algorithm. And the manifold regularized constraint is introduced, so that the result of feature extraction can describe the spatial structure of the original data more completely.

2. Related Work

2.1. LRR

LRR is a combination of matrix low-rank decomposition and sparse decomposition. In recent years, it has been widely used in subspace clustering. LRR assumes that the original data comes from different subspaces and performs feature extraction by trying to find the lowest rank representation of the original data. And this low-rank representation coefficient is the reflection of the original data in the spatial distribution of structural information. If the original data , each column represents a sample, and generally the LRR uses the data itself as a dictionary. Then the model can be as shown in

The LRR matrix , and is the linear representation coefficient of the sample under the data dictionary . The original data usually contains a lot of noise, while the sparse constraint can maintain the robustness of the algorithm effectively. Ref. [24] shows the specific solution process of LRR.

Let ; we construct the following Augmented Lagrangian function:

The specific update algorithm is as follows.

Keep , ; update :Keep ,  , and  ; update :Keep , ; update :

2.2. Lat-LRR

LRR has two conditions; one is that the original data contains enough samples, and the other is that contains enough nonpolluting data. However, these two conditions are almost impossible to achieve for gene data. On the one hand, the available number of gene samples for research is small because of the high prices of gene sequencing. On the other hand, due to process, instrument electromagnetic interference, and other factors, noise pollution will be produced inevitably in the process of genetic sequencing. To overcome the limitation of LRR, [25] proposed a method of Lat-LRR which expressed the original observation data as a linear combination of principal feature and latent feature for feature extraction. Considering the characteristics of heavy noise in gene expression profile, we added sparsity constraints to the model to construct the following Lat-LRR function:

The solution of Lat-LRR is given in [26]. Alternating direction method (ADM) is adopted to solve the model (6). Let ,  ; we constructed the following Augmented Lagrangian function:

Keep and  ; update :

Keep ,  ; update :

Keep ,  ,  ,  , and  ; update :

Keep ,  ,  , and  ; update :

Keep ,  ; update :

3. Method

3.1. NNDGLLRR

Lat-LRR overcomes the problem of too many constraints of LRR dictionary; however, Lat-LRR has limited ability to recover the subspace, and too many auxiliary variables are involved in the process of algorithm solving that involves a lot of matrix singularity value decomposition (SVD) and matrix inversion, which will affect the performance of the algorithm. Ref. [27] proposed a feature extraction method combining manifold constraint and nonnegative matrix factorization (NMF). In the case of NMF reducing dimensionality, the internal spatial structure of the data is maintained by manifold regularized constraint, and good experimental results are obtained. Ref. [28, 29] proposed an image clustering method combining manifold regularized constraint with Lat-LRR. Similar to the image data, the gene expression profile is also constituted by numerical matrix with high redundancy and heavy noise. Considering this characteristic, we constructed a new NNDGLLRR model on the basis of the original model.where , , and are nonnegative constants; the model is a nonnegative latent low-rank representation (NNLLRR) when α and are equal to zero. Model (13) takes a more general form. The dual regularized constraint is used to preserve the internal spatial structure of the original data, and sparse constraints and nonnegative constraints are used to maintain and enhance the robustness of the algorithm. and are Laplacian matrices, ,  . , and are weight matrix, and there are many ways to solve , and here we use Gaussian thermal weight. The specific solution is as follows:where is a constant; and represent the th column and th column of   (th and th sample); and represent the th row and the th row of  ,  .

ADM is used to solve model (12), and the following augmented Lagrange function is constructed:where is a Lagrangian multiplier; is a constant and .

Data in real life is generally nonnegative, and nonnegative constraints will make the calculation with a certain degree of sparseness and enhance the robustness of the algorithm. To maintain the nonnegative of feature extraction, we define the following operators:

The solution of model (15) is divided into three subproblems: first, the solution of variable , second, the solution of variables , and, third, the solution variable of  .

(1) Solving the First Subproblem. Update :

Regarding Taylor second-order expansion to (17), the approximate solution of is as follows:

Nonnegative constraints to are as follows:

Define ;  ;  ;  . Ref. [30] gives the solution of ; the solution process is as follows:

In (20), is the singular value decomposition (SVD) of , is the vector form of the singular value contraction operator (SVT), and is defined as follows:

(2) Solving the Second Subproblem. Similarly, update  :

Nonnegative constraints to   are as follows:

Define ;  .

(3) Solving the Third Subproblem. Update :where is a soft threshold operator (ST); is defined as follows:

The iterative process of each variable of NNDGLLRR is given above. The concrete updating process is shown in Algorithm 1.

Algorithm 1: Solving NNDGLLRR model with ALM.
3.2. Sparse Representation Classifier (SRC)

Sparse representation is a hotspot in the field of pattern recognition in recent years. SRC has been successfully applied in the field of image classification and has achieved relatively ideal experimental results [31]. Similar to the image data, the gene expression profile is also composed by a series of high redundancy and heavy noise of gene samples. In this paper, the latent features extracted by NNDGLLRR are regarded as data dictionary to construct the following SRC model:

According to the result of SRC, we can get the classification result of unknown gene sample :

The detailed flow of the SRC is shown in Algorithm 2.

Algorithm 2: The flow of SRC.
3.3. Algorithm Flow

To sum up, the algorithm can be divided into two parts; one is to use NNDGLLRR to extract latent features of the original gene expression profile, and the other is to use SRC to classify the latent features. The overall flow is as shown in Algorithm 3.

Algorithm 3: Algorithm flow.

4. Results and Discussion

4.1. Selecting the Test Data

To test the feature extraction performance of the algorithm, we used diffuse large B-cell lymphoma [32] (DLBCL), mixed lineage leukemia [33] (MLL), lung cancer [34] (LC), acute lymphoblastic leukemia [35] (ALL) gene sequences to make test, and the sample information of each group of genes as is shown in Table 1.

Table 1: Test data information.
4.2. Accuracy Test

-means and sparse representation classifier (SRC) are simple and common classifiers. To compare the clustering results of -means and SRC, the two kinds of classifiers are used to classify the original gene expression profile. Clustering results are shown in Table 2. It is not difficult to find that the classification effect of SRC is significantly higher than that of -means, which is due to the small number of gene expression profiles. To verify the effectiveness of the algorithm for feature extraction, the extracted features from LRR, Lat-LRR, and NNDGLLRR are classified by SRC. Classification results as shown in Table 2.

Table 2: Algorithm identification accuracy under different data sets.

Table 2 shows that any one of LRR, Lat-LRR, and NNDGLLRR can achieve feature extraction effectively. However, the feature extraction effect of NNDGLLRR is better than that of Lat-LRR. The category and number of samples, as well as dimension of the gene expression profile, will have an impact on the final recognition effect.

4.3. The Influence of Graph Regularized Coefficients

Generally, we set . To verify the influence of graph regularized coefficients on feature extraction, we have compared the recognition results of LRR, Lat-LRR, and NNDGLLRR under the condition of different values. The results are shown in Figure 1.

Figure 1: Clustering performance of algorithms with different graph regularized coefficients.

Through the test results of MLL and LC, we can find that manifold regularized constraint has obvious optimization effect on the gene expression profile feature extraction when the values of and are appropriate, and it can significantly improve the recognition effect of feature extraction. However, and should not be too large or too small. The optimal graph regularized coefficients may be different for different test data sets.

4.4. The Influence of Sparse Representation Coefficients

During the process of gene sequencing, the resulting gene expression profile will usually contain heavy noise due to the sequencing process. To verify the effect of the sparse constraint on the feature extraction, we tested the classification accuracy of LRR, Lat-LRR, and NNDGLLRR for feature extraction under different sparse constraint coefficients . The test results are shown in Figure 2.

Figure 2: Clustering performance of algorithms with different sparse representation coefficients.

Figure 2 shows that different sparse constraint coefficients have a considerable effect on the final feature extraction results. When the value of is appropriate, the performance of Lat-LRR and NNDGLLRR on feature extraction is better than that of LRR. In general, the performance of NNDGLLRR is better than that of Lat-LRR, which proves the validity of manifold constraint again.

4.5. Complexity Analysis

,  , and  , and we set the lowest ranks of and obtained by the algorithm as and . Then the complexity of SVT operation for and is about and , and the complexity of ST operation for is about . The complexity of construction the Laplacian matrix of and is about and ; and the complexity of one positive operation for and is about and . If the iteration of the algorithm is , then the overall complexity of LRR, Lat-LRR, and NNDGLLRR algorithms is shown in Table 3.

Table 3: Algorithm complexity calculation.

Generally, it is considered that for gene expression profile. It can be seen from Table 3 that LRR is the simplest in terms of computational complexity, but the performance of LRR on feature extraction is less effective than that of Lat-LRR and NNDGLLRR, and it is difficult to meet the actual demand. The result of Lat-LRR on feature extraction can be not bad, but the partitioning ability of the subspace is limited, and the operation speed is slow because of too many introduced variables. The variable update algorithm of NNDGLLRR not only reduces the calculated amount, but also achieves satisfactory results on feature extraction.

5. Conclusion

Aiming at the characteristics of high redundancy and heavy noise of gene expression profile, a feature extraction model of NNDGLLRR is proposed in this paper. In the process of experiment, we extracted the features of different gene expression profile by LRR, Lat-LRR, and NNDGLLRR and classified the extracted features by SRC. The experimental results show that the performance of NNDGLLRR on feature extraction is better than that of LRR and better than Lat-LRR slightly, which verified the comparative advantages of NNDGLRR. At the same time, compared with Lat-LRR, the overall complexity of NNDGLLRR is reduced through the improvement of the variable update algorithm. The experiments using different gene expression data sets for testing have made comparatively ideal experimental results, which proves the validity of the dual graph regularized constraint. In summary, the proposed nonnegative low-rank sparse constraint and dual graph regularized constraint are reasonable, and NNDGLLRR has good adaptability to different gene expression profile with high redundancy and heavy noise.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This paper is supported by the National Nature Science Foundation of China (nos. 51365017 and 61305019) and the Science and Technology Project of Jiangxi Province Education Department (no. GJJ150680).

References

  1. S. Bicciato, A. Luchini, and C. Di Bello, “PCA disjoint models for multiclass cancer analysis using gene expression data,” Bioinformatics, vol. 19, no. 5, pp. 571–578, 2003. View at Publisher · View at Google Scholar · View at Scopus
  2. D. Lee, W. Lee, Y. Lee, and Y. Pawitan, “Super-sparse principal component analyses for high-throughput genomic data,” BMC Bioinformatics, vol. 11, article 296, 2010. View at Publisher · View at Google Scholar · View at Scopus
  3. J.-X. Liu, Y.-T. Wang, C.-H. Zheng, W. Sha, J.-X. Mi, and Y. Xu, “Robust PCA based method for discovering differentially expressed genes,” BMC Bioinformatics, vol. 14, no. 8, article S3, 10 pages, 2013. View at Publisher · View at Google Scholar · View at Scopus
  4. A. Sharma and K. K. Paliwal, “Cancer classification by gradient LDA technique using microarray gene expression data,” Data and Knowledge Engineering, vol. 66, no. 2, pp. 338–347, 2008. View at Publisher · View at Google Scholar · View at Scopus
  5. J. Ye, J. Chen, R. Janardan, and S. Kumar, “Developmental stage annotation of Drosophila gene expression pattern images via an entire solution path for LDA,” ACM Transactions on Knowledge Discovery from Data, vol. 2, no. 1, article 4, 2008. View at Publisher · View at Google Scholar · View at Scopus
  6. K. K. Paliwal and A. Sharma, “Improved direct LDA and its application to DNA microarray gene expression data,” Pattern Recognition Letters, vol. 31, no. 16, pp. 2489–2492, 2010. View at Publisher · View at Google Scholar · View at Scopus
  7. D. Lutter, K. Stadlthanner, F. Theis et al., “Analyzing gene expression profiles with ICA,” in Proceedings of the 24th IASTED International Conference on Biomedical Engineering (BioMed '06), pp. 25–30, ACTA Press, Innsbruck, Austria, 2006.
  8. I. Hiroyuki, S. Hiroki, A. Kazuhiko et al., “Classification of gastric cancer subtypes by applying ICA to gene expression data and pathway analysis using Bayesian network,” IPSJ SIG Technical Reports, vol. 2012, no. 12, pp. 1–2, 2012. View at Google Scholar
  9. H. Chen, Y. Zhang, and I. Gutman, “A kernel-based clustering method for gene selection with gene expression data,” Journal of Biomedical Informatics, vol. 62, pp. 12–20, 2016. View at Publisher · View at Google Scholar · View at Scopus
  10. B. Liu, Q. Cui, T. Jiang, and S. Ma, “A combinational feature selection and ensemble neural network method for classification of gene expression data,” BMC Bioinformatics, vol. 5, no. 1, pp. 239–242, 2004. View at Publisher · View at Google Scholar · View at Scopus
  11. A.-H. Tan and H. Pan, “Predictive neural networks for gene expression data analysis,” Neural Networks, vol. 18, no. 3, pp. 297–306, 2005. View at Publisher · View at Google Scholar · View at Scopus
  12. J. Lee and C. Zhang, “Classification of gene-expression data: the manifold-based metric learning way,” Pattern Recognition, vol. 39, no. 12, pp. 2450–2463, 2006. View at Publisher · View at Google Scholar · View at Scopus
  13. H. Huang and H. Feng, “Gene classification using parameter-free semi-supervised manifold learning,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 9, no. 3, pp. 818–827, 2012. View at Publisher · View at Google Scholar · View at Scopus
  14. X. Hang, “Cancer classification by sparse representation using microarray gene expression data,” in Proceedings of the IEEE International Conference on Bioinformatics and Biomeidcine Workshops, pp. 174–177, IEEE, Philadelphia, Pa, USA, November 2008. View at Publisher · View at Google Scholar · View at Scopus
  15. C.-H. Zheng, L. Zhang, T.-Y. Ng, C. K. Shiu, and D.-S. Huang, “Metasample-based sparse representation for tumor classification,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 8, no. 5, pp. 1273–1282, 2011. View at Publisher · View at Google Scholar · View at Scopus
  16. R. Mehra, S. Varambally, L. Ding et al., “Identification of GATA3 as a breast cancer prognostic marker by global gene expression meta-analysis,” Cancer Research, vol. 65, no. 24, pp. 11259–11264, 2005. View at Publisher · View at Google Scholar · View at Scopus
  17. G. Ye, M. Tang, J. F. Cai et al., “Correction: low-rank regularization for learning gene expression programs,” PLoS ONE, vol. 9, no. 1, Article ID e82146, 2014. View at Google Scholar
  18. A. Kapur, K. Marwah, and G. Alterovitz, “Gene expression prediction using low-rank matrix completion,” BMC Bioinformatics, vol. 17, no. 1, article 243, 2016. View at Publisher · View at Google Scholar · View at Scopus
  19. C.-H. Zheng, T.-Y. Ng, L. Zhang, C.-K. Shiu, and H.-Q. Wang, “Tumor classification based on non-negative matrix factorization using gene expression data,” IEEE Transactions on Nanobioscience, vol. 10, no. 2, pp. 86–93, 2011. View at Publisher · View at Google Scholar · View at Scopus
  20. Y.-S. Lee, A. Krishnan, Q. Zhu, and O. G. Troyanskaya, “Ontology-aware classification of tissue and cell-type signals in gene expression profiles across platforms and technologies,” Bioinformatics, vol. 29, no. 23, pp. 3036–3044, 2013. View at Publisher · View at Google Scholar · View at Scopus
  21. L. Nanni, S. Brahnam, S. Ghidoni, E. Menegatti, and T. Barrier, “A comparison of methods for extracting information from the co-occurrence matrix for subcellular classification,” Expert Systems with Applications, vol. 40, no. 18, pp. 7457–7467, 2013. View at Publisher · View at Google Scholar · View at Scopus
  22. M. Vimaladevi and B. Kalaavathi, “A microarray gene expression data classification using hybrid back propagation neural network,” Genetika, vol. 46, no. 3, pp. 1013–1026, 2014. View at Google Scholar
  23. G. Lee, A. Singanamalli, H. Wang et al., “Supervised Multi-view Canonical Correlation Analysis (sMVCCA): integrating histologic and proteomic features for predicting recurrent prostate cancer,” IEEE Transactions on Medical Imaging, vol. 34, no. 1, pp. 284–297, 2015. View at Publisher · View at Google Scholar · View at Scopus
  24. R. Vidal and P. Favaro, “Low rank subspace clustering (LRSC),” Pattern Recognition Letters, vol. 43, no. 1, pp. 47–61, 2014. View at Publisher · View at Google Scholar · View at Scopus
  25. P. Li, J. Bu, J. Yu, and C. Chen, “Towards robust subspace recovery via sparsity-constrained latent low-rank representation,” Journal of Visual Communication and Image Representation, vol. 37, pp. 46–52, 2016. View at Publisher · View at Google Scholar · View at Scopus
  26. G. Liu and S. Yan, “Latent low-rank representation for subspace segmentation and feature extraction,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV '11), pp. 1615–1622, November 2011. View at Publisher · View at Google Scholar · View at Scopus
  27. D. Cai, X. He, J. Han, and T. S. Huang, “Graph regularized nonnegative matrix factorization for data representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 8, pp. 1548–1560, 2011. View at Publisher · View at Google Scholar · View at Scopus
  28. Z. Zhang, S. Yan, and M. Zhao, “Similarity preserving low-rank representation for enhanced data representation and effective subspace learning,” Neural Networks, vol. 53, pp. 81–94, 2014. View at Publisher · View at Google Scholar · View at Scopus
  29. M. Yin, J. Gao, Z. Lin, Q. Shi, and Y. Guo, “Dual graph regularized latent low-rank representation for subspace clustering,” IEEE Transactions on Image Processing, vol. 24, no. 12, pp. 4918–4933, 2015. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  30. Z. Zhang, G. Ely, S. Aeron et al., “Novel methods for multilinear data completion and de-noising based on tensor-SVD,” Computer Science, vol. 44, no. 9, pp. 3842–3849, 2014. View at Google Scholar
  31. J. Wang, C. Lu, M. Wang, P. Li, S. Yan, and X. Hu, “Robust face recognition via adaptive sparse representation,” IEEE Transactions on Cybernetics, vol. 44, no. 12, pp. 2368–2378, 2014. View at Publisher · View at Google Scholar · View at Scopus
  32. M. A. Shipp, K. N. Ross, P. Tamayo et al., “Diffuse large B-cell lymphoma outcome prediction by gene-expression profiling and supervised machine learning,” Nature Medicine, vol. 8, no. 1, pp. 68–74, 2002. View at Publisher · View at Google Scholar · View at Scopus
  33. S. A. Armstrong, J. E. Staunton, L. B. Silverman et al., “MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia,” Nature Genetics, vol. 30, no. 1, pp. 41–47, 2002. View at Publisher · View at Google Scholar · View at Scopus
  34. A. Bhattacharjee, W. G. Richards, J. Staunton et al., “Classification of human lung carcinomas by mRNA expression profiling reveals distinct adenocarcinoma subclasses,” Proceedings of the National Academy of Sciences of the United States of America, vol. 98, no. 24, pp. 13790–13795, 2001. View at Publisher · View at Google Scholar · View at Scopus
  35. E.-J. Yeoh, M. E. Ross, S. A. Shurtleff et al., “Classification, subtype discovery, and prediction of outcome in pediatric acute lymphoblastic leukemia by gene expression profiling,” Cancer Cell, vol. 1, no. 2, pp. 133–143, 2002. View at Publisher · View at Google Scholar · View at Scopus