Review Article  Open Access
Ching Siang Tan, Wai Soon Ting, Mohd Saberi Mohamad, Weng Howe Chan, Safaai Deris, Zuraini Ali Shah, "A Review of Feature Extraction Software for Microarray Gene Expression Data", BioMed Research International, vol. 2014, Article ID 213656, 15 pages, 2014. https://doi.org/10.1155/2014/213656
A Review of Feature Extraction Software for Microarray Gene Expression Data
Abstract
When gene expression data are too large to be processed, they are transformed into a reduced representation set of genes. Transforming largescale gene expression data into a set of genes is called feature extraction. If the genes extracted are carefully chosen, this gene set can extract the relevant information from the largescale gene expression data, allowing further analysis by using this reduced representation instead of the full size data. In this paper, we review numerous software applications that can be used for feature extraction. The software reviewed is mainly for Principal Component Analysis (PCA), Independent Component Analysis (ICA), Partial Least Squares (PLS), and Local Linear Embedding (LLE). A summary and sources of the software are provided in the last section for each feature extraction method.
1. Introduction
The advances of microarray technology allow the expression levels of thousands of genes to be measured simultaneously [1]. This technology has caused an explosion in the amount of microarray gene expression data. However, the gene expression data generated are highdimensional, containing a huge number of genes and small number of samples. This is called the “large small problem” [2]. The highdimensional data are the main problem when analysing the data. As a result, instead of using gene selection methods, feature extraction methods are also important in order to reduce the dimensionality of highdimensional data. Instead of eliminating irrelevant genes, feature extraction methods work by transforming the original data into a new representation. Feature extraction is usually better than gene selection in terms of causing less information loss. As a result, the highdimensionality problem can be solved using feature extraction.
Software is a set of machine readable instructions that direct a computer’s processor to perform specific operations. With increases in the volume of data generated by modern biomedical studies, software is required to facilitate and ease the understanding of biological processes. Bioinformatics has emerged as a discipline in which emphasis is placed on easily understanding biological processes. Gheorghe and Mitrana [3] relate bioinformatics to computational biology and natural computing. Higgs and Attwood [4] believe that bioinformatics is important in the context of evolutionary biology.
In this paper, the software applications that can be used for feature extraction are reviewed. The software reviewed is mainly for Principal Component Analysis (PCA), Independent Component Analysis (ICA), Partial Least Squares (PLS), and Local Linear Embedding (LLE). In the last section for each feature extraction method, a summary and sources are provided.
2. Software for Principal Component Analysis (PCA)
In the domain of dimension reduction, PCA is one of the renowned techniques. The fundamental concept of PCA is to decrease the dimensionality of a given data set, whilst maintaining as plentiful as possible the variation existing in the initial predictor variables. This is attained by transforming the initial variables to a latest set of predictor variables. Linear amalgamation of the initial variables is . In mathematical domain, PCA successively optimizes the variance of a linear amalgamation of the initial predictor variables: conditional upon the constraint , for every . The orthogonal constraint makes sure that the linear combinations are uncorrelated; that is, , . These linear combinations are denoted as the principle components (PCs): The projection vectors (or known as the weighting vectors) can be attained by eigenvalue decomposition on the covariance matrix : where is the th eigenvalue in the decreasing order, for , and is the resultant eigenvector. The eigenvalue calculates the variance of the th PC and the eigenvector gives the weights for the linear transformation (projection).
2.1. FactoMineR
FactoMineR is an package that provides various functions for the analysis of multivariate data [5]. The newest version of this package is maintained by Hussen et al. [6]. There are a few main features provided by this package; for example, different types of variables, data structures, and supplementary information can be taken into account. Besides that, it offers dimension reduction methods such as Principal Component Analysis (PCA), Multiple Correspondence Analysis (MCA), and Correspondence Analysis (CA). The steps in implementing PCA are described in Lê et al. [5] and Hoffmann [7]. For PCA, there are three main functions for performing the PCA, plotting it, and printing its results. This package is mainly for Windows, MacOS, and Linux.
2.2. ExPosition
ExPosition is an R package for the multivariate analysis of quantitative and qualitative data. ExPosition stands for Exploratory Analysis with the Singular Value Decomposition. The newest version of this package is maintained by Beaton et al. [8]. A variety of multivariate methods are provided in this package such as PCA, multidimensional scaling (MDS), and Generalized PCA. All of these methods can be performed by using the corePCA function in this package. Another function, epPCA, can be applied to implement PCA. Besides that, Generalized PCA can be implemented using the function epGPCA as well. All of these methods are used to analyse quantitative data. A plotting function is also offered by this package in order to plot the results of the analysis. This package can be installed on Windows, Linux, and MacOS.
2.3. amap
The R package “amap” was developed for clustering as well as PCA for both parallelized functions and robust methods. It is an R package for multidimensional analysis. The newest version is maintained by Lucas [9]. Three different types of PCA are provided by this package. The methods are PCA, Generalized PCA, and Robust PCA. PCA methods can be implemented using the functions acp and pca for PCA, acpgen for Generalized PCA, and acprob for Robust PCA. This package also allows the implementation of correspondence factorial analysis through the function afc. Besides that, a plotting function is also provided for plotting the results of PCA as a graphical representation. The clustering methods offered by this package are kmeans and hierarchical clustering. The dissimilarity matrix and distance matrix can be computed using this package as well. This package is mainly for Windows, Linux, and MacOS.
2.4. ADE4
ADE4 was originally developed by Thioulouse et al. [10] as software for analyzing multivariate data and displaying graphics. This software includes a variety of methods such as PCA, CA, Principal Component Regression, PLS, Canonical Correspondence Analysis, Discriminant Analysis, and others. Besides that, this software is implemented in an R environment as an R package, “ade4.” The newest version of this package is maintained by Penel [37]. In this package, PCA can be performed by using the dudi.pca function. A visualization function is also provided in order to visualize the results as a graphical representation. In previous studies, this package was implemented by Dray and Dufour [38] to identify and understand ecological community structures. This package is mainly for Linux, Windows, and MacOS.
2.5. MADE4
MADE4 (microarray ade4) was developed by Culhane et al. [11] for multivariate analysis of gene expression data based on the R package “ade4.” Basically, it is the extensions of the R package “ade4” for microarray data. The purpose of writing this software was to help users in the analysis of microarray data using multivariate analysis methods. This software is able to handle a variety of gene expression data formats, and new visualization software has been added to the package in order to facilitate the visualization of microarray data. Other extra features such as data preprocessing and gene filtering are included as well. However, this package was further improved by the addition of the LLSimpute algorithm to handle the missing values in the microarray data by Moorthy et al. [39]. It is implemented in an R environment. The advance of this package is that multiple datasets can be integrated to carry out analysis of microarray data. The newest version is maintained by Culhane [40]. This package can be installed on Linux, Windows, and MacOS.
2.6. XLMiner
XLMiner is addin software for Microsoft Excel that offers numerous data mining methods for analysing data [12]. It offers a quick start in the use of a variety of data mining methods for analysing data. This software can be used for data reduction using PCA, classification using Neural Networks or Decision Trees [41, 42], class prediction, data exploration, affinity analysis, and clustering. In this software, PCA can be implemented using the Principle Component tab [43]. This software is implemented in Excel. As a result, the dataset should be in an Excel spreadsheet. In order to start the implementation of XLMiner, the dataset needs to be manually partitioned into training, validation, and test sets. Please see http://www.solver.com/xlminerdatamining for further details. This software can be installed on Windows and MacOS.
2.7. ViSta
ViSta stands for Visual Statistics System and can be used for multivariate data analysis and visualization in order to provide a better understanding of the data [13]. This software is based on the LispStat system [44]. It is an open source system that can be freely distributed for multivariate analysis and visualization. PCA and multiple and simple CA are provided in this software. Its main advance is that the data analysis is guided in a visualization environment in order to generate more reliable and accurate results. The four stateoftheart visualization methods offered by this software are GuideMaps [45], WorkMaps [46], Dynamic Statistical Visualization [47], and Statistical ReVision [48]. The plugins for PCA can be downloaded from http://www.mdp.edu.ar/psicologia/vista/vista.htm. An example of implementation of the analysis using PCA can be viewed in ValeroMora and Ledesma [49]. This software can be installed on Windows, Unix, and Macintosh.
2.8. imDEV
Interactive Modules for Data Exploration and Visualization (imDEV) [14] is an application of RExcel that integrates and Excel for the analysis, visualization, and exploration of multivariate data. It is used in Microsoft Excel as addins by using an R package. Basically, it is implemented in Visual Basic and . In this software, numerous dimension reduction methods are provided such as PCA, ICA, PLS regression, and Discriminant Analysis. Besides that, this software also offers clustering, imputing of missing values, feature selection, and data visualization. The visualization methods are offered such as dendrograms, distribution plots, biplots, and correlation networks. This software is compatible with a few versions of Microsoft Excel such as Excel 2007 and 2010.
2.9. Statistics Toolbox
Statistical Toolbox offers a variety of algorithms and tools for data modelling and data analysis. Multivariate data analysis methods are offered by this toolbox. The methods include PCA, clustering, dimension reduction, factor analysis, visualization, and others. In the statistical toolbox of MATLAB, several PCA functions are provided for multivariate analysis, for example, pcacov, princomp, and pcares (MathWorks). Most of these functions are used for dimensional reduction. pcacov is used for covariance matrices, princomp for raw data matrices, and pcares for residuals from PCA. All of these functions are implemented in MATLAB.
2.10. Weka
Weka [16] is data mining software that provides a variety of machine learning algorithms. This software offers feature selection, data preprocessing, regression, classification, and clustering methods [50]. This software is implemented in a Java environment. PCA is used as a dimension reduction method in Weka to reduce the dimensionality of complex data through transformation. However, not all of the datasets are complete. Prabhume and Sathe [51] introduced a new filter PCA for Weka in order to solve the problem of incomplete datasets. It works by estimating the complete dataset from the incomplete dataset. This software is mainly for Windows, Linux, and MacOS.
2.11. NAG Library
In NAG Library, the function of PCA is provided as the g03aa routine [17] in both C and Fortran. This routine performs PCA on data matrices. This software was developed by the Numerical Algorithms Group. In the NAG Library, more than 1700 algorithms are offered for mathematical and statistical analysis. For PCA, it is suitable for multivariate methods, G03. Other methods provided are correlation analysis, wavelet transforms, and partial differential equations. Please refer to http://www.nag.com/numeric/MB/manual_22_1/pdf/G03/g03aa.pdf for further details about the g03aaa routine. This software can be installed on Windows, Linux, MacOS, AIX, HP UX, and Solaris.
2.12. Case Study
In this section, we will discuss the implementation of coinertia analysis (CIA) to crossplatform visualization in MADE4 and ADE4 to perform multivariate analysis of microarray datasets. To demonstrate, PCA was applied on 4 childhood tumors (NB, BLNHL, EWS, and RMS) from a microarray gene expression profiling study [52]. From these data, a subset (khan$train, 206 genes × 64 cases), each case’s factor denoting the respective class (khan$train classes, length = 64), and a gene annotation’s data frame are accessible in aforementioned dataset in MADE4: < library (made4) < data (khan) < dataset = khan$train < fac = khan$train.classes < geneSym = khan$annotation$Symbol < results.coa < ord (dataset, type = “coa”) < par (mfrow = c (1, 2)) < plotarrays (results.coa, classvec = fac) < plotgenes (results.coa, genelabels = geneSym).Figure 1 shows the PCA of a 306gene subset. As origin as the point of reference, the more advanced gene and case are projected in the similar direction, the stronger the association between involved gene and case is (gene is upregulated in that array sample).
2.13. Summary of PCA Software
Tables 1 and 2 show the summary and sources of PCA software, respectively. Table 3 discusses the related work of this software.


3. Software for Independent Component Analysis (ICA)
ICA is considered as a valuable extension of PCA that has been established considering the blind separation of independent sources from their linear combination [53]. In a way, the initial point of ICA is the property of uncorrelation of general PCA. Based on data matrix , whose rows tally to observational variables and whose columns are the individuals of the corresponding variables, the ICA model of can be written as With generality intact, is a mixing matrix, whereas is a source matrix under the necessity of being statistically independent as possible. “Independent components” are the new variables confined in the rows of , to wit, the variables observed are linearly collected independent components. Mutual information , where is the marginal entropy of the variables , is the probabilistic density function, and is the joint entropy [54]. Value the independent components able to be attained by discovering the correct linear mixtures of the observational variables, since mixing can be inverted as
3.1. FastICA
FastICA is the most widely used method of ICA [55]. It is implemented in an R environment as the R package “FastICA” for performing ICA and Projection Pursuit by using the FastICA algorithm. FastICA was first introduced by Hyvärinen [54] for single and multiple component extraction. The FastICA algorithm is based on a fixedpoint iteration scheme maximizing nonGaussianity as a measure of statistical independence. This package is maintained by Marchini et al. [18]. ICA is used to extract the informative features through a transformation of the observed multidimensional random vectors into independent components. This package is mainly for Windows, Linux, and MacOS. FastICA is also implemented in MATLAB. In MATLAB, FastICA implements a fast fixedpoint algorithm for ICA as well as projection pursuit. It provides a simple user interface and also a powerful algorithm for computation.
3.2. JADE
JADE is an R package that provides a function for implementing ICA. This package is maintained by Nordhausen et al. [19]. In this package, Cardoso’s JADE algorithm [56] is provided for ICA. Instead of the JADE algorithm, other Blind Source Separation (BSS) methods such as the SOBI [57] and AMUSE [58] methods are offered. Both of these methods are mainly used for solving second order BSS problems. Amari error [59] is offered to evaluate the performance of the ICA algorithm. This package can be installed on Linux, Windows, and MacOS.
3.3. High Performance Signal Analysis Tools (HiPerSAT)
HiPerSAT is written in C++ for processing electroencephalography (EEG) data with whitening of data and ICA [20]. MPI and OpenMP are used to perform parallel analysis of ICA. Basically, this software is used to analyse EEG data in order to understand the neurological components of brain activity. In this software, FastICA, SOBI, and Informax algorithms are offered. HiPerSAT is integrated into MATLAB and EEGLAB [60]. EEGLAB is MATLABbased software that is used for analysing EEG data. However, the advantage of HiPerSAT is that it can handle larger datasets compared to MATLAB. In comparison to EEGLAB, HiPerSAT is able to handle large datasets without partitioning but EEGLAB requires data partitioning. Data whitening is performed before implementing the algorithms. This software can be installed on all platforms.
3.4. MineICA
MineICA is an R package that supplies the implementation of ICA on transcriptomic data [21]. The main purpose of MineICA is to provide an easier way of interpreting the decomposition results from ICA. Besides that, this software also provides a correlationbased graph for comparing the components from different datasets. The newest version of this package is maintained by Biton [61]. This package provides some features such as storage of ICA results, annotation of features, and visualization of the results of ICA. This package can be installed on Linux, MacOS, and Windows.
3.5. Pearson Independent Component Analysis
Karnanen [22] developed an R package for a feature extraction technique based on the Pearson ICA algorithm. This is a mutual informationbased blind source separation approach which applies the Pearson system as a parametric model. In order to extract the independent components using the ICA algorithm, the mutual information of the components has to be minimized. However minimization of mutual information is required to use a score function. The Pearson system was used to model the score function. The parameters of the Pearson system are estimated by the method of moments. In order to speed up the algorithm, tanh nonlinearity is used when the distribution is far from Gaussian.
3.6. Maximum Likelihood Independent Component Analysis
Teschenforff [23] developed an R package for ICA by using maximum likelihood estimation. This method was first introduced by Hyvaerinen et al. [62]. This method uses a fixedpoint algorithm as the Maximum Likelihood estimation. For a fixed set of data and underlying statistical model, Maximum Likelihood selects the set of values of the model parameters that maximizes the likelihood function. Maximum Likelihood estimation gives a unified approach to estimation, which is welldefined in the case of normal distribution. By using a maximum likelihood framework and controlling the number of algorithm runs, this fixedpoint algorithm provides a very fast implementation for maximization of likelihood.
3.7. Sample Case Study
In this section, we utilize MineICA for microarraybased gene expression data of 200 breast cancer tumors kept in the package breastCancerMAINZ [63] based on a study done by Biton et al. [21]. In this study, we focused on how MineICA can be utilized to study an ICAbased decomposition. Pseudo code for this case study is as follows:(1)Loading the library and the data(2)Creation of an IcaSet object(2.1)Load an example of expression data(2.2)Run ICA(2.3)Create a MineICAParams object, function buildMineICAParams(2.4)Create an IcaSet instance, function buildIcaSet(2.5)IcaSet basics(3)Run global analysis(4)Run analysis by calling individual functions(4.1)Write description of contributing genes or features, function writeProjByComp(4.2)Plot heatmaps of the contributing elements, function plot_heatmapsOnSel(4.3)Gene enrichment analysis, function runEnrich(4.4)Association with sample variables(4.5)Clustering of the samples according to each component(4.6)Comparison of IcaSet objects, function runCompareIcaSets.Figure 2 explains the correlation based graph denoting relationship between independent components (IC) attained on four breast cancer samples’ microarray data. Every node represents an IC and respective colors denote the origin of dataset. Thickness of edge represents the extent of correlation among the linked ICs. Black edges represent reciprocal nodes.
3.8. Summary of ICA Software
Tables 4 and 5 show the summary and sources of ICA software, respectively.


4. Software for Partial Least Squares (PLS)
The fundamental hypothesis of PLS is that the experimental information is created by a framework or methodology which is determined by a small number of latent characteristics. Thusly, PLS goes for discovering uncorrelated linear transformation of the initial indicator characteristics which have high covariance with the reaction characteristics. In light of these latent components, PLS predicts reaction characteristics , the assignment of regression, and reproduce initial matrix , the undertaking of data modelling, in the meantime. The purpose of building components in PS is to optimize the covariance among the variable and the initial predictor variables : Restricted to constraint , for all . The crucial assignment of PLS is to attain the vectors of maximum weights to build a small number of components, while PCA is an “unsupervised” method that utilizes the data only. To develop the components, , PLS decomposes and to yield a bilinear denotation of the data [64]: where ’s are vectors of weights for building the PLS components , ’s are scalars, and and are the residuals. The concept of PLS is to assume and by regression.
4.1. Partial Least Squares Discriminant Analysis
Barker and Rayens [24] developed a PLS for discriminant analysis. However the original PLS was not designed for discriminant purposes. PLS Discriminant Analysis is used to find a linear regression model by projecting the dependent features and the independent features to a new space. Then the fundamental relations can be extracted from the latent variables. This method was developed for software called Unscrambler, which was first developed by Martens and Naes [65]. Unscrambler is a commercial software product for multivariate data analysis. Unscrambler is used for analysing large and complex datasets quickly and easily using the power of multivariate analysis. Moreover this multivariate data analysis also offers exceptional data visualization.
4.2. Least Squares: Partial Least Squares
Jørgensen et al. [25] proposed a method of using an iterative combination of PLS and ordinary least squares to extract the relationship between the predictor variable and the responses. This method is based on a combination of least squares estimates for the design variables and PLS regression on the spectra. The PLS scores were incorporated into the ordinary least squares equation on the spectra. The idea is to separate the information from the spectral and design matrices in a nice way. However this method is able to extract the information even when fewer components are used. In addition, this method is insensitive to the relative scaling of the spectra and the process. Moreover this combination method is also less biased than the individual PLS technique.
4.3. Powered Partial Least Squares Discriminant Analysis
Liland and Indahl [26] extended the Powered PLS to Powered PLS Discriminant Analysis to overcome the extraction of information for the multivariate classification problem. This method can construct more efficient group separation and generate more interpretive outcomes than the ordinary Partial Least Square Discriminant Analysis technique. The features extracted by the Powered PLS can contribute to revealing the relevance of particular predictors and often requires smaller and simpler components than ordinary PLS. Moreover the optimization task is equivalent to maximizing the correlation between the transformed predictors and the groups. This makes it possible to discard the influence of less important predictors. This method was also developed by the authors for availability in an R package.
4.4. Penalized Partial Least Squares
Krämer et al. [27] proposed a combination of the feature extraction technique PLS with a penalization framework. This method is an extension of PLS regression using a penalization technique. Ordinary PLS is suited for regression problems by minimizing a quadratic loss function iteratively. In addition, the representation in terms of kernel matrices provides an intuitive geometric interpretation of the penalty term. The penalty terms control the roughness of the estimated functions. With the incorporation of penalization into this framework, the research direction became more promising. This method is used to extract relevant information for highdimensional regression problems and also for noisy data. This method was also developed by the Krämer and her colleagues colleagues [66] for availability in an R package.
4.5. SlimPLS
Gutkin et al. [33] proposed a feature extraction method based on PLS called SlimPLS. Rankingbased filters usually utilize a univariate method when selecting features. The filter methods can produce reasonable feature sets especially when the original feature sets are uncorrelated. However the chosen feature set will be suboptimal when the features of the original set are dependent. Some of the features will add little discriminative power on top of previously selected features. SlimPLS is a multivariate feature extraction method which incorporates feature dependencies into calculation. This multivariate property is constructed by combining the highly predictive feature with some less predictive but correlated features. This is because the added features will provide more information on the behaviour of the samples.
4.6. Sparse Partial Least Squares Discriminant Analysis and Sparse Generalized Partial Least Squares
Chung and Keles [28] proposed two extension feature extraction approaches based on Sparse PLS. These approaches are Sparse PLS Discriminant Analysis and Sparse Generalized PLS for highdimensional datasets. These two approaches improved ordinary PLS by employing feature extraction and dimension reduction simultaneously. These two approaches perform well even with unbalanced sample sizes of the classes. Sparse PLS Discrimination Analysis is computationally efficient because it only requires computational time for one run of Sparse PLS and a classifier. Moreover, Sparse Generalized PLS extends Sparse PLS to the generalized linear model framework. These methods were also developed by Chung and Keles for availability in an R package.
4.7. Degrees of Freedom of Partial Least Squares
Kramer and Sugiyama [29] proposed a method of unbiased estimation of the degrees of freedom for PLS regression. The authors stated that the construction of latent components from the independent variable also depended on the dependent variable. However for PLS regression, the optimal number of components needs to be determined first. One of the ways of determining the optimal number of components is through the degrees of freedom for the complexity of fitted models. Moreover the degrees of freedom estimate can be used for the comparison of different regression methods. Furthermore, the two implementations for the degrees of freedom utilize the connection between PLS regression and numerical linear methods from numerical linear. The authors also developed an R package for this unbiased estimation of the degrees of freedom of PLS.
4.8. Surrogate Variable Analysis Partial Least Squares
Chakraborty and Datta [30] proposed a surrogate variable analysis method based on PLS. In differential gene expression analysis, one of the important issues is to avoid the hidden confounders in the dataset. The hidden confounders of gene expression are caused by different environmental conditions of the samples. However this problem cannot be simply overcome by modifying the gene expression data by using a normalizing technique. This method can extract the informative features by identifying the hidden effects of the underlying latent factors using ordinary PLS and applying analysis of covariance (ANCOVA). ANCOVA is applied with the PLS signatures of these hidden effects as covariates in order to identify the genes that are truly differentially expressed. This method was also developed by the authors for availability in an R package.
4.9. Partial Least Squares Path Modelling
Sanchez and Trinchera [31] developed an R package for Partial Least Squares Path Modelling (PLSPM). PLSPM was first introduced by Wold [67] and is also known as Structural Equation Modelling (SEM). It can be used as a compositebased alternative to factorbased SEM. PLSPM can be used when the distributions are highly skewed. Moreover, PLSPM can also be used to estimate relationships between latent variables with several indicators even though the sample size is small. Basically, PLSPM consists of two sets of linear equations: the inner model and the outer model. The inner model specifies the relations between latent variables, while the outer model specifies the relations between a latent variable and its observed indicator. PLSPM is a multivariate feature extraction analysis technique based on the causeeffect relationships of the unobserved and observed features.
4.10. Partial Least Squares Regression for Generalized Linear Models
Bertrand et al. [32] developed a software application of PLS regression for generalized linear models. Generalized linear models are important to allow the response features to have a distribution other than normal. Generalized linear models can be viewed as a case of generalized linear models with an identity link. From the perspective of generalized linear models, however, it is useful to suppose that the distribution function is the normal distribution with constant variance and the link function is the identity, which is the canonical link if the variance is known. However, the generalized linear models preserve all the predictive power of the features where the predicted means are not assumed to be normally distributed. PLS regression is used to extract the predictive features from the generalized linear models.
4.11. Case Study
In this section, we will discuss the R package consists of svpls. This function will call fitModel function in order to appropriate a number of ANCOVA models that are identified by pmax to the data and opt for the best model by looking the minimum value of the Akaike’s information Criterion (AIC) [68]. Subsequently, this model is utilized to forecast the real pattern of genes’ differential expression. The command lines in R are as follows: > ## Fitting the optimal ANCOVA model to the data gives: > fit <svpls (10, 10, hidden_fac.dat, pmax = 5, fdr = 0.05) > ## The optimal ANCOVA model, its AIC value and the positive genes detected > ## from it are givenL > fit$opt.model > fit$AIC.opt > fit$genes > ## The corrected gene expression matrix obtained after removing the effects of the hidden variability is given by: > Y.corrected < fit$Y.corr > pval.adj <fit$pvalues.adj.For instance, we study the efficacy of svapls on ALL/AML preprocessed dataset [69]. This data consists of expression levels of 7129 genes that have been logtransformed over two samples of patients. These two sets of 47 patients and 25 patients reported to suffer from Acute lymphoblastic Leukemia (ALL) and Acute Myeloid Leukemia (AML), respectively. By using svpls function, we yielded initial 1000 genes with corrected expression matrix. Random samples’ distribution from four sources in the abovementioned matrix removes the extra effects owing to reported batch specific clustering in the initial data. In this context svapls performed equally efficient relative to another popular R package ber for removing batch effects in microarray data as shown in Figure 3.
(a)
(b)
(c)
4.12. Summary of PLS Software
Tables 6 and 7 show the summary and sources of PLS software, respectively. Table 8 shows the related works on discussed software.



5. Software for Local Linear Embedding (LLE)
Straightforward geometric intuitions are the basis for LLE algorithm. Assume that given data comprise of realvalued vectors , for each dimensionality, tested by some core manifold. Given that there is adequate data, every data point and their neighbors are expected to be situated on or near to a locally linear patch of the manifold. Abovementioned patches are described by linear coefficients that rebuild every data point from respective neighbors. Equation (8) is the cost function used to calculate reconstruction errors which sums the squared distances between all the data points and their reconstructions. The weights summarize the contribution of the th data point to the th reconstruction. The optimal weights are found by solving a leastsquares problem [70]:
5.1. lle
An package “lle” has been developed in order to implement LLE for feature extraction. This package provides the algorithm of LLE in order to transform highdimensional data into lowdimensional data. The newest version of this package is maintained by Diedrich and Abel [34]. The main functions of this package allow users to perform LLE and also to plot the results of LLE. The implementation of LLE is based on the idea of Ridder and Duin [71]. Besides that, some enhancements such as selection of the subset and calculation of the intrinsic dimension are offered. This package can be installed on Windows, Linux, and MacOS.
5.2. RDRToolbox
RDRToolbox is an R package developed for nonlinear dimension reduction with LLE and Isomap. The package is maintained by Bartenhagen [35]. It offers the transformation of highdimensional to lowdimensional data by using either LLE or Isomap. Besides that, a plotting function is provided to plot the results. In addition, the DavisBouldin Index is provided for the purposes of validating clusters. It is mainly for Linux, MacOS, and Windows.
5.3. ScikitLearn
Scikitlearn is software implemented in Python by integrating machine learning algorithms [36]. It is a simpletouse software that allows users to implement a variety of machine learning algorithms. The machine learning algorithms include classification, clustering, feature extraction, model selection, manifold learning, and other methods. Isomap, LLE, and Local Tangent Space Alignment (LTSA) are provided by this software. Please see http://scikitlearn.org/stable/ for further details. This software can be installed on a variety of platforms such as Windows and Ubuntu.
5.4. Case Study
This section demonstrates the dimension reduction workflow for the publicly available the Golub et al. leukemia dataset (see Figure 5). The data is available as R package and can be downloaded and loaded via > source (“http://bioconductor.org/biocLite.R”) > biocLite (“golubEsets”) > library (golubEsets) > data (Golub_Merge).The dataset consists of 72 samples, divided into 47 ALL and 25 AML patients, and 7129 expression values. In this example, we compute a twodimensional LLE and Isomap embedding and plot the results. At first, we extract the features and class labels: > golubExprs = t (exprs (Golub_Merge)) > labels = pData (Golub_Merge)$ALL.AML > dim (golubExprs).The residual variance of Isomap can be used to estimate the intrinsic dimension of the dataset: > Isomap (data = golubExprs, dims = 1 : 10, plotResiduals = TRUE, ).Based on Figure 4, regarding the dimensions for which the residual variances stop to decrease significantly, we can expect a low intrinsic dimension of two or three and, therefore, visualization true to the structure of the original data. Next, we compute the LLE and Isomap embedding for two target dimensions: > golubIsomap = Isomap (data = golubExprs, dims = 2, ) > golubLLE = LLE(data = golubExprs, dim = 2, ).The DavisBouldinIndex shows that the ALL and AML patients are well separated into two clusters: > DBIndex(data = golubIsomap$dim2, labels = labels) > DBIndex(data = golubLLE, labels = labels).Finally, we use plotDR to plot the twodimensional data: > plotDR(data = golubIsomap$dim2, labels = labels, axesLabels = c(“”, “”), legend = TRUE) > title (main = “Isomap”) > plotDR (data = golubLLE, labels = labels, axesLabels = c(“”,“”), legend = TRUE) > title (main = “LLE”).Both visualizations, using either Isomap or LLE, show distinct clusters of ALL and AML patients, although the cluster overlaps less in the Isomap embedding. This is consistent with the DBIndex, which is very low for both methods, but slightly higher for LLE. A threedimensional visualization can be generated in the same manner and is best analyzed interactively within R.
(a)
(b)
5.5. Summary of LLE Software
Tables 9 and 10 show the summary and sources of LLE software, respectively. Table 11 shows the related works in discussed software.



6. Conclusion
Nowadays, numerous software applications have been developed to help users implement feature extraction of gene expression data. In this paper, we present a comprehensive review of software for feature extraction methods. The methods are PCA, ICA, PLS, and LLE. These software applications have some limitations in terms of statistical aspects as well as computational performance. In conclusion, there is a need for the development of better software.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
The authors would like to thank Universiti Teknologi Malaysia for funding this research by Research University Grants (Grant nos. J130000.2628.08J80 and J130000.2507.05H50). The authors would also like to thank Research Management Centre (RMC), Universiti Teknologi Malaysia, for supporting this research.
References
 S. Van Sanden, D. Lin, and T. Burzykowski, “Performance of gene selection and classification methods in a microarray setting: a simulation study,” Communications in Statistics. Simulation and Computation, vol. 37, no. 12, pp. 409–424, 2008. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 Q. Liu, A. H. Sung, Z. Chen et al., “Gene selection and classification for cancer microarray data based on machine learning and similarity measures,” BMC Genomics, vol. 12, supplement 5, article S1, 2011. View at: Publisher Site  Google Scholar
 M. Gheorghe and V. Mitrana, “A formal languagebased approach in biology,” Comparative and Functional Genomics, vol. 5, no. 1, pp. 91–94, 2004. View at: Publisher Site  Google Scholar
 P. G. Higgs and T. Attwood, “Bioinformatics and molecular evolution,” Comparative and Functional Genomics, vol. 6, pp. 317–319, 2005. View at: Google Scholar
 S. Lê, J. Josse, and F. Husson, “FactoMineR: an R package for multivariate analysis,” Journal of Statistical Software, vol. 25, no. 1, pp. 1–18, 2008. View at: Google Scholar
 F. Hussen, J. Josse, S. Le, and J. Mazet, “Package ‘FactoMineR’,” 2013, http://cran.rproject.org/web/packages/FactoMineR/FactoMineR.pdf. View at: Google Scholar
 I. Hoffmann, “Principal Component Analysis with FactoMineR,” 2010, http://www.statistik.tuwien.ac.at/public/filz/students/seminar/ws1011/hoffmann_ausarbeitung.pdf. View at: Google Scholar
 D. Beaton, C. R. C. Fatt, and H. Abdi, Package 'ExPosition', 2013, http://cran.rproject.org/web/packages/ExPosition/ExPosition.pdf.
 A. Lucas, “Package ‘amap’,” 2013, http://cran.rproject.org/web/packages/amap/vignettes/amap.pdf. View at: Google Scholar
 J. Thioulouse, D. Chessel, S. Dolédec, and J.M. Olivier, “ADE4: a multivariate analysis and graphical display software,” Journal of Statistics and Computing, vol. 7, no. 1, pp. 75–83, 1997. View at: Publisher Site  Google Scholar
 A. C. Culhane, J. Thioulouse, G. Perriere, and D. G. Higgins, “MADE4: an R package for multivariate analysis of gene expression data,” Bioinformatics, vol. 21, no. 11, pp. 2789–2790, 2005. View at: Google Scholar
 I. H. Witten and E. Frank, Data Mining: Practical Machine Learning Tools and Techniques, Elsevier, 2nd edition, 2005.
 F. W. Young and C. M. Bann, “ViSta: a visual statistics system,” in Statistical Computing Environments for Social Research, R. A. Stine and J. Fox, Eds., pp. 207–235, Sage, 1992. View at: Google Scholar
 D. Grapov and J. W. Newman, “imDEV: a graphical user interface to R multivariate analysis tools in Microsoft Excel,” Bioinformatics, vol. 28, no. 17, Article ID bts439, pp. 2288–2290, 2012. View at: Publisher Site  Google Scholar
 The MathWorks, Statistics Toolbox for Use with MATLAB, User Guide Version 4, 2003, http://www.pi.ingv.it/~longo/CorsoMatlab/OriginalManuals/stats.pdf.
 M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten, “The WEKA data mining software: an update,” ACM SIGKDD Explorations Newsletter, vol. 11, no. 1, pp. 10–18, 2009. View at: Publisher Site  Google Scholar
 NAG Toolbox for Matlab: g03aa, G03Multivariate Methods, http://www.nag.com/numeric/MB/manual_22_1/pdf/G03/g03aa.pdf.
 J. L. Marchini, C. Heaton, and B. D. Ripley, “Package ‘fastICA’,” http://cran.rproject.org/web/packages/fastICA/fastICA.pdf. View at: Google Scholar
 K. Nordhausen, JF. Cardoso, J. Miettinen, H. Oja, E. Ollila, and S. Taskinen, “Package ‘JADE’,” http://cran.rproject.org/web/packages/JADE/JADE.pdf. View at: Google Scholar
 D. Keith, C. Hoge, R. Frank, and A. D. Malony, HiPerSAT Technical Report, 2005, http://nic.uoregon.edu/docs/reports/HiPerSATTechReport.pdf.
 A. Biton, A. Zinovyev, E. Barillot, and F. Radvanyi, “MineICA: independent component analysis of transcriptomic data,” 2013, http://www.bioconductor.org/packages/2.13/bioc/vignettes/MineICA/inst/doc/MineICA.pdf. View at: Google Scholar
 J. Karnanen, “Independent component analysis using score functions from the Pearson system,” 2006, http://cran.rproject.org/web/packages/PearsonICA/PearsonICA.pdf. View at: Google Scholar
 A. Teschenforff, Independent Component Analysis Using Maximum Likelihood, 2012, http://cran.rproject.org/web/packages/mlica2/mlica2.pdf.
 M. Barker and W. Rayens, “Partial least squares for discrimination,” Journal of Chemometrics, vol. 17, no. 3, pp. 166–173, 2003. View at: Publisher Site  Google Scholar
 K. Jørgensen, V. Segtnan, K. Thyholt, and T. Næs, “A comparison of methods for analysing regression models with both spectral and designed variables,” Journal of Chemometrics, vol. 18, no. 10, pp. 451–464, 2004. View at: Publisher Site  Google Scholar
 K. H. Liland and U. G. Indahl, “Powered partial least squares discriminant analysis,” Journal of Chemometrics, vol. 23, no. 1, pp. 7–18, 2009. View at: Publisher Site  Google Scholar
 N. Krämer, A. Boulesteix, and G. Tutz, “Penalized Partial Least Squares with applications to Bspline transformations and functional data,” Chemometrics and Intelligent Laboratory Systems, vol. 94, no. 1, pp. 60–69, 2008. View at: Publisher Site  Google Scholar
 D. Chung and S. Keles, “Sparse partial least squares classification for high dimensional data,” Statistical Applications in Genetics and Molecular Biology, vol. 9, no. 1, 2010. View at: Publisher Site  Google Scholar  MathSciNet
 N. Kramer and M. Sugiyama, “The degrees of freedom of partial least squares regression,” Journal of the American Statistical Association, vol. 106, no. 494, pp. 697–705, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 S. Chakraborty and S. Datta, “Surrogate variable analysis using partial least squares (SVAPLS) in gene expression studies,” Bioinformatics, vol. 28, no. 6, pp. 799–806, 2012. View at: Publisher Site  Google Scholar
 G. Sanchez and L. Trinchera, Tools for Partial Least Squares Path Modeling, 2013, http://cran.rproject.org/web/packages/plspm/plspm.pdf.
 F. Bertrand, N. Meyer, and M. M. Bertrand, “Partial Least Squares Regression for generalized linear models,” http://cran.rproject.org/web/packages/plsRglm/plsRglm.pdf. View at: Google Scholar
 M. Gutkin, R. Shamir, and G. Dror, “SlimPLS: a method for feature selection in gene expressionbased disease classification,” PLoS ONE, vol. 4, no. 7, Article ID e6416, 2009. View at: Publisher Site  Google Scholar
 H. Diedrich and M. Abel, “Package ‘lle’,” http://cran.rproject.org/web/packages/lle/lle.pdf. View at: Google Scholar
 C. Bartenhagen, “RDRToolbox: a package for nonlinear dimension reduction with Isomap and LLE,” 2013, http://bioconductor.org/packages/2.13/bioc/vignettes/RDRToolbox/inst/doc/vignette.pdf. View at: Google Scholar
 F. Pedregosa, G. Varoquaux, A. Gramfort et al., “Scikitlearn: machine learning in Python,” Journal of Machine Learning Research, vol. 12, pp. 2825–2830, 2011. View at: Google Scholar
 S. Penel, Package “ade4”, 2013, http://cran.rproject.org/web/packages/ade4/ade4.pdf.
 S. Dray and A. Dufour, “The ade4 package: implementing the duality diagram for ecologists,” Journal of Statistical Software, vol. 22, no. 4, pp. 1–20, 2007. View at: Google Scholar
 K. Moorthy, M. S. Mohamad, S. Deris, and Z. Ibrahim, “Multivariate analysis of gene expression data and missing value imputation based on llsimpute algorithm,” International Journal of Innovative Computing, Information and Control, vol. 6, no. 5, pp. 1335–1339, 2012. View at: Google Scholar
 A. Culhane, Package ‘made4’, 2013, http://bioconductor.org/packages/release/bioc/manuals/made4/man/made4.pdf.
 C. V. Subbulakshmi, S. N. Deepa, and N. Malathi, “Comparative analysis of XLMiner and WEKA for pattern classification,” in Proceedings of the IEEE International Conference on Advanced Communication Control and Computing Technologies (ICACCCT '12), pp. 453–457, Ramanathapuram Tamil Nadu, India, August 2012. View at: Publisher Site  Google Scholar
 S. Jothi and S. Anita, “Data mining classification techniques applied for cancer disease—a case study using Xlminer,” International Journal of Engineering Research & Technology, vol. 1, no. 8, 2012. View at: Google Scholar
 T. Anh and S. Magi, Principal Component Analysis: Final Paper in Financial Pricing, National Cheng Kung University, 2009.
 L. Tierney, LispStat: An ObjectOriented Environment for Statistical Computing & Dynamic Graphics, AddisonWesley, Reading, Mass, USA, 1990.
 F. W. Young and D. J. Lubinsky, “Guiding data analysis with visual statistical strategies,” Journal of Computational and Graphical Statistics, vol. 4, pp. 229–250, 1995. View at: Google Scholar
 F. W. Young and J. B. Smith, “Towards a structured data analysis environment: a cognitionbased design,” in Computing and Graphics in Statistics, A. Buja and P. A. Tukey, Eds., vol. 36, pp. 253–279, Springer, New York, NY, USA, 1991. View at: Google Scholar
 F. W. Young, R. A. Faldowski, and M. M. McFarlane, “Multivariate statistical visualization,” in Handbook of Statistics, C. R. Rao, Ed., pp. 958–998, 1993. View at: Google Scholar
 M. McFarlane and F. W. Young, “Graphical sensitivity analysis for multidimensional scaling,” Journal of Computational and Graphical Statistics, vol. 3, no. 1, pp. 23–34, 1994. View at: Google Scholar
 P. M. ValeroMora and R. D. Ledesma, “Using interactive graphics to teach multivariate data analysis to psychology students,” Journal of Statistics Education, vol. 19, no. 1, 2011. View at: Google Scholar
 E. Frank, M. Hall, G. Holmes et al., “Weka—a machine learning workbench for data mining,” in Data Mining and Knowledge Discovery Handbook, O. Maimon and L. Rokach, Eds., pp. 1269–1277, 2010. View at: Google Scholar
 S. S. Prabhume and S. R. Sathe, “Reconstruction of a complete dataset from an incomplete dataset by PCA (principal component analysis) technique: some results,” International Journal of Computer Science and Network Security, vol. 10, no. 12, pp. 195–199, 2010. View at: Google Scholar
 J. Khan, J. S. Wei, M. Ringnér et al., “Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks,” Nature Medicine, vol. 7, no. 6, pp. 673–679, 2001. View at: Publisher Site  Google Scholar
 P. Comon, “Independent component analysis, a new concept?” Signal Processing, vol. 36, no. 3, pp. 287–314, 1994. View at: Publisher Site  Google Scholar
 A. Hyvärinen, “Fast and robust fixedpoint algorithms for independent component analysis,” IEEE Transactions on Neural Networks, vol. 10, no. 3, pp. 626–634, 1999. View at: Publisher Site  Google Scholar
 V. Zarzoso and P. Comon, “Comparative speed analysis of FastICA,” in Independent Component Analysis and Signal Separation, M. E. Davies, C. J. James, S. A. Abdallah, and M. D. Plumbley, Eds., vol. 4666 of Lecture Notes in Computer Science, pp. 293–300, Springer, Berlin, Germany, 2007. View at: Google Scholar
 J. F. Cardoso and A. Souloumiac, “Blind beamforming for nonGaussian signals,” IEE Proceedings, Part F: Radar and Signal Processing, vol. 140, no. 6, pp. 362–370, 1993. View at: Google Scholar
 A. Belouchrani, K. AbedMeraim, J. Cardoso, and E. Moulines, “A blind source separation technique using secondorder statistics,” IEEE Transactions on Signal Processing, vol. 45, no. 2, pp. 434–444, 1997. View at: Publisher Site  Google Scholar
 L. Tong, V. C. Soon, Y. F. Huang, and R. Liu, “AMUSE: a new blind identification algorithm,” in Proceedings of the IEEE International Symposium on Circuits and Systems, pp. 1784–1787, May 1990. View at: Google Scholar
 S. Amari, A. Cichocki, and H. H. Yang, “A new learning algorithm for blind signal separation,” in Proceedings of the Advances in Neural Information Processing Systems Conference, pp. 757–763, 1996. View at: Google Scholar
 A. Delorme and S. Makeig, “EEGLAB: an open source toolbox for analysis of singletrial EEG dynamics including independent component analysis,” Journal of Neuroscience Methods, vol. 134, no. 1, pp. 9–21, 2004. View at: Publisher Site  Google Scholar
 A. Biton, “Package ‘MineICA’,” 2013, http://www.bioconductor.org/packages/2.13/bioc/manuals/MineICA/man/MineICA.pdf. View at: Google Scholar
 A. Hyvaerinen, J. Karhunen, and E. Oja, Independent Component Analysis, John Wiley & Sons, New York, NY, USA, 2001.
 M. Schmidt, D. Böhm, C. von Törne et al., “The humoral immune system has a key prognostic impact in nodenegative breast cancer,” Cancer Research, vol. 68, no. 13, pp. 5405–5413, 2008. View at: Publisher Site  Google Scholar
 I. S. Helland, “On the structure of partial least squares regression,” Communications in Statistics. Simulation and Computation, vol. 17, no. 2, pp. 581–607, 1988. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 H. Martens and T. Naes, Multivariate calibration, John Wiley & Sons, London, UK, 1989. View at: MathSciNet
 N. Krämer and A. Boulesteix, “Package “ppls”,” 2013, http://cran.rproject. org/web/packages/ppls/ppls.pdf. View at: Google Scholar
 H. Wold, “Soft modeling: the basic design and some extensions,” in Systems under Indirect Observations: Causality, Structure, Prediction, K. G. Joreskog and H. Wold, Eds., Part 2, pp. 1–54, NorthHolland, Amsterdam, The Netherlands, 1982. View at: Google Scholar
 H. Akaike, “Likelihood and the Bayes procedure,” Trabajos de Estadística y de Investigación Operativa, vol. 31, no. 1, pp. 143–166, 1980. View at: Google Scholar
 T. R. Golub, D. K. Slonim, P. Tamayo et al., “Molecular classification of cancer: class discovery and class prediction by gene expression monitoring,” Science, vol. 286, no. 5439, pp. 531–537, 1999. View at: Publisher Site  Google Scholar
 S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction linear embedding,” Science, vol. 290, no. 5500, pp. 2323–2326, 2000. View at: Publisher Site  Google Scholar
 D. Ridder and R. P. W. Duin, Locally Linear Embedding, University of Technology, Delft, The Netherlands, 2002.
Copyright
Copyright © 2014 Ching Siang Tan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.