Abstract

Cancer can be considered as one of the leading causes of death widely. One of the most effective tools to be able to handle cancer diagnosis, prognosis, and treatment is by using expression profiling technique which is based on microarray gene. For each data point (sample), gene data expression usually receives tens of thousands of genes. As a result, this data is large-scale, high-dimensional, and highly redundant. The classification of gene expression profiles is considered to be a (NP)-Hard problem. Feature (gene) selection is one of the most effective methods to handle this problem. A hybrid cancer classification approach is presented in this paper, and several machine learning techniques were used in the hybrid model: Pearson’s correlation coefficient as a correlation-based feature selector and reducer, a Decision Tree classifier that is easy to interpret and does not require a parameter, and Grid Search CV (cross-validation) to optimize the maximum depth hyperparameter. Seven standard microarray cancer datasets are used to evaluate our model. To identify which features are the most informative and relative using the proposed model, various performance measurements are employed, including classification accuracy, specificity, sensitivity, F1-score, and AUC. The suggested strategy greatly decreases the number of genes required for classification, selects the most informative features, and increases classification accuracy, according to the results.

1. Introduction

Cancer can be considered as one of the leading death causes [1], and gene expression profiles derived from microarray data have been identified as promising cancer diagnostic indices [2].

Microarrays are used to measure thousands of genes interactions at the same time and create a cellular function global picture [3, 4].

The classification of microarray data is one of the most common and important applications of functional genomics microarray data which means classifying patients samples to many classes based on their gene expression profiles [5, 48]. In literature, there are many machine learning methods which have been used in the application of microarray data classification [7, 8]. However, the classification process of microarray data is still challenging and difficult due to the small samples numbers and its high dimensionality [9]. Microarray gene expression experiments often generate a lot of features for a small patients number which leads to a high-dimension dataset with a small samples number. Gene expression data is very challenging and complex; genes are correlated with one another directly or indirectly which make classification process a very hard and difficult mission which generally requires using an accurate and powerful feature selection technique.

The main phase in categorization systems is feature selection. To improve classification performance, feature selection-based classification approaches have been investigated. Any statistical technique’s [10] success is contingent on the predetermination of independent properties. As a result, identification is kept to a minimum, but the purpose of feature selection is to find the most informative subset of characteristics and/or reduce the number of dimensions.

In order to be able to select an informative genes subset while eliminating/declining redundant or irrelevant genes and to be able to improve the performance of microarray high-dimension data classification, this research study introduces a hybrid feature selection approach, called PCC-DTCV, which combines different methods, Pearson’s correlation coefficients (PCC) and Decision Tree (DT) [11] as classification approach and feature selection [1217] and Grid Search CV which can be employed as an optimization technique [13, 14, 1820, 4247], to optimize the tuning parameters of DT (max-depth) to be able to get the optimal feature subset.

In order to evaluate the suggested PCC-DTCV model, 7 popular datasets from the most well-known used microarray datasets for different cancer types are used. The proposed model evaluation was carried out using classification accuracy, k-fold cross-validation, sensitivity, and specificity values.

The proposed method, according to the experimental methods, reduces dimensionality and selects the most important and informative features (genes) and improves the identification of cancer tissues from benign tissues. Furthermore, PCC-DTCV improves the accuracy classification performance.

Recently, there have been several major research efforts to study the classification and diagnosis of cancer microarray data. As a framework for the study presented in the remainder of the article, we offer an outline of some of this analysis and the applied techniques in Table 1.

3. Preliminaries

3.1. DNA Microarray

One of the main tools employed in molecular biology and genetics to track gene expression is the DNA microarray, which refers to the degree of development of genome-determined protein molecules. Although the protein structure varies from the mRNA measurement gene instead of protein structure, it is a popular method for calculating gene expression, so thousands of proteins would be difficult to analyse, and mRNA sequences [39] are hybridized with their additional DNA or RNA sequences, although proteins do not have these properties. Figure 1 shows experimental microarray steps that include the extraction of mRNA from a cell or tissue. The structure of the DNA microarray matrix is shown in Figure 2, described by N (gene measurement) × M (a sample or condition involved in a specific microarray experiment).

The data on gene expression given by microarray experiments are processed as the broad matrix, where the samples are defined by the columns and measured the expression strength of the i-th gene in the j-th sample samples or experimental conditions). The row of genes represents G (). In this study, we concentrate on solving two problems: the analysis and understanding of microarray data which have faced several challenges:(1)High dimensionality and noise(2)Increasing the accuracy performance taking into consideration interpretations that are biologically significant in gene expression when analysing microarray data

3.2. Decision Tree (DT)

Decision Tree learning can be considered as one of the most practical and commonly used methods for inductive inference.

It is a technique used for approximating discrete-valued functions that can learn disjunctive expressions and is resilient to noisy data [21]. These inductive inference algorithms are among the most common, and they have been successfully employed for a wide range of tasks, from learning up to diagnosis of the medical cases to learn to assess the loan applicants credit risk. Decision Tree learning is a nonbacktracking, one-step lookahead (hill-climbing) quest through the space of all possible Decision Trees using a heuristic. Decision Trees are commonly used in bioinformatics, especially in decision analysis to help identify a type of disease that is more likely to produce a specific symptom. A tree-based classifier was used in our proposed model PCC-DTCV for its basic properties, explicit context, and fast transformation to if-then law. The Decision Tree is a device for decision support which uses a dendritic graph and its potential problems, including unintended incident effects, resource costs, and utility. In bioinformatics, Decision Trees are widely used, especially in decision analysis to help classify a type of disease more likely to achieve a certain symptom. As a descriptive means for manipulating conditional probabilities, the Decision Tree may be used. This works to isolate and conquer [22] methods for creating a Decision Tree. As shown in equations (1)–(5), Entropy controls how a Decision Tree determines to divide the data. It influences how the borders are drawn by a Decision Tree. The Gini (Gini impurity) index measures the degree of probability of a feature that is incorrectly labelled to be randomly selected. The mistakes are in classification. The gain ratio is the parameter of the Decision Tree for measuring output and it is defined aswhere n is the number of classes and 0 log_20 = 0 in entropy calculations:and the function SplitInfo is defined as follows:where gives the probability distribution of the data sample and “bit” due to calculating statistics. Equation (5) is used with log function with basis 2. The relation between Entropy, misclassification error, and Gini impurity is seen in Figure 3.

3.3. Correlation Coefficients

The Correlation-based Feature Selection algorithm computes the correlation coefficient using a correlation-based heuristic evaluation function (CFS). It gets over the limitation of univariate filter approaches which do not take into account feature interaction [23, 24]. The process of correlation [25] can be used to measure the relationship between variables (genes). The linear relationship between 2 variables can be described using Pearson’s correlation coefficients or correlation coefficients in statistics. The well-known similarity measure between two features is defined as correlation. If two features depend on each other linearly then their coefficient of correlation is . The correlation coefficient is 0 if the features are uncorrelated. For a pair of variables (X, Y), the coefficient of linear correlation r is given by the following equation [26]:

In our model, we used Pearson’s correlation coefficient (PCC) as feature selection [38]. This is a mathematical strength measure of a paired data linear relationship. It is defined by r and is limited as follows:

, in which positive values mean that the linear correlation is positive, negative values imply that the linear correlation is negative, zero value means there is a nonlinear correlation, and the closer the value is to or 1, the stronger the linear correlation exists.

The proposed model used Pearson’s correlation to calculate the correlation between the features. Also using symmetric uncertainty measures, the class feature is calculated. If the value is higher than the threshold value (0.5), then the feature will be chosen. The selection ends when the number of features equals n log n. According to the decreasing order, each selected feature is ranked. To remove the redundant feature, the feature-feature correlation is done. Threshold values (0.4, 0.5, and 0.6) were used and their results were compared with the Decision Tree before and after the optimization of its maximum depth (max-depth).

3.4. Grid Search Cross-Validation

A Grid Search is a parameter tuning method which builds and evaluates a model methodically for each algorithm parameters combination specified in a grid. The well-known terms that should be addressed while using Grid Search CV are discussed as follows [27].

(A)Estimator. This term is used in order to implement the estimator interface in scikit-learn. This parameter receives the classifier to be trained [28].(B)Parameter Grid. It is a Python key-value dictionary with pairs of parameter names and parameter settings. All combinations of these parameters are tested to ensure the highest level of accuracy.(C)Cross-Validation. This determines the strategy for cross-validation splitting. Cross-validation is a technique to resample available data to evaluate machine learning models. The main goal of this is to evaluate the performance of machine learning models on previously unseen data. The main advantage of using this is that it produces less biased or optimistic results compared to a simple train-test split. The way it works is that it first shuffles the dataset randomly. The entire dataset is then divided into groups. Each group is used as test data, while the others are used as training data. The evaluation band related to each group is saved to be summarized at the end to be able to check the performance of the model. Importantly, each sample appears once in the testing data and is used to train times. Figure 4 explains the Grid Search cross-validation process in detail.During the training step, hyperparameters that are optimizer variables are executed to get optimal average values after several trial-and-error processes. The tree maximum depth (max-depth) is the most critical hyperparameter that influences the difficulty of the Decision Tree model [6], whereas the maximum depth is the length of the longest distance from the tree root to a node. The root node is 0.0 in depth. The higher overall depth value induces overfitting, and the lower value results in underfitting. In the proposed hybrid model PCC-DTCV, Grid Search cross-validation is used to overcome the overfitting constraint with the regular Grid Search. To obtain better parameters, Grid Search cross-validation is used to optimize the tree maximum depth hyperparameter to get optimal hyperparameter value.

4. Proposed Model PCC-DTCV

The proposed hybrid model PCC-DTCV for classifying and diagnosing cancer microarray data (as shown in Figure 5) will be clarified in detail. Key model phases are as follows: (1) The datasets (cancer microarray) are preprocessed. This stage is essential to (a) avoid features in wide digital ranges dominating fewer ranges, (b) avoid numeric complexities through calculation, and (c) adjust each feature to the range [0,1]. (2) PCC is applied for finding the association between the continuous features and the class feature, evaluation to genes is based on how it is related to targeting the features that have a high correlation with the target class having a correlation value near to 1, and the genes with the correlation coefficient 0.5 are removed and the output subsets are informative and important genes (features). (3) The sample data output subset is the partition used to train data to match the model and test data used to validate the model as soon as a model is fully trained (trained DT). (4) Grid Search CV with 10-fold cross-validation is used to obtain the optimal parameters (max-depth) of the tree on the training set. (5) Model accuracy is evaluated, using 10-fold cross-validation, and our dataset is 10-fold split, using train split to train the classifier, and prediction quality on unseen data is estimated (test split). The confusion matrix is measured for test split. The AUC, specificity, and sensitivity in addition to accuracy are computed. (6) If the termination conditions are not met, the process from three to six is repeated until the termination criteria are satisfied. If the termination is fulfilled (max folds, number = 10), the result will be a subset of optimal genes (features) with the most important and informative genes.

5. Results and Discussion

In the following section, the findings of numerous studies that have been performed to determine the efficacy of the proposed hybrid model PCC-DTCV are discussed. A PC with the following properties is the platform used to test the proposed model: RAM 4 GB and 32-bit OS (Windows 7), Intel(R) Core (TM) i5-7500CPU, as well as frameworks such as Pandas, NumPy, Keras, SciPy, and Matplotlib and the Python programming language. The datasets for human cancer research included the following: breast cancer, prostate, colon tumor, lung cancer, and leukemia were applied to 7 different microarray datasets with a limited sample and high-dimensional and binary class. Also, all datasets have two types, and Table 2 lists the definitions of all datasets.

In all studies, 10-fold cross-validation techniques were applied. The data were randomly divided into 10 different subsets (with the same size in 10-fold cross-validation), and the experiment was conducted 10 times. One subset was used as a test set for each run, and one was used as a validation set, and the other sets were used as a training set (Figure 5). To obtain a single estimate, the mean of the k results from the folds can then be determined. To approximate the model, 10-fold cross-validation was used in our experiments, and the results obtained are illustrated in the form of mean standard deviation. Also, in both analyses, the number of iterations was 50. The value of the tree (max-depth) hyperparameter is [3, 30]. Our evaluation included the four following measurements:(1)Accuracy (ACC): it is the most used evaluation standard for the proportion of correctly predicted pairs, but using it alone is usually insufficient. Accuracy .(2)Sensitivity (Sen.): a diseased person is likely to be recognized as diseased through the test. , where TN is the true negatives number and TP is the true positives number.(3)Specificity (Spec.): the likelihood is that a person without the illness is defined by the test as nondiseased (or healthy). It is described as , where FP implies the number of false positives and TN is the number of true negatives.(4)The AUC shows the area under the receiver operating characteristics (ROC) curve, and it is calculated as .

TP is true positive, FP is false positive, TN is true negative, and FN is false negative. Based on the confusion matrix, we evaluated the performance of the proposed method and rival gene selection. To assess the output of our model, a statistical test method called Chi-Square test [21] is used to check how well the observed values for a given distribution fit the distribution when the variables are independent.

6. Experimental Results

The experimental results are discussed in this section to validate the performance of the our model PCC-DTCV. The classification accuracy performance and the selected feature (genes) numbers of the PCC-DTCV method with PCC (0.4, 0.5, and 0.6) for all microarray datasets, respectively, are summarized in Tables 38 and Figures 69. The statistical model is described in Table 9. Six different performance metrics were chosen for result estimation: AUC, ACC, specificity, sensitivity, recall, and F1-score [10, 22]. To reduce the high dimensionality, improve the classification performance of the problem at hand, and select the most informative genes, we run PCC-DTCV using all datasets and get the informative selected features (genes) number. It can be noted that PCC-DTCV gives the genes (features) ordered list according to the highest correlation informative, importance, and relevant genes. It is obvious that PCC-DTCV achieves the highest level of dimensional reduction by choosing smallest number of informative genes and the biggest dimensional dataset is Chowdary for breast cancer with 22283 features (genes) and 104 samples, as seen in Tables 3 and 4. The highly correlated subset selected genes with the target class by the PCC-DTCV with PPC = 0.4 are 410 features (genes) from 22283. PCC-DTCV compared the performance of DT classifier without any optimization method and optimized DT as given in Tables 3 and 4, respectively, as well as the accuracy of the highest-dimensional dataset with PPC = 0.4 with optimized DT classifier and DT classifier. With accuracy of (0.92  0.09), the optimized DT classifier shows better accuracy than DT classifier with accuracy of (0.90  0.09).

Tables 5 and 6 show that the number of selected (features) genes with PPC 0.5 is lower than the number of selected (features) genes with PPC 0.4 for all datasets; it means that the PCC-DTCV achieved higher-dimensional reduction with PPC 0.5 than with PPC 0.4. In terms of ACC, AUC, sensitivity, and specificity, the PCC-DTCV achieved higher results compared to those achieved with PPC 0.4 with both optimized DT classifier and DT classifier; for example, the number of selected (features) genes of Gordon dataset for lung cancer is 274 compared with 743 for PPC 0.4, in addition to the values of 94%, 88%, 97%, and 79% for terms ACC, AUC, sensitivity, and specificity with DT classifier and 94%, 90%, 97%, and 87% with optimized DT classifier. From Tables 7 and 8, we can see that the PCC-DTCV with PPC 0.6 has the best ACC, AUC, sensitivity, specificity, recall, and F1-score compared to PPC 0.4 and 0.5, which means that the classification performance of PCC-DTCV is best with PPC 0.6 and the selected genes (features) are the most relevant, important, and informative genes.

Figures 69 represent the classification performance obtained using PPC 0.4, 0.5, and 0.6 and DT classifier with the PCC-DTCV model for prostate cancer, colon cancer, leukemia, lung cancer, and breast cancer datasets. From Figure 6, the achieved results show that the suggested PCC-DTCV model gives better accuracy for breast cancer, lung cancer, and leukemia datasets and can obtain the highest accuracy with PPC 0.6 in the Chowdary dataset for breast cancer, 96%. From Figure 7, the AUC metric obtains more than 90% for both Chowdary and Singh datasets and more than 80% for all datasets except Alon dataset for colon cancer, obtaining 71% for PPC 0.6. From Figure 8, the sensitivity metric obtains more than 90 with Gordon, Chowdary, and Golub datasets for lung cancer, breast cancer, and Leukemia, respectively, with PPC 0.4, 0.5, and 0.6. Specificity metric shown in Figure 9 achieves 96% for Chowdary dataset with PPC 0.6.

Figures 1013 represent the classification performance obtained using optimized DT classifier and PPC with PCC-DTCV model for prostate cancer, colon cancer, leukemia, lung cancer, and breast cancer datasets. In Figure 10, accuracy metric can obtain more than 90% for Gordon, Chowdary, and Golub datasets for lung cancer, breast cancer, and leukemia, respectively, with PPC 0.6. From Figure 11, the AUC metric obtains more than 80% for all datasets except Alon dataset for colon cancer, obtaining 76% for PPC 0.6. From Figure 12, the sensitivity metric obtains more than 90% with Gordon, Chowdary, and Golub datasets for lung cancer, breast cancer, and leukemia, respectively, with PPC 0.4, 0.5, and 0.6. Specificity metric shown in Figure 13 obtains more than 80% for all datasets with PPC 0.6. Finally, the proposed model gives the best classification performance with PPC 0.6 for both optimized DT classifier and DT classifier and the selected features (genes) have minimum redundancy and are maximally relevant and informative with high correlation with target class.

These results clarify that the proposed models follow a normal distribution as the value in both Kolmogorov–Smirnov and Shapiro–Wilk Sig. 0.02 .05 shown in Table 9. The first condition is fulfilled by following a normal distribution. The second condition is that there are no outliers and anomalies, as well as applying regression analysis to use the Manhattan distance method, which calculates the outlier values. In addition, the emergence of a new variable and the largest value in Manhattan distance is compared and it is less valuable than the number of existing variables using the Chi-Square. The value of Manhattan distance = 18.466 compared to using the Chi-Square. The critical value of the Chi-Square distribution will be applied at df = 6 and the value of 0.001 is found to be at 0.999 confidence interval and uncertainty interval 0.001. If the value of 22.46 is greater than 18.466, then it is assumed that there are no multiple-choice outliers. Sig = .0155, which is greater than 0.0001. It is a fulfilled condition and it means that the homogeneity condition variance exists and is fulfilled by a choice of Finney greater than 0.5. This indicates that the variance of the dependent variables is equal.

7. Conclusions

This paper presents a PCC-DTCV model for cancer diagnosis and classification. PCC-DTCV optimizes the tuning parameter (max-depth) of the DT classifier using Grid Search and Pearson’s correlation used as a gene (feature) selection method. The proposed model was effective in identifying an optimal or near-optimal subset of informative and important genes and yielded high classification results. Several experiments were carried out to evaluate the proposed PCC-DTCV model for selecting the most informative genes to improve the performance classification of cancer microarray data. In addition to reducing the dimensionality of microarray data, the result demonstrates the model’s effective performance in selecting the most informative genes with high importance. The results showed that PCC-DTCV provides the best fit for cancer type prediction in terms of specificity, AUC, sensitivity, and accuracy. In future work, for Decision Tree optimization, we anticipate using ski driver and other algorithms.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This research was supported by the Researchers Supporting Project number (RSP-2021/244), King Saud University, Riyadh, Saudi Arabia.