Abstract

Aiming at the shortcomings of high feature reduction using traditional rough sets, such as insensitivity with noise data and easy loss of potentially useful information, combining with genetic algorithm, in this paper, a VPRS-GA (Variable Precision Rough Set--Genetic Algorithm) model for high-dimensional feature reduction of medical image is proposed. Firstly, rigid inclusion of the lower approximation is extended to partial inclusion by classification error rate β in the traditional rough set model, and the ability dealing with noise data is improved. Secondly, some factors of feature reduction are considered, such as attribute dependency, attributes reduction length, and gene coding weight. A general framework of fitness function is put forward, and different fitness functions are constructed by using different factors such as weight and classification error rate β. Finally, 98 dimensional features of PET/CT lung tumor ROI are extracted to build decision information table of lung tumor patients. Three kinds of experiments in high-dimensional feature reduction are carried out, using support vector machine to verify the influence of recognition accuracy in different fitness function parameters and classification error rate. Experimental results show that classification accuracy is affected deeply by different weight values under the invariable classification error rate condition and by increasing classification error rate under the invariable weigh value condition. Hence, in order to achieve better recognition accuracy, different problems use suitable parameter combination.

1. Introduction

Rough set theory was developed by Pawlak in 1982 [1], and it is a mathematical tool to deal with vagueness and uncertainty. The classification ability unchanged in its main idea, decision or classification rules of problem are derived by knowledge reduction [2]. The Variable Precision Rough Set (VPRS) theory, proposed by Ziarko, and is an extension of original rough set model. For inconsistent information system, the VPRS model allows a flexible approximation boundary region by a precision variable β [3]. When β = 0, Pawlak rough set model is a special case of variable precision rough set model. The main task of variable precision rough set model is to solve the problem of data classification with no function or uncertainty. The hierarchical model of attribute reduction for variable precision rough set is studied by Xiaowei [4]. There is abnormal phenomenon in existing attribute reduction models; therefore, a variable precision rough set attribute reduction algorithm with the property of interval is proposed, and the reduction abnormal problem is transformed into a hierarchical model representation, and the reduction anomaly is gradually eliminated by the layer-by-layer reduction model; Jie and Jiayang [5] puts forward that there may be a reduction jump phenomenon in variable precision rough set feature reduction, which affects the quality of reduction and brings the problem of attribute reduction of variable precision rough set; Pei and Qinghua [6] proposes an FCM clustering algorithm based on variable precision rough set; according to the threshold characteristics of the variable precision rough set model, the algorithm divides the objects in the edge of the cluster into the positive, negative, and boundary regions, to improve the accuracy of clustering. Two different solutions of variable precision rough set attribute reduction algorithm are proposed by Hao and Junan [7]; based on tolerance matrix and minimal reduction of attribute core, the attribute kernel idea of variable precision rough set is proposed. The experimental results show that the two algorithms can reduce the search space and improve the efficiency of the algorithm.

Feature reduction is one of the core contents of rough set theory; in the condition of keeping the classification ability for knowledge base unchanged, we delete irrelevant or unimportant knowledge, which can reduce the dimension of the decision system, reduce the time complexity, and improve the efficiency of the algorithm [8]. People want to find the minimum reduction, but it has been proved to be an NP-Hard problem [9]; the main research is how to find the second optimal solution. Genetic algorithm is a computational model which is based on the natural selection and evolution mechanism; its core idea is inspired by the natural selection rule of the survival of the fittest, can achieve a highly parallel, random, and adaptive search, is not easy to fall into local optimal [10], can find the global optimal solution with high probability, and has great advantage in solving the NP-Hard problem.

In this paper, a new algorithm of PET/CT high-dimensional feature selection is proposed based on genetic algorithm and variable precision rough set model. On one hand, the algorithm considers the value of chromosome coding, the minimum number of attributes, and the dependency of attributes to construct a general fitness function framework and adjusting weight coefficient of each factor to achieve different fitness function; on the other hand, aiming at the limitation of Pawlak rough set model, introducing the classification error rate of , it extends rigid inclusion of the lower approximation for traditional rough set to partial inclusion, not only improving the concept of approximate space, but also enhancing the ability to deal with noise data and changing the range of β to achieve different fitness function. Finally, through extracting PET/CT lung cancer ROI 98-dimensional feature to construct the information decision table of lung cancer patients, 8 group experiments of high-dimensional features selection are done by using support vector machine to classify and recognize reduction subsets, to verify the degree of influence on the different weights and different classification error rate, and find a set of parameters suitable for this problem (; β = 0.6). The experimental results show that different parameters can be used to get different experimental results, so we should choose the appropriate parameter combination according to different problems so as to get better recognition accuracy.

2. Materials and Methods

2.1. PET/CT

PET/CT is a kind of advanced medical imaging technology, which is a combination of the good performance of PET and CT on the same device, and provides the anatomical and functional metabolism of the subjects under the same conditions [11]. PET is a functional image; it can provide metabolic information of tissue and organ and reflect functional changes of the human body from the molecular level, such as the physiological, pathological, biochemical, and metabolic, but has poor spatial resolution, cannot be accurately located, and cannot display the anatomical information of the lesions [12]. CT belongs to the anatomical structure of images, with high spatial resolution and density resolution; it has unique advantages in displaying the anatomical structure and density of the body [13], it also can provide detailed anatomical information of human organs and tissues, but can not reflect the functional information of tissues and organs [14] (Figure 1).

2.2. Genetic Algorithm

Genetic algorithm is a computational model which is based on the natural selection and evolution mechanism; its core idea is inspired by the natural selection rule of the survival for the fitness, so the search algorithm is an iterative process of survival and detection, is a very effective search and optimization technique, can achieve a highly parallel, random, and adaptive search, cannot easily fall into local optimum, and can find the global optimal solution with high probability and its robustness is good [15]. General use of genetic algorithm for reduction is achieved by a binary coding, 1 indicates that the position selects the corresponding attribute, while 0 indicates that the corresponding attribute is not selected. Genetic algorithm consists of four parts: encoding and decoding, fitness function, genetic operator, and control parameters, genetic operators include selection operator, crossover operator, and mutation operator, The selection operator is generally selected by roulette wheel selection method, according to the selection probability , crossover operator is a single point crossover, with a certain probability p to select individuals to participate in crossover, mutation operator selects the individual with the probability p and randomly selects the corresponding gene of the variant individuals to operate [16]. The general steps are as follows: determining the initial population and calculating the target value of each individual in the population and the corresponding value of the fitness function, choosing the chromosomes with high fitness value, and forming a matching set (selection), according to certain rules of reproduction (crossover and mutation), to meet the conditions to stop the genetic iteration, or return to step 3 (Figure 2).

2.3. Variable Precision Rough Set

Ziarko proposed the variable precision rough set model in 1993, he first proposed the concept of classification error rate; in the case of a given classification error rate, the objects with the same attributes can be classified into classes as many as possible [17].

Definition 1 (equivalence class). Assuming that R is an equivalence relation on K, a collection of all elements equivalent to an element k in K is called an equivalence class of k, denoted as [k].

Definition 2 (indiscernibility relation). If and , then is also an equivalence relation, it is called the indiscernibility relation on P, denoted as ind [P].

Definition 3 (upper approximation and lower approximation). The knowledge base of K = (, R), , R is equivalent to the relationship between . The lower approximation of X can be understood as all of the classification errors that are not greater than are included in the R equivalence class in X. The upper approximation of X can be understood as the intersection of all those with classification error not greater than and equivalence classes with X is not empty [18]. Expressions are as follows:Assume that the decision information table S = (, A, , f), where is a sample of the universe and a nonempty finite sample set,  = {x1, X2, x3,…, Xn}, and Xi represents each sample. A = , P represents a collection of conditional attributes, Q represents a set of decision attributes. represents the range of attribute. f: is an information function that gives an attribute value for each attribute of each xi, that is, . X and Y represent nonempty set in finite field . represents the condition attribute set and decision attribute set; ind (P), ind (Q) is an indiscernibility relation determined by P, Q; ind (P) is a collection of equivalence classes called condition class, expressed in , i.e.,  = {P1, P2, P3,…, Pn}; ind (P) is a collection of equivalence classes called decision class, expressed in , that is  = {Q1, Q2, Q3, …, Qn}.

Definition 4. (majority inclusion relation). If there is a for each , then Y contains X, denoted as Y X, thenwhere represents the cardinality of the set X and is the relative classification error rate of set X on set Y.
Make ; the majority inclusion relation is defined as ; the “majority” requirement implies that the number of common elements in X and Y is greater than 50% of the number of elements in X.

Definition 5. (-reduction). Conditional attribute set P is a subset of P for -reduction or approximate reduction of decision attribute set Q, and the subset is and meets the following two conditions [19]:(1)(2)To remove any attribute from , condition (1) is not valid

2.3.1. Attribute Dependency

Definition 6. The dependency of the decision attribute set Q and the conditional attribute set P is defined as, is the number of objects contained in the domain, is the number of objects contained in the positive domain of all equivalence classes that are not greater than the classification error, which indicates that the conditional attributes can correctly divide the object to . is the dependency of the decision attribute Q to the conditional attribute P and is an evaluation of the ability to classify objects with the classification error . = 0 means that P cannot be used to divide objects into equivalence classes in Q,  = 1 means that P can be used to divide objects into equivalence classes in Q completely, means that P can be used to divide objects into equivalence classes in Q partly.

2.3.2. Attribute Importance

Definition 7. Assume that the decision information table is S = (, A, , f), A = , sh, the relative importance of attribute s isAn attribute is able to distinguish an object; the greater the value, the stronger the ability.
The selection of threshold for variable precision rough set needs to meet the following requirements.
The choice of to make the classification accuracy as high as possible:(1)(2) makes the attributes contained in the reduction results as little as possible

2.4. SVM

Support vector machine (SVM) is a supervised learning model for data analysis, pattern recognition, and regression analysis in the field of machine learning. The best compromise between model complexity (the learning accuracy of a particular training sample) and learning ability (ability to identify an arbitrary sample without error) should be found based on limited sample information. In order to obtain the best generalization ability, the basic idea is to use the structural risk minimization principle to construct the optimal classification hyperplane in the attribute space. SVM has some advantages such as good generalization ability, simple data structure, low computational complexity, short training time, few parameters selection, high fitting precision, strong robustness, and so on [20, 21]. It has great advantages in dealing with small sample, nonlinear, and high-dimensional pattern recognition. It is often used in pattern recognition [12, 22], regression estimation, and so on.(1)After the introduction of kernel function and penalty parameter by SVM, the optimal discriminant function model isAmong it, 0 < a < C, yi ∈ {1, −1}.(2)The optimization function of SVM is(3)The radial basis kernel function is a widely used kernel function; the kernel function is used in this paper:

Among them, , is an important parameter in the kernel function, which affects the complexity of SVM classification algorithm.

The kernel function parameter and penalty coefficient C of support vector machine (SVM) is an important parameter which affects the performance of SVM classification, so (C, ) is used as the optimization variable. In the process of learning SVM, 5-fold cross validation is used to calculate the optimal classification performance of kernel function parameter and penalty coefficient, and then the diagnosis result of optimization is applied to the SVM classifier for lung cancer, the final selection of the sensitivity, specificity, accuracy, and computation time as the evaluation indexes of related experiments.

3. Results and Discussion

3.1. Main Idea

The main idea of the model is as follows.

3.1.1. Parameters

Population size M, chromosome length N (the number of condition attributes), crossover probability Pc, mutation probability Pm, fitness function , and the maximum number of iterations K are the parameters.

3.1.2. Coding

The binary coding method is used, which is represented by a binary string whose length is equal to the number of condition attributes; each bit corresponds to a condition attribute, a bit of 1 indicates that the corresponding condition attribute is selected, 0 indicates that the condition attribute is not selected, e.g., represents a chromosome with a length of 8, and it is known that the corresponding 1, 2, 5, 7 of 0 indicates that the corresponding condition attribute is not selected, then{c3, C4, C6, c8} is the last individual to choose the attributes set.

3.1.3. The Initial Population

Assuming the population size M (the number of chromosomes in the population is M), M length of Lr chromosome (0, 1) is the randomly generated as the initial population.

3.1.4. Genetic Operators

Genetic operators include selection operator, crossover operator, and mutation operator. The selection operator generally uses the roulette wheel selection method, according to the selection probability to select. Crossover operator uses a single-point crossover, with a certain probability Pc to select the individual uniform crossover. The mutation operator selects the individuals with the probability Pm to carry on the variation, and randomly selects the corresponding bit of the nonnuclear attribute.

3.1.5. Fitness Function

The fitness function is the core of the genetic algorithm, the fitness value is the only index to evaluate the fitness function; this paper from the gene encoding value, the minimum number of attributes reduction, attribute dependency, and other aspects constructs a fitness function framework, by adjusting the weights of various factors and changing the classification error rate to achieve different fitness function. The fitness function is set as follows:target1. Attribute dependency: , it represents the dependency of decision attribute Q for conditional attribute P.target2. Attributes reduction Length: , is the number of condition attributes represented by 0, 1. represents the number of 1 in attribute P, the shorter the better results.target3. Gene coding weight function: i.e., Penalty function, target3 = . Gene values can only take 0 and 1, but the chromosome will show not 1 and not 0, such that the value is less than 0 or greater than 1. The value must be punished; therefore, the gene coding weight function is constructed. If the gene is 0, , but the gene is 1, . So do not punish the genes of 0 or 1, but if there is a chromosome with a length of 6: r = [0 0 −2 −1 2 1],  = [−1 −1 −3 −2 1 0],  = [0 0 6 2 2 0], then  = 10, length is 6, so target3 = 10/6 = 1.67.

Therefore, the fitness function constructed in this paper iswhere is the weight coefficient of fitness function,  = (0, 1, 2, 3), because the genetic algorithm can only find the minimum value, and the bigger the fitness value, the better it is, so the objective function is minus and the penalty function is plus.

Flow chart about this model is given in Figure 3.

3.2. Model Concrete Steps

Input: A decision information table S = (, A, , f)Output: generate (M), Lr = 98//Initial population M, 01 sequence of Chromosome length 98Setting , , crossover probability Pc, mutation probability Pm, iteration number KBegin for i = 1 : K target1 =  target2 =  target3 = Fitness function = ;//Fitness functionP = Select (M, 2, Pc);//Crossover probability PcQ = Crossover (P, 2, Pc);//Crossover algorithmQ′ = Mutation (Q, Pm);//Mutation algorithmEnd

3.3. Experimental Environment and Data
3.3.1. Hardware Environment

Intel Core i5 4670-3.4 GHz with 8.0 GB memory and 500 GB hard disk were used.

3.3.2. Software Environment

Matlab R2012b, LibSVM, and Windows 7 operating system were used.

3.3.3. Experimental Data

The PET/CT images of 2000 lung cancer patients were collected as the study samples (1000 cases of benign lung tumor, 1000 cases of malignant lung tumor). Firstly, ROI was extracted from the lung tumor and pretreated; then 8-dimensional shape features, 7-dimensional gray features, 3-dimensional Tamura features, 56-dimensional GLCM features, and the 24-dimensional frequency domain features were extracted from the lung tumor ROI, and 98-dimensional feature vectors are discreted and normalized. In the decision attribute, 1 represents the lung malignant tumor and −1 represents the lung benign tumor. Figure 4(a) shows four PET/CT images, ROI of Lung malignant tumor, and Figure 4(b) shows four PET/CT images, ROI of lung benign tumor. Table 1 gives the feature values of two patients with lung cancer (one patient was a malignant tumor and the other was a benign tumor).

3.4. Analysis of Experimental Results

In this paper, 3 kinds of experiments are designed according to the weight value of the fitness function and the classification error rate (namely, the inclusion degree ). For first type of experiments, , according to the different values of to do the three groups of experiments totally. For second type of experiments, , , according to the different values of to do three groups of experiments. For third type of experiments, , by increasing the value to achieve the best fitness function, to achieve the best results.

3.4.1. Experiment 1—Research on Different Weight Coefficients under the Condition in the Same Classification Error Rate

values and values of experiment 1 are shown in Table 2, and  = 0.6.

(1) 1st Group Experiment.  = {1, 0, 0}: The algorithm is run 5 times according to this group weights; the results of the 5 groups are given in Table 3, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency degree, and the time. The convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 5.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples; the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 4.

The weight value in this group experiment is  = {1, 0, 0}, attribute dependency degree are regarded as fitness function. The average attribute dependency degree is 0.973, the average length of reduction is 39.2, and the average optimal fitness value is −0.973. The average recognition accuracy of the experiment is 96.92%. The premature phenomena are shown in Figure 5, and evolution progress is terminated early.

(2) 2nd Group Experiment.  = {1, 1, 0}: This experiment introduces an objective function to control the length of reduction (the shorter the length of reduction, the better it is), the influence degree of the objective function which controls the reduction on the fitness function and the final recognition accuracy is verified. The algorithm is run 5 times according to this group weights, and the results of the 5 groups are given in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 6.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples, the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 6.

The experimental weight of this group is  = {1, 1, 0}; attribute dependency degree and the length of reduction are regarded as fitness function. As can be seen from Table 5, the reduction length is 13, 17, and so on. The average length of reduction was 17, which was significantly reduced compared with the average length of the reduction in Table 3 in the experiment of the first groups, which reduced the time, improved the efficiency of the algorithm, and increased the attribute dependency of the algorithm, even up to 1. The average recognition accuracy of the experimental group was 96.98%, which was increased by 0.06% compared with that in the first groups.

(3) 3rd Group Experiment.  = {1, 1, 1}: On the basis of attribute dependency degree and attribute reduction length, this experiment introduces gene coding weight function, in order to verify the effect of gene coding weight function on fitness function and the final recognition accuracy. The algorithm are run 5 times according to this group weights, the results of the 5 groups are given in Table 7, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 7.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples; the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 8.

The experimental weight of this group is  = {1, 1, 1}; attribute dependency degree, the length of reduction, and gene coding weight value are regarded as fitness function. However, the premature phenomena are shown in Figure 7, and evolution progress is terminated early. From Table 7, we can see that the attribute dependency decreases gradually and even the attribute dependency of Reduction 1 is reduced to 0.759. The average recognition accuracy of the experimental group was 96.85%. The accuracy of recognition was decreased compared with the second groups, and hence, using gene encoding weight function in the fitness function to improve recognition accuracy is useless, only for the samples with different results, to analyze specific issues. The 3 experiment runs of experiment 1 verified the necessity of the fitness function, by continuously introducing fitness objective function, such as target1, target2, and target3, the conclusion is that the fitness function is better when it is not bigger, but after the introduction of target3, the accuracy declines; therefore, the introduction of target1 and target2 in this algorithm can get better results.

3.4.2. Experiment 2—Research on Different Classification Error Rates under the Condition in the Same Weight Coefficient

According to experiment 1, we can see that when , , , the experimental results are the best, so in experiment 2, the case of the weight value of was unchanged, , , , and the value was changed, and they are shown in Table 9.

(1) 1st Group Experiment.  = 0.6: The algorithm is run 5 times according to this group weights; the results of the 5 groups are shown in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 6 (i.e., not repeated in the second group of experiment 1).

(2) 2nd Group Experiment.  = 0.8: The algorithm are is 5 times according to this group weights, the results of the 5 groups are given in Table 10, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 8.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples; the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 11.

In the case of constant weight and classification error rate of  = 0.2 (which contains  = 0.8), the classification error rate is changed on the basis of the attribute dependency and the length of control reduction. The premature phenomena are shown in Figure 8, and evolution progress is terminated early, such that the attribute dependency of reduction 5 was 0.8405, appeared in attribute dependency on less than 0.9, compared with first groups of experiment 2 attribute dependency declined. The average recognition accuracy of the experimental group was 96.74%. Compared with the classification error rate of 0.4 and inclusion degree of 0.6, the accuracy was decreased by 0.24%.

(3) 3rd Group Experiment.  = 1: The algorithm is run 5 times according to the group weights, and the results of the 5 groups are shown in Table 12, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 9.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples, the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 13.

In the case of constant weight and classification error rate of  = 0 (which contains  = 1), the classification error rate is reduced and the inclusion degree is improved on the basis of the attribute dependency and the length of control reduction. The premature phenomena are shown in Figure 9, and evolution progress is terminated early. The average recognition accuracy of the experimental group was 95.73%, which was decreased by 0.06% compared with that in the second group. In experiment 2, the effect of changing the classification error rate on the recognition accuracy was verified by the 3 groups of experiments. By continually reducing the classification error rate, the final recognition accuracy has been declining, when inclusion degree is 1,  = 0; variable precision rough set becomes Pawlak rough set, the recognition accuracy of the recognition accuracy of is minimum, which was verified the advantages of variable precision rough set.

3.4.3. Experiment 3—Research on Increasing the Weight Coefficient under the Condition in the Same Classification Error Rate

According to experiment 1 and experiment 2, we can know that when  = 0.6, , , , the recognition accuracy is the best; therefore, in the third experiment, by increasing the weight of , 3 groups of experiments are performed, fitness goals: target 1 (attribute dependency), target 2 (the minimum number of attributes reduction), and target 3 (gene encoding weight function), the three objective functions play an important role in the evaluation of fitness function. However, the importance of fitness function is reduced in these three objectives. Therefore, in this experiment, when the other conditions are unchanged, the weight coefficient of the target 1 is increased, to verify the influence of the change in the weight coefficient on the experimental results, and they are shown in Table 14.

(1) 1st Group Experiment. , , : The algorithm are run 5 times according to this group weights, the results of the 5 groups are given in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency and the time. The convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 6 (i.e., not repeated for the second group of experiment 1).

(2) 2nd Group Experiment. , , : The algorithm is run 5 times according to this group weights, the results of the 5 groups are given in Table 15, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 10.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples, the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 16.

The experimental group in the case of  = 0.6 unchanged, the weight of was increased. The attribute dependency and the optimal fitness function are relatively high; the average precision of reduction 1, reduction 2, and reduction 3 in Table 16 is more than 97% and that of reduction 2 is even up to 97.25%. The average recognition accuracy of the experimental group is 97.03%, which is higher than that of the first group.

(3) 3rd Group Experiment. , , : The algorithm is run 5 times according to the group weights, and the results of the 5 groups are given in Table 17, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. The convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 11.

Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples; the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 18.

The experiment in the case of  = 0.6 unchanged, increase the weight of the . , , . In Table 18, the recognition accuracy of reduction 1 is over 97% only. The average recognition accuracy of the experimental group is 96.75%, and the accuracy is decreased by 0.28% compared with that in the other second groups. The three groups of experiment 3, the purpose is to verify whether the experimental accuracy is influenced by increasing the weight. The experimental results show that the accuracy of second group for experimental weight {2, 1, 0} compared with first group experiment when {1, 1, 0} is high, but the third groups of experimental weight was {3, 1, 0}, and the accuracy declined; therefore, the weight of {2, 1, 0} is the best choice for this experiment.

4. Conclusions

Aiming at the deficiency of the traditional rough set model, this paper proposes a new feature selection model based on genetic algorithm and variable precision rough set; by introducing the value, the rigid inclusion of the approximation for the traditional rough set is relaxed, and then we design the 3 kinds of experiments by constructing the decision information table of the PET/CT feature for lung tumor ROI. The first type of experiment is the inclusion of , and different values of made a total of three groups of experiments; the experimental results show that the better recognition accuracy can be obtained when the weight is {1, 1, 0}, and the results show that the gene coding weight function has no effect on the fitness function. For the second type of experiments, , , according to the different values of to do three groups of experiments, the results show that the recognition accuracy of is the best, which shows that the larger the value is, the lower approximate cardinality will be larger, then the relative accuracy will increase. For the third type of experiments, , value is increased to achieve the best fitness function, in order to achieve the best results. The experimental results show that the recognition accuracy is better than the others when the weight value is {2, 1, 0}, so it is better to solve the problem by increasing the proportion of attribute dependency. Through the above experiments, it is shown that the high-dimensional feature selection algorithm based on genetic algorithm and variable precision rough set can solve the multiobjective optimization problem well. However, when the fitness function and its parameters are applied in the specific application, it is necessary to analyze the specific problems.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Acknowledgments

This work was supported by National Natural Science Foundation of China (Grant nos. 61561040 and 61876121), First-Class Disciplines Foundation of Ningxia (No. NXYLXK2017B09).