A Model of High-Dimensional Feature Reduction Based on Variable Precision Rough Set and Genetic Algorithm in Medical Image

School of Computer Science and Engineering, North Minzu University, Yinchuan 750021, China School of Science, Ningxia Medical University, Yinchuan 750004, China School of Electronic & Information Engineering, Suzhou University of Science and Technology, Suzhou 215009, China Key Laboratory of Spectral Imaging Technology CAS, Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China Human Resources Department, 4e Second Affiliated Hospital of Xiamen Medical College, Xiamen 361021, China


Introduction
Rough set theory was developed by Pawlak in 1982 [1], and it is a mathematical tool to deal with vagueness and uncertainty. e classification ability unchanged in its main idea, decision or classification rules of problem are derived by knowledge reduction [2]. e Variable Precision Rough Set (VPRS) theory, proposed by Ziarko, and is an extension of original rough set model. For inconsistent information system, the VPRS model allows a flexible approximation boundary region by a precision variable β [3]. When β � 0, Pawlak rough set model is a special case of variable precision rough set model. e main task of variable precision rough set model is to solve the problem of data classification with no function or uncertainty. e hierarchical model of attribute reduction for variable precision rough set is studied by Xiaowei [4]. ere is abnormal phenomenon in existing attribute reduction models; therefore, a variable precision rough set attribute reduction algorithm with the property of interval is proposed, and the reduction abnormal problem is transformed into a hierarchical model representation, and the reduction anomaly is gradually eliminated by the layer-by-layer reduction model; Jie and Jiayang [5] puts forward that there may be a reduction jump phenomenon in variable precision rough set feature reduction, which affects the quality of reduction and brings the problem of attribute reduction of variable precision rough set; Pei and Qinghua [6] proposes an FCM clustering algorithm based on variable precision rough set; according to the threshold characteristics of the variable precision rough set model, the algorithm divides the objects in the edge of the cluster into the positive, negative, and boundary regions, to improve the accuracy of clustering. Two different solutions of variable precision rough set attribute reduction algorithm are proposed by Hao and Junan [7]; based on tolerance matrix and minimal reduction of attribute core, the attribute kernel idea of variable precision rough set is proposed. e experimental results show that the two algorithms can reduce the search space and improve the efficiency of the algorithm.
Feature reduction is one of the core contents of rough set theory; in the condition of keeping the classification ability for knowledge base unchanged, we delete irrelevant or unimportant knowledge, which can reduce the dimension of the decision system, reduce the time complexity, and improve the efficiency of the algorithm [8]. People want to find the minimum reduction, but it has been proved to be an NP-Hard problem [9]; the main research is how to find the second optimal solution. Genetic algorithm is a computational model which is based on the natural selection and evolution mechanism; its core idea is inspired by the natural selection rule of the survival of the fittest, can achieve a highly parallel, random, and adaptive search, is not easy to fall into local optimal [10], can find the global optimal solution with high probability, and has great advantage in solving the NP-Hard problem.
In this paper, a new algorithm of PET/CT high-dimensional feature selection is proposed based on genetic algorithm and variable precision rough set model. On one hand, the algorithm considers the value of chromosome coding, the minimum number of attributes, and the dependency of attributes to construct a general fitness function framework and adjusting weight coefficient of each factor to achieve different fitness function; on the other hand, aiming at the limitation of Pawlak rough set model, introducing the classification error rate of β, it extends rigid inclusion of the lower approximation for traditional rough set to partial inclusion, not only improving the concept of approximate space, but also enhancing the ability to deal with noise data and changing the range of β to achieve different fitness function. Finally, through extracting PET/CT lung cancer ROI 98-dimensional feature to construct the information decision table of lung cancer patients, 8 group experiments of high-dimensional features selection are done by using support vector machine to classify and recognize reduction subsets, to verify the degree of influence on the different weights and different classification error rate, and find a set of parameters suitable for this problem (ϖ1 � 1, ω2 � 1, ω3 � 0; β � 0.6). e experimental results show that different parameters can be used to get different experimental results, so we should choose the appropriate parameter combination according to different problems so as to get better recognition accuracy.

PET/CT.
PET/CT is a kind of advanced medical imaging technology, which is a combination of the good performance of PET and CT on the same device, and provides the anatomical and functional metabolism of the subjects under the same conditions [11]. PET is a functional image; it can provide metabolic information of tissue and organ and reflect functional changes of the human body from the molecular level, such as the physiological, pathological, biochemical, and metabolic, but has poor spatial resolution, cannot be accurately located, and cannot display the anatomical information of the lesions [12]. CT belongs to the anatomical structure of images, with high spatial resolution and density resolution; it has unique advantages in displaying the anatomical structure and density of the body [13], it also can provide detailed anatomical information of human organs and tissues, but can not reflect the functional information of tissues and organs [14] (Figure 1).

Genetic Algorithm.
Genetic algorithm is a computational model which is based on the natural selection and evolution mechanism; its core idea is inspired by the natural selection rule of the survival for the fitness, so the search algorithm is an iterative process of survival and detection, is a very effective search and optimization technique, can achieve a highly parallel, random, and adaptive search, cannot easily fall into local optimum, and can find the global optimal solution with high probability and its robustness is good [15]. General use of genetic algorithm for reduction is achieved by a binary coding, 1 indicates that the position selects the corresponding attribute, while 0 indicates that the corresponding attribute is not selected. Genetic algorithm consists of four parts: encoding and decoding, fitness function, genetic operator, and control parameters, genetic operators include selection operator, crossover operator, and mutation operator, e selection operator is generally selected by roulette wheel selection method, according to the selection probability p i � (f i / M i�1 f i ), crossover operator is a single point crossover, with a certain probability p to select individuals to participate in crossover, mutation operator selects the individual with the probability p and randomly selects the corresponding gene of the variant individuals to operate [16]. e general steps are as follows: determining the initial population and calculating the target value of each individual in the population and the corresponding value of the fitness function, choosing the chromosomes with high fitness value, and forming a matching set (selection), according to certain rules of reproduction (crossover and mutation), to meet the conditions to stop the genetic iteration, or return to step 3 ( Figure 2).

Variable Precision Rough Set.
Ziarko proposed the variable precision rough set model in 1993, he first proposed the concept of classification error rate; in the case of a given classification error rate, the objects with the same attributes can be classified into classes as many as possible [17].
Definition 1 (equivalence class). Assuming that R is an equivalence relation on K, a collection of all elements equivalent to an element k in K is called an equivalence class of k, denoted as [k].
Definition 2 (indiscernibility relation). If P ⊆ R and P ≠ ϕ, then ∩ P is also an equivalence relation, it is called the indiscernibility relation on P, denoted as ind [P].
Definition 3 (upper approximation and lower approximation). e knowledge base of K � (U, R), X ⊆ U, R is equivalent to the relationship between U. e lower approximation of X can be understood as all of the classification errors that are not greater than β are included in the R equivalence class in X. e upper approximation of X can be understood as the intersection of all those with classification error not greater than β and equivalence classes with X is not empty [18]. Expressions are as follows: Assume that the decision information table S � (U, A, V, f ), where U is a sample of the universe and a nonempty finite sample set, U � {x1, X2, x3,. . ., Xn}, and Xi represents each sample. A � P ∪ Q, P represents a collection of conditional attributes, Q represents a set of decision attributes. V represents the range of attribute. f: U × A ⟶ V is an information function that gives an attribute value for each attribute of each and Y represent nonempty set in finite field U. P, Q ⊆ A represents the condition attribute set and decision attribute set; ind (P), ind (Q) is an indiscernibility relation determined by P, Q; ind (P) is a collection of equivalence classes called condition class, expressed in U/P, i.e., U/ind(P) � {P 1 , P 2 , P 3 ,. . ., P n }; ind (P) is a collection of equivalence classes called decision class, expressed in U/Q, that is U/ind(Q) � {Q 1 , Q 2 , Q 3 , . . ., Q n }.
Definition 4 (majority inclusion relation). If there is a e ∈ Y for each e ∈ X, then Y contains X, denoted as Y X, then where |X| represents the cardinality of the set X and c(X, Y) is the relative classification error rate of set X on set Y. Make (0 ≤ β < 0.5); the majority inclusion relation is defined as Y⊇ β X ⟺ c(X, Y) ≤ β; the "majority" requirement implies that the number of common elements in X and Y is greater than 50% of the number of elements in X.
Definition 5 (β-reduction). Conditional attribute set P is a subset of P for β-reduction or approximate reduction of decision attribute set Q, and the subset is red(P, Q, β) and meets the following two conditions [19]: (1) c(P, Q, β) � c(red(P, Q, β), Q, β) (2) To remove any attribute from red(P, Q, β), condition (1) is not valid  e dependency of the decision attribute set Q and the conditional attribute set P is defined as pos(P, Q, β) � ∪ Y∈U/Qind(P) β Q, |U| is the number of objects contained in the domain, |pos(P, Q, β)| is the number of objects contained in the positive domain of all equivalence classes that are not greater than the β classification error, which indicates that the conditional attributes can correctly divide the object to U|Q. c(0 ≤ c ≤ 1) is the β dependency of the decision attribute Q to the conditional attribute P and is an evaluation of the ability to classify objects with the classification error β.c � 0 means that P cannot be used to divide objects into equivalence classes in Q, c � 1 means that P can be used to divide objects into equivalence classes in Q completely, 0 < c < 1 means that P can be used to divide objects into equivalence classes in Q partly.

Attribute Importance
An attribute is able to distinguish an object; the greater the value, the stronger the ability. e selection of threshold β for variable precision rough set needs to meet the following requirements. e choice of β to make the classification accuracy as high as possible: (1) 0 ≤ β < 0.5 (2) β makes the attributes contained in the reduction results as little as possible 2.4. SVM. Support vector machine (SVM) is a supervised learning model for data analysis, pattern recognition, and regression analysis in the field of machine learning. e best compromise between model complexity (the learning accuracy of a particular training sample) and learning ability (ability to identify an arbitrary sample without error) should be found based on limited sample information. In order to obtain the best generalization ability, the basic idea is to use the structural risk minimization principle to construct the optimal classification hyperplane in the attribute space. SVM has some advantages such as good generalization ability, simple data structure, low computational complexity, short training time, few parameters selection, high fitting precision, strong robustness, and so on [20,21]. It has great advantages in dealing with small sample, nonlinear, and high-dimensional pattern recognition. It is often used in pattern recognition [12,22], regression estimation, and so on.
(1) After the introduction of kernel function and penalty parameter by SVM, the optimal discriminant function model is Among it, 0 < a < C, y i ∈ {1, − 1}. (2) e optimization function of SVM is (3) e radial basis kernel function is a widely used kernel function; the kernel function is used in this paper: Among them, g > 0, g is an important parameter in the kernel function, which affects the complexity of SVM classification algorithm. e kernel function parameter g and penalty coefficient C of support vector machine (SVM) is an important parameter which affects the performance of SVM classification, so (C, g) is used as the optimization variable. In the process of learning SVM, 5-fold cross validation is used to calculate the optimal classification performance of kernel function parameter and penalty coefficient, and then the diagnosis result of optimization is applied to the SVM classifier for lung cancer, the final selection of the sensitivity, specificity, accuracy, and computation time as the evaluation indexes of related experiments.

Main Idea.
e main idea of the model is as follows.

Parameters.
Population size M, chromosome length N (the number of condition attributes), crossover probability Pc, mutation probability Pm, fitness function F(x), and the maximum number of iterations K are the parameters.

3.1.2.
Coding. e binary coding method is used, which is represented by a binary string whose length is equal to the number of condition attributes; each bit corresponds to a condition attribute, a bit of 1 indicates that the corresponding condition attribute is selected, 0 indicates that the condition attribute is not selected, e.g., 00110101 { } represents a chromosome with a length of 8, and it is known that the corresponding 1, 2, 5, 7 of 0 indicates that the corresponding condition attribute is not selected, then{c3, C4, C6, c8} is the last individual to choose the attributes set.

e Initial Population.
Assuming the population size M (the number of chromosomes in the population is M), M length of Lr chromosome (0, 1) is the randomly generated as the initial population.

Genetic Operators.
Genetic operators include selection operator, crossover operator, and mutation operator. e selection operator generally uses the roulette wheel selection method, according to the selection probability p i � (f i / M i�1 f i ) to select. Crossover operator uses a single-point crossover, with a certain probability Pc to select the individual uniform crossover. e mutation operator selects the individuals with the probability Pm to carry on the variation, and randomly selects the corresponding bit of the nonnuclear attribute.

Fitness Function.
e fitness function is the core of the genetic algorithm, the fitness value is the only index to evaluate the fitness function; this paper from the gene encoding value, the minimum number of attributes reduction, attribute dependency, and other aspects constructs a fitness function framework, by adjusting the weights of various factors and changing the classification error rate to achieve different fitness function. e fitness function is set as follows: target1.
Attribute dependency: c(P, Q, β) � (|pos β P (Q)|/|U|), it represents the β dependency of decision attribute Q for conditional attribute P. target2. Attributes reduction Length: |C reduct| �((|P | − |Lr|)/|P|), |P| is the number of condition attributes represented by 0, 1. |Lr| represents the number of 1 in attribute P, the shorter the better results. target3. Gene coding weight function: i.e., Penalty function, target3 � abs(r × (r − 1))/|r|. Gene values can only take 0 and 1, but the chromosome will show not 1 and not 0, such that the value is less than 0 or greater than 1. e value must be punished; therefore, the gene coding weight function is constructed. If the gene is 0, r × (r − 1) � 0, but the gene is 1, r × (r − 1) � 0. So do not punish the genes of 0 or 1, but if there is a chromosome with a length of 6: abs(r × (r − 1)) � 10, length is 6, so target3 � 10/6 � 1.67. erefore, the fitness function constructed in this paper is where ω is the weight coefficient of fitness function, ω � (0, 1, 2, 3), because the genetic algorithm can only find the minimum value, and the bigger the fitness value, the better it is, so the objective function is minus and the penalty function is plus.
Flow chart about this model is given in Figure 3.

Hardware Environment.
Intel Core i5 4670-3.4 GHz with 8.0 GB memory and 500 GB hard disk were used.

Software Environment.
Matlab R2012b, LibSVM, and Windows 7 operating system were used.

Experimental Data.
e PET/CT images of 2000 lung cancer patients were collected as the study samples (1000 cases of benign lung tumor, 1000 cases of malignant lung tumor). Firstly, ROI was extracted from the lung tumor and pretreated; then 8-dimensional shape features, 7-dimensional gray features, 3-dimensional Tamura features, 56-dimensional GLCM features, and the 24-dimensional frequency domain features were extracted from the lung tumor ROI, and 98-dimensional feature vectors are discreted and normalized. In the decision attribute, 1 represents the lung malignant tumor and − 1 represents the lung benign tumor. Figure 4(a) shows four PET/CT images, ROI of Lung malignant tumor, and Figure 4(b) shows four PET/CT images , ROI of lung benign tumor. Table 1 gives the feature values of two patients with lung cancer (one patient was a malignant tumor and the other was a benign tumor). value to achieve the best fitness function, to achieve the best results.

Experiment 1-Research on Different Weight Coefficients under the Condition in the Same Classification Error
Rate. ϖ values and β values of experiment 1 are shown in Table 2, and 1 − β � 0.6.
(1) 1st Group Experiment. ϖ1, ω2, ω3 { } � {1, 0, 0}: e algorithm is run 5 times according to this group weights; the results of the 5 groups are given in Table 3, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency degree, and the time. e convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 5.
Each reduction is classified by SVM classifier, using the method of 5 Initial population Fitness value    (2) 2nd Group Experiment. ϖ1, ω2, ω3 { } � {1, 1, 0}: is experiment introduces an objective function to control the length of reduction (the shorter the length of reduction, the better it is), the influence degree of the objective function which controls the reduction on the fitness function and the final recognition accuracy is verified. e algorithm is run 5 times according to this group weights, and the results of the 5 groups are given in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. e convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 6.
Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples, the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by    Table 6. e experimental weight of this group is ϖ1, { ω2, ω3} � {1, 1, 0}; attribute dependency degree and the length of reduction are regarded as fitness function. As can be seen from Table 5, the reduction length is 13, 17, and so on. e average length of reduction was 17, which was significantly reduced compared with the average length of the reduction in Table 3 in the experiment of the first groups, which reduced the time, improved the efficiency of the algorithm, and increased the attribute dependency of the algorithm, even up to 1. e average recognition accuracy of the experimental group was 96.98%, which was increased by 0.06% compared with that in the first groups. e algorithm are run 5 times according to this group weights, the results of the 5 groups are given in Table 7, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time.
e convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 7.  Table 8. e experimental weight of this group is ϖ1, { ω2, ω3} = {1, 1, 1}; attribute dependency degree, the length of reduction, and gene coding weight value are regarded as fitness function. However, the premature phenomena are shown in Figure 7, and evolution progress is terminated early. From Table 7, we can see that the attribute dependency decreases gradually and even the attribute dependency of Reduction 1 is reduced to 0.759. e average recognition accuracy of the experimental group was 96.85%. e accuracy of recognition was decreased compared with the second groups, and hence, using gene encoding weight function in the fitness function to improve recognition accuracy is useless, only for the samples with different results, to analyze specific issues. e 3 experiment runs of experiment 1 verified the necessity of the fitness function, by continuously introducing fitness objective function, such as target1, target2, and target3, the conclusion is that the fitness function is better when it is not bigger, but after the introduction of target3, the accuracy declines; therefore, the introduction of target1 and target2 in this algorithm can get better results.

Experiment 2-Research on Different Classification Error Rates under the Condition in the Same Weight
Coefficient. According to experiment 1, we can see that when ω1 � 1, ω2 � 1, ω3 � 0, the experimental results are the best, so in experiment 2, the case of the weight value of was unchanged, ω1 � 1, ω2 � 1, ω3 � 0, and the β value was changed, and they are shown in Table 9.
(1) 1st Group Experiment. 1 − β � 0.6: e algorithm is run 5 times according to this group weights; the results of the 5 groups are shown in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. e convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 6 (i.e., not repeated in the second group of experiment 1).
(2) 2nd Group Experiment. 1 − β � 0.8: e algorithm are is 5 times according to this group weights, the results of the 5 groups are given in Table 10, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. e  convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 8.    Table 9: Fitness function weight proportion and β value.
changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 11. In the case of constant weight ϖ and classification error rate of β � 0.2 (which contains 1 − β � 0.8), the classification error rate is changed on the basis of the attribute dependency and the length of control reduction. e premature phenomena are shown in Figure 8, and evolution progress is terminated early, such that the attribute dependency of reduction 5 was 0.8405, appeared in attribute dependency on less than 0.9, compared with first groups of experiment 2 attribute dependency declined. e average recognition accuracy of the experimental group was 96.74%. Compared with the classification error rate of 0.4 and inclusion degree of 0.6, the accuracy was decreased by 0.24%.
(3) 3rd Group Experiment. 1 − β � 1: e algorithm is run 5 times according to the group weights, and the results of the 5 groups are shown in Table 12, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. e convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 9.
Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples, the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 13.
In the case of constant weight ϖ and classification error rate of β = 0 (which contains 1 − β = 1), the classification error rate is reduced and the inclusion degree is improved on the basis of the attribute dependency and the length of control reduction. e premature phenomena are shown in Figure 9, and evolution progress is terminated early. e average recognition accuracy of the experimental group was 95.73%, which was decreased by 0.06% compared with that in the second group. In experiment 2, the effect of changing the classification error rate on the recognition accuracy was verified by the 3 groups of experiments. By continually reducing the classification error rate, the final recognition accuracy has been declining, when inclusion degree is 1, β = 0; variable precision rough set becomes Pawlak rough set, the recognition accuracy of the recognition accuracy of is minimum, which was verified the advantages of variable precision rough set.

Experiment 3-Research on Increasing the Weight Coefficient under the Condition in the Same Classification
Error Rate. According to experiment 1 and experiment 2, we can know that when β � 0.6, ω1 � 1, ω2 � 1, ω3 � 0, the recognition accuracy is the best; therefore, in the third experiment, by increasing the weight of ω, 3 groups of experiments are performed, fitness goals: target 1 (attribute     dependency), target 2 (the minimum number of attributes reduction), and target 3 (gene encoding weight function), the three objective functions play an important role in the evaluation of fitness function. However, the importance of fitness function is reduced in these three objectives. erefore, in this experiment, when the other conditions are unchanged, the weight coefficient of the target 1 is increased, to verify the influence of the change in the weight coefficient on the experimental results, and they are shown in Table 14.
(1) 1st Group Experiment. ω1 � 1, ω2 � 1, ω3 � 0: e algorithm are run 5 times according to this group weights, the results of the 5 groups are given in Table 5, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency and the time. e convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 6 (i.e., not repeated for the second group of experiment 1).
(2) 2nd Group Experiment. ω1 � 2, ω2 � 1, ω3 � 0: e algorithm is run 5 times according to this group weights, the results of the 5 groups are given in Table 15, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time.
e convergence of VPRS-GA under the variation for fitness function value in one time is shown in Figure 10.  Table 16.
e experimental group in the case of 1 − β = 0.6 unchanged, the weight of ω1 was increased. e attribute dependency and the optimal fitness function are relatively high; the average precision of reduction 1, reduction 2, and reduction 3 in Table 16 is more than 97% and that of reduction 2 is even up to 97.25%. e average recognition accuracy of the experimental group is 97.03%, which is higher than that of the first group.        (3) 3rd Group Experiment. ω1 � 3, ω2 � 1, ω3 � 0: e algorithm is run 5 times according to the group weights, and the results of the 5 groups are given in Table 17, including the reduction of the conditional attributes, the length of reduction, the optimal fitness value, the attribute dependency, and the time. e convergence of VPRS-GA under the variation for fitness function value for one time is shown in Figure 11.
Each reduction is classified by SVM classifier, using the method of 5-fold cross validation, through changing the training samples and test samples; the results of the five groups were obtained, including accuracy, sensitivity, specificity, and time (training samples are constructed by 800 malignant samples and 800 benign samples; testing samples are constructed by 200 malignant samples and 200 benign samples; the experiment is repeated 5 times by changing training samples and testing samples). Finally, the average values of these five groups are obtained as the final result of the reduction and shown in Table 18.
e experiment in the case of 1 − β � 0.6 unchanged, increase the weight of the ω1. ω1 � 3, ω2 � 1, ω3 � 0. In Table 18, the recognition accuracy of reduction 1 is over 97% only. e average recognition accuracy of the experimental group is 96.75%, and the accuracy is decreased by 0.28% compared with that in the other second groups. e three groups of experiment 3, the purpose is to verify whether the experimental accuracy is influenced by increasing the weight. e experimental results show that the accuracy of second group for experimental weight {2, 1, 0} compared with first group experiment when {1, 1, 0} is high, but the third groups of experimental weight was {3, 1, 0}, and the accuracy declined; therefore, the weight of {2, 1, 0} is the best choice for this experiment.

Conclusions
Aiming at the deficiency of the traditional rough set model, this paper proposes a new feature selection model based on genetic algorithm and variable precision rough set; by introducing the β value, the rigid inclusion of the approximation for the traditional rough set is relaxed, and then we design the 3 kinds of experiments by constructing the decision information table of the PET/CT feature for lung tumor ROI. e first type of experiment is the inclusion of 1 − β � 0.6, and different values of ω made a total of three groups of experiments; the experimental results show that the better recognition accuracy can be obtained when the weight is {1, 1, 0}, and the results show that the gene coding weight function has no effect on the fitness function. For the second type of experiments, ϖ1 � 1, ω2 � 1, ω3 � 0, according to the different values of β to do three groups of experiments, the results show that the recognition accuracy of 1 − β � 0.6 is the best, which shows that the larger the β value is, the lower approximate cardinality will be larger, then the relative accuracy will increase. For the third type of experiments, 1 − β � 0.6, ω value is increased to achieve the best fitness function, in order to achieve the best results. e experimental results show that the recognition accuracy is better than the others when the weight value is {2, 1, 0}, so it is better to solve the problem by increasing the proportion of attribute dependency. rough the above experiments, it is shown that the high-dimensional feature selection algorithm based on genetic algorithm and variable precision rough set can solve the multiobjective optimization problem well. However, when the fitness function and its parameters are applied in the specific application, it is necessary to analyze the specific problems.

Data Availability
e data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e authors declare no conflicts of interest.