Abstract

Attribute reduction is a popular approach of preprocessing data. Discernibility matrix is a typical method that focuses on attribute reduction. Faced with the processing of modern information systems with large amounts of data and rapid changes, the traditional static discernibility matrix reduction model is powerless. To overcome this shortcoming, this paper first proposes an indistinguishable element pair method that does not need to store discernibility information, which retains the advantages of institution and easy-to-understand, and at the same time effectively solves the problem of space consumption. In order to make the model adapt to the processing of dynamic data sets, we further study the incremental mechanism and design a set of dynamic reduction models, which can adjust the reduction set in time according to the changes of objects. Theoretical analysis and experimental results indicate that the proposed algorithm is obviously superior to the discernibility matrix model, and can effectively deal with the reduction of dynamic data sets.

1. Introduction

Rough sets theory (RST) is a valid mathematical tool, which was proposed by Pawlak and Skowron in 1982, for dealing with inaccurate, incomplete, and vague information [1]. RST has been widely used in many fields such as machine learning [2], data mining [3], decision supporting [4], expert system [5], pattern recognition [6], and music emotions annotation [7]. Attribute reduction is one of the hot research focuses in RST [8], which aims to delete redundant data, while keeping the distinguishing power of the original data in information systems. For the convenience of the following description, Table 1 summarizes the list of abbreviations in the article. In the last two decades, many heuristic attribute reduction approaches have been developed based on the positive region [9], discernibility matrix [10, 11], information entropy [12], fuzzy rough [13, 14], m-polar fuzzy [15, 16], and knowledge granularity [17].

Among the abovementioned approaches, DMA is a typical reduction model. Since DMA consumes a lot of space to store distinguishable information, it cannot reduce large data sets. In order to effectively express the distinguishable information among samples, Hu and Cercone [18] proposed a concise definition of a discernibility matrix. Ye and Chen [19] proposed a discernibility matrix-elements that retains all bases of 1. Yang and Sun [20] use the sample comparison of the upper and lower approximation to obtain the discernibility matrix. Dong et al. [21] proposed a fast algorithm of attribute reduction for covering the decision system with minimal elements in discernibility matrix. Wei et al. [22] proposed two discernibility matrices in the sense of entropies. However, these approaches only consider how to improve the distinguishing ability of samples, and do not consider the space consumption. In order to reduce the space computation, Jiang [10] proposed a minimal element selection tree. Li et al. [23] proposed a simple object-attribute discernibility matrix approach. Although scholars have improved the discernibility matrix, the space consumption problem has not been fundamentally solved. To overcome this deficiency, this paper proposes a method based on IEP without discernibility matrix. Firstly, we divide the data set according to conditional attributes and decision attributes and calculate the number of indistinguishable element pairs. Then, select the conditional attribute with the smallest values of IEP. Finally, repeat the abovementioned two steps until the value is 0.

With the rapid development of communication and network techniques, the actual data may change over time. However, the IEP method is only suitable for static data sets. Hence, it is desired to design an incremental attribute reduction algorithm with IEP to deal with dynamic decision systems.

Incremental learning is an efficient approach making full use of the precious results of the original decision system, which can obtain the efficient reduced results by recomputing the updated part of the dynamic data set. Many incremental algorithms have been proposed with different models for dynamic data. Yang proposed an incremental algorithm for updating an object or attribute [24]. Ge et al. developed an incremental attribute reduction based on a simplified discernibility matrix, which is equivalent to attribute reduction based on a positive region [25]. Liu et al. proposed a strong discernibility matrix method for incremental attribute reduction on fuzzy decision tables [26]. In literature [27], Wei proposed three new types of discernibility matrices by compacting a decision table. Zhang et al. proposed a method based on a relation matrix under the change attribute reduction in set-valued information systems [28]. Ma et al. [29] proposed a compressed binary discernibility matrix to process the group dynamic data. Obviously, the abovementioned matrix methods mainly focus on updating the elements of discernibility matrix. These approaches are ineffective in obtaining the reduction results with large-scale decision systems due to the limited memory space. Hence, we incorporate the incremental update mechanism into the IEP approach. Verifies the feasibility and efficiency of proposed algorithm through extensive experiments on UCI data sets.

2. Preliminaries

In this section, we review some basic concepts about rough set, discernibility matrix, and indistinguishable element-pair.

2.1. Basic Concepts

Definition 1. (see [1]). Given the decision system is a quadruple tuple , where is a finite nonempty object set and is a finite nonempty attribute set, is a set of its values, and is an information function with for each and . If , where is the conditional attribute set, and is the decision attribute set. For every subset , an indiscernibility relation is defined as follows: Obviously, if denotes as , is an equivalence relation. We assume includes x, the equivalence relation x is defined as:

Definition 2. (see [1]). Given the decision system for every subset and indiscernibility relation , the upper approximation set and the lower approximation set of Y can be defined by the basic set of P as follows:The universe is partitioned into three disjoint regions by these two approximations and : the positive region , the negative region , and the boundary region . Then the three different regions are defined as following, respectively:

Definition 3. (see [18]). Let be a decision table, C be the condition attribute set, and D be the decision attribute. The discernibility matrix in all samples is defined as , where:

Definition 4. (see [31]). Let be a decision table, C be the condition attribute set, and D be the decision attribute. In terms of a positive region, the discernibility matrix is defined as ,where: is the consistent part of the decision table and is the inconsistent part .

Definition 5. (see [22]). Let be a decision table, C be the condition attribute set, and D be the decision attribute. The discernibility matrix in the sense of complement entropy is defined as ,where: and are the same as Definition 4.

Theorem 1. In the discernibility matrix information in Definition 5, the total number of pairwise comparisons between elements is only related to the division of decision attributes of the data objects.

Proof. Suppose is a nonempty finite set of data objects. is the consistent part of which belongs to the set of positive regions. is the inconsistent part of , which is included in the negative region set. The samples in and are not duplicated. Let , then , and . Suppose is the cardinality of the data set, we can have . According to Definition 5, it can be seen that there are three cases for comparison between samples.
Firstly, the positive region samples with different values of decision attribute should be compared in pairs where and . Due to the symmetry of and , the sample and are compared twice, the actual number of comparisons should be halved. Let count1 be the number of comparisons between samples of the positive region set, then we can achieve the following conclusion:Secondly, let count2 be the number of comparisons between samples of the positive region and negative region, then we have where .
Thirdly, all samples among the negative region are compared with other samples, and repeated comparisons should be subtracted, we have .
Overall, the total number () of comparisons in discernibility matrix based on Definition 5 is as follows:Obviously, when a data set is given, the total number of comparisons is only related to the division of decision attributes, and irrelevant to conditional attributes.

2.2. The Presentation of the Indistinguishable Element-Pair

The discernibility matrix algorithm records the differences between samples by different values of conditional attributes between them, and the amount of distinguishable information measures the importance of the attributes. The larger the value, the more important the attribute. For the discernibility matrix proposed in literature [24] and literature [22], the number of comparisons between samples in the discernibility matrix is determined when the data set is given. Among the samples to be compared, the value of certain condition attributes is either the same or different. The same values of one conditional attribute mean being indistinguishable, while different means being distinguishable. If the amount of distinguishable information is larger, the amount of indistinguishable information is smaller when the total number of comparisons does not change. Here, we use the amount of indistinguishable information to measure the importance of conditional attributes.

Definition 6. Suppose is a universe that is nonempty finite data set, is a conditional attribute set. is the division of data set U on conditional attribute A and is the division on decision attribute D. is the division of inconsistent samples in decision attributes The indistinguishable element-pair of A relative to D is defined as follows:where .
In fact, all data objects among the subdivision are indistinguishable from each other. There are pairs. However, among these element pairs, some comparisons should be subtracted due to some data objects with the same decision attribute value. We have the definition of indistinguishable element-pair.
The asterisked (∗) data objects belong to the negative region set

Example 1. Suppose U is a simplified decision table without repeated samples in Table 2. A is a conditional attribute and D is a decision attribute. Let , the data objects belong to the negative region, Let and In subdivisions , the underlined data objects have the same decision attribute value. Data set has three subdivisions and on conditional attribute A. According to Definition 6, we have indistinguishable information of the three subdivisions as follows:

Theorem 2. The smaller the indistinguishable element pair, the stronger the distinguishing ability.

Proof. According to Theorem 1, if the data set is given, the number of data objects in the positive and negative regions is also definite. Obviously, according to Definition 5, the number of comparisons between all data objects in the data set is also an invariable number. Suppose TotalCount is the total number of comparisons and the indistinguishable element-pair based on conditional attribute A is , then the discernibility element-pair’s TotalCount-. So, the smaller the , the bigger the TotalCount-, the stronger the discernibility.

Theorem 3. Given the decision system and . Then is detailed of , and we have

Proof. Let and . Suppose is a subdivision of . The detail subdivision based on P is and and , the other subdivision is unchanged. Let . For convenience below, let and , we have the result as follows:Since ,

Theorem 4. Let be a decision table and is a data set without duplicate samples, then .

Proof. Assume , Since is a data set without duplicate samples, , So .

Definition 7. Let be a decision table and . U is a data set without duplicate samples. Then B is a relative reduction based on the following indistinguishable element-pair of S if B satisfies:(1);(2);

Definition 8. (see [30]). Let be an information system. For any the value of object about attribute a is . Let and , we called Dis(a) and Dis(A) the discernibility relations in terms of a and A, respectively.

Definition 9. (see [30]). Suppose , let denote a matrix, where for , otherwise , is called the discernibility matrix of the information system .

3. The Algorithm Based on Indistinguishable Element-Pair

The indistinguishable element-pair algorithm obtains the importance of attributes by means of discernibility matrix information and does not create the discernibility matrix. We should compute the positive region and negative region at first, then achieve the simplified decision table. Reduce the calculation of duplicate data objects, saving a lot of time.

3.1. Compute the Positive and Negative Region (CPNR)

In each detailed subdivision, if the decision value of the sample is different, then we put the first sample x into the negative region set and let and the rest put the x into the positive region set and let . In the process of calculating positive and negative regions, equivalence class division needs to be calculated continuously. Here is an ingenious method, which can greatly speed up the calculation speed of equivalence class partitioning. The details are described as follows: for i in range (n): list [array[i] AppendixAppendix for clarity. (i).

If the data object has an integer value on the attribute, the characteristics of an integer can be used. By collecting all the objects with the same value in the same subdivision, equivalence class division can be obtained quickly and accurately.

Example 2. There are six data objects, the values of attribute A are 1, 2, 1, 3, 2, and 1, respectively. Let Array [1–6] = {1, 2, 1, 3, 2, 1}, collecting the data objects with the same value into the same list. Array [1] = Array [3] = Array [6] = 1, we have list [1] = {1, 3, 6}. Array [2] = Array [5] = 2, we have list [2] = {2, 5}. Array [4] = 3, we have list [3] = {4}.
Abovementioned all, data objects 1, 3, and 6 are divided into the same subdivision, and data objects 2 and 5 are divided into the same subdivision. (Algorithm 1)

Input:
Output:
 / is a data set without duplicate samples, is a positive region, data set and is a negative region data set. /
Step 1:
Step 2:
Step 3: while () do{
 Step 3.1: for any , let
 Step 3.2: compute ,
 Step 3.3: statistics of the subdivisions regarded as
, let , count add 1. }
Step 4: scan the remaining subdivisions.
3.2. The Attribute Reduction Algorithm Based on Indistinguishable Element-Pair (IEP)

Suppose f (x, a) is the value of the data object x on the conditional attribute A. There have two different data objects xi and xj, if , then a is recorded in the discernibility matrix. Another way to think about it is to take down indistinguishable data objects. After research, all samples divided by the same subdivision are indistinguishable. Algorithm IEP is described as follows: (Algorithm 2)

Input:
Output: red
Step 1:
Step 2: Calculate the positive region and negative region with CPNR
Stet 3: get the simplified decision table through Step 2
Step 4: ,
Step 5: while () do {
 Step5.1: IEP(b, B, D) = min{IEP(a, B, D),
 Step 5.2:
 Step 5.3: delete all the subdivisions and the card is 1 of
 Step 5.4: }
Step 6: return red

Example 3. Suppose is a simplified decision table without repeated samples in Table 3. Based on the definition equivalence class, we have  = {{1, 4},{2, 3},{5, 6}}.The bold data objects of the subdivision have the same value on the attribute . The asterisked () data objects belong to the negative region set. Based on the Definition 6, we have , {{1, 2, 3, 4},{5, 6}},  = {{1, 3, 6},{2, 5},{4}},  = 3,  = {{1, 3, 5, 6},{2, 4}}
When the conditional attribute with the lower indistinguishable degree has a stronger distinguishing ability, we select the attribute a, and let red = {a}. If the amount of information is not zero, we enter the next cycle. On the basis of it is divided by conditional attributes {b, c, e}, we obtain the following:  = {{1, 4},{2, 3},{5, 6}} ,  = {{1}, {2}, {3}, {4}, {5}, {6}}, ,  = {{1}, {2}, {3}, {4}, {5, 6}} , .
Because is the smallest, we select c to merge into the reduced set red. Now, the amount of information is 0, and the algorithm terminates. Reduce result is .

3.3. The Existing Static Reduction Algorithms

The typical discernibility matrix algorithm and the related improved algorithms constantly revise the definition of discriminant matrix from the perspective of distinguishable data objects, leading to the inevitable consumption of a large amount of space resources to store the discernibility matrix. The phenomenon of memory overflow often occurs during the reduction of large data sets, which leads to the failure to complete the reduction task.

The IEP method does not need to store discernibility matrix and is suitable for reduction of large-scale data sets. In order to further verify the effectiveness of the IEP presented in this paper, let’s analyze the complexity of time and space and other similar algorithms based on discernibility matrix. In IEP algorithm, U is a decision table. Steps 2-3 focus on calculating the simplified decision table and . The time complexity of computing is . The time complexity of steps 5 is . The space complexity of data set is , steps 5 want space . Therefore, the total time complexity of algorithm IEP is and the space complexity is . Table 3 shows a comparison of the time and space complexity of computing the reductions by Algorithms HU [18], DDMSE [22], and MEDA [30]. From Table 4, we can obtain the time complexity of IEP is usually much smaller than the algorithms HU because of is lower than . The time complexity of the algorithm MEDA is . The space complexity of algorithm IEP is but the space complexity of storing the discernibility matrix of HU, DDMSE is . Therefore, the space consumption of algorithm IEP is much less than that of algorithms HU, DDMSE, and MEDA.

4. Incremental Attribute Reduction Algorithm Based on Indistinguishable Element-Pair

The abovementioned algorithm IEP only adapts to the static data set. In reality, most data sets are dynamic. The traditional static methods are ineffective. Therefore, it is necessary to study some algorithms for dynamic data sets.

4.1. An Incremental Method to Calculate Indistinguishable Element-Pair after Adding Some Objects (IEPAO)

There are two kinds of data objects as to updating in data set: increase and decrease. Let’s introduce the first one: to increase the data objects. When some data objects are added to data set, we only need to calculate the IEP of the updating part and obtain the amount of information with the help of the previous reduction result red. If the amount of information is zero, the updated reduction result is red. Otherwise, the added part objects will be merged with the basic data, we compute the amount of information based on detailed subdivision according to attribute set .

Theorem 5. Let be a decision system, . It is assumed that is the new data objects, . According to the division of , we have . Then

Proof. See Appendix 1 for the proof process.
According to Theorem 5, the value of is related to , where add data objects. We propose the algorithm IEPAO based on this characteristic. (Algorithm 3)
In Table 5, is the reduction result before adding data, is the final reduction result IEP can only reduce static data set the time complexity is If data objects is added, the time complexity becomes . The IEPAO algorithm uses the previous reduction results, the time complexity is . It clearly shows that the calculation time of IEPAO is less than IEP.

Input: , and incremental object sets , where is the simplified decision table before update.
Output: Updated reduction set
Step 1: Mark the negative region data objects, ,
Step 2: Compute
Step 3: We may assume that has k subdivisions and , then compute .
Step 4: Compute
Step 5: If Info = 0, algorithm is terminated else
 While do {
  Let
   for a in {
     Compute
   ,
   ,
    }
 Compute }
Step 6: return
4.2. Incremental Updating Attribute Reduction Algorithm When Delete Some Objects (IEPDO)

In reality, some data will be discarded after a long time. IEPDO algorithm can reduce the deleted data objects dynamically.

Theorem 6. Let S =  be a decision system and . We assume that the deleted data object set is and . From the definition of equivalence class divided, we have , where (i = 1, 2, …, k). If delete the data objects from U, the indistinguishable amount of information is , then .

Proof. According to Theorem 6, the value of is related to where delete data objects. We propose the algorithm IEPDO based on this characteristic. (Algorithm 4)
When some objects are deleted, we have the indistinguishable information of updated data set, which needs a small amount of computation through the previous reduced result and deleted data objects. Suppose the final reduction result is , usually, less than or equal to and less than . The time complexity comparison of IEPDO and IEP is provided in Table 6. Obviously, the time complexity of IEPDO is smaller than IEP.

Input: , and delete object sets , where is the simplified decision table before update.
Output:
Step 1: Mark the deleted data objects is , ,
Step 2: Compute ;
Step 3: We may assume that has k subdivisions and , then compute
Step 4: Compute
Step 5: Let
 for a in {
  Compute ,
  if Info = 0 then and
 else
  break;
  }
Step 6: return

5. Experiment Analysis

In this section, lots of experiments are conducted on both static and dynamic data sets to verify the efficiency of the proposed attribute reduction algorithms. In the experiments, fifteen data sets are downloaded from UCI. Table 7 displays the basic information of each data set, where represents the number of samples, represents the number of conditional attributes, represents the number of decision classes, and Type represents the decision system is consistency (Y in short) or inconsistency (N in short), respectively. For the convenience of the following description, the data set Letters recognition is abbreviated as Letters, Mammographic Mass as Mass. All the character or string features are normalized into an integer. All of the experiments have been implemented on a PC with Windows 10, Core™ i7-10710U CPU 1.10 GHz 1.61 Hz and 8 G memory. All of the algorithms are coded in python, and the used software isPyCharm Community Edition 2020.2.3 × 64 and Weak3.2.

5.1. Performance Comparison between Algorithm IEP and Other Discernibility Matrix Algorithms Based on Static Data Sets

In experiment, we consider the fifteen data sets from UCI listed in Table 7. These selected data sets are reasonably distributed, including large data sets for Letters and small data sets for Hepatitis, Audiology, consistent data sets (Gene and Mushroom, etc.), and inconsistent data sets (Mass and Spect heart). In order to show the time effect of each algorithm, we refer to the SpeedupRatio= Tbaseline/T method proposed by literature [31], is the executing time of a typical algorithm. Tbaseline reaches its maximum when the typical algorithm cannot perform the reduction task. Then .

For the different data sets, the SpeedupRatio of IEP and the other three algorithms (Hu, DDMSE, and MEDA) is also different. Table 8 shows the SpeedupRatio of IEP, Hu, DDMSE, and MEDA. The data in bold indicates that the algorithm runs the fastest on a certain data set. In Table 8, we have the unpredictable speed of Hu, DDMSE, and MEDA on Letters and Connect-4 data sets because these two data sets are too large. The SpeedupRatios of IEP are 1.6109 and 2.4286, respectively, on data sets Audiology and Hepatitis. Since the values are greater than 1, the speed of IEP is faster than Hu on Audiology and Hepatitis. From Table 8, IEP is the fastest but the difference is not obvious among the Hu, DDMSE and MEDA based on small data sets. Overall, the is related to . The smaller the , the faster the speed.

Table 9 shows the reduction and performance time of the comparisons of four algorithms. In Table 9, time is measured in seconds, red is reduction and the data in bold represents the minimum reduction time of many algorithms. The IEP takes only 1.953 seconds to reduce the Mushroom data set, while DDMSE takes 230.071 seconds. The main reason is that DDMSE modifies the definition of discernibility information to improve the distinguishing ability, leading to the increasing number of compared element pairs. Then, it makes the space for storing discernibility matrix larger and larger. On data sets Letters and Connect-4, IEP can quickly and effectively obtain the reduction results, while Hu, DDMSE, and MEDA cannot complete the reduction task due to insufficient memory. On the small data set Breast and Balance-scale, IEP needs to waste 0.035 seconds, while the other three algorithms take 0.191,0.152 and 0.044 seconds respectively. For the reduction on a small data set, the time effect of IEP and other algorithms is not obvious.

Compared with other discernibility matrix algorithms, IEP has less time consumption, while getting the same reduction results based on the same data sets. Especially, the reduction effect is more obvious on large-scale data sets.

5.2. Time Comparison of IEPAO and IEP When Adding Data Objects

In the following experiments, we select nine data sets for dynamic update experiments from Table 7. For each data set, 50% of the data objects are randomly selected as the original object set, and the remaining data are randomly generated at the proportions of 10%, 20%, 30%, 40%, and 50% as incremental object sets, respectively. The incremental part is divided into 5 groups of experiments, each group is executed 10 times and computes the average time. Experimental results are outlined in Figure 1. In the experiment of IEP, time statistics do not include the calculating time of original objects.

In each subfigures of Figure 1, the x-coordinate represents the increment ratio, and the unit is the proportion of the increased part of the data in the total data. The value of the y-coordinate y is the time of computing reduction in different incremental, which are measured by seconds. In Figure 1, the curve with a five-pointed star mark shows the change in the running time of IEPAO, while the curve with a circle mark indicates the variation of IEP.

It can be seen in Figure 1 that as the size of data set expands, the time of calculating reduction will increase. The calculation time of IEPAO is much less than IEP. The main reason for this phenomenon is that when we add the new objects into the data set, IEPAO only needs to calculate the added part of the data, and then combine the previous reduction results in obtaining the changed result quickly. But, IEP can only process static data sets. When new data is added, it takes longer time to recalculate the original data and the added part. On the whole, the performance of IEPAO is relatively stable in Figure 1. With the increase of updated data, the calculation time is also increasing. But, IEPAO has an anomaly in that the calculation time decreased as the data increasing. The subfigure (f) of Figure 1 displays that IEPAO takes 4.543 seconds to reduce the Letters data set with a 20% increment, and 4.253 seconds to reduce the data with 30% increment. The main reason is that when the data increases, it accelerates the division of data on conditional attributes. Since the amount of information is zero based on subdivision of cardinality 1, these objects are constantly deleted during the reduction process to accelerate the convergence speed.

5.3. Time Comparison of IEPDO and IEP When Deleting Data Objects

The same as Section 5.2, nine data sets are selected in Table 7. Take the original data of each data sets as the basic data, and randomly select 10%, 20%, 30%, 40%, and 50% objects in the remaining data to delete respectively. Use IEP and IEPDO to reduce the updated data. Each group of data was selected to repeat the experiment for 10 times and averaged the time of 10 times. The experimental results show in Figure 2. In each subfigure of Figure 2, x-axis represents the proportion of decrement data objects, while y-axis represents the computational time. The time is measured by seconds. The curve with five-pointed star mark in Figure 2 shows the change in the running time of the IEPDO, while

As shown in Figure 2, when we delete some data objects from data set, the computational time of IEP and IEPDO will decrease accordingly. In the same computing environment, IEPDO takes less time than IEP. The IEPDO method only needs to calculate the updated objects when deleting data objects. Some updated objects when deleting data objects. Some conditional attributes with zero indistinguishable element-pair are removed from the reduction set. However, IEP takes longer time because of calculating the reduced data set from all conditional attributes. For the IEP selected, it takes 0.0338 seconds on the decreased 30% data objects of Hepatitis, while taking 0.0352 seconds to compute the deleted 40% data. It costs 1.233 seconds to calculate the decrease of 30% data of Nursery but takes 1.2451 seconds on the decrease of 40% data. Why the calculation time increases with the decreasing data is that the speed of dividing equivalence class is slowed down as it randomly selects some weak distinguish power on conditional attributes. The performance of IEPDO is relatively stable. By deleting data, the calculation time decreases. From nine subfigure of Figure 2, when we reduce the large-scale data, the time-consuming effect of IEPDO is more obvious with the decreased data objects, while the effect is not significant on small data sets, such as data sets of Chess kr-kp and Hepatitis.

5.4. Classification Accuracy Analysis of IEP, IEPAO, and IEPDO

In this section, the precision of classification is calculated on the selection of reducts obtained by the algorithms IEP, IEPAO, and IEPDO. Firstly, we take 50% of objects of each data set in nine data sets from Table 7 as the basic data set, the rest 50% data as the incremental objects and select the algorithms IEP and IEPAO to reduce. Secondly, we delete 50% of objects randomly from each data set, using the algorithms IEP and IEPDO to process. Then, the classification accuracies are acquired by using J48, NaiveBayes (NB), RandomForest (RF), SMO classifier, and 10-fold cross-validation. The experimental results are shown in Tables 10 and 11.

From Table 10, it is clear that when some objects are added into the information systems, the average classification accuracy of the reduction found by incremental algorithm IEPAO is better than those of algorithm IEP in data sets Chess, Connect-4, Gene, Handwritten, Hepatitis, Letters and Ticdata2000 are coincide with those of algorithm IEP in data sets, e.g., Chess kr-kp, Mushroom, and Nursery. The experimental results show that the incremental algorithm IEPAO can find a feasible attribute reduction when incremental algorithm IEPAO replaces algorithm IEP. Moreover, the algorithm IEPAO can obtain high-quality attribute reduction with less time consumption. Similarly, when some objects are deleted from the original object set, the average classification accuracy of the reduction obtained by the algorithm IEPDO is better than IEP in data sets Connect-4, Gene, Handwritten, Hepatitis, Letters, and Ticdata2000.

Accordingly, we can conclude that the incremental algorithm IEPDO can find a feasible attribute reduction.

Hence, the experimental results verified that the proposed incremental methods IEPAO and IEPDO can obtain an efficient attribute reduction and provide a quick data preprocessing method for dynamic data sets.

6. Conclusions and Further Study

Attribute reduction can effectively eliminate redundant information. Though the discernibility matrix method is one of the intuitive and effective reduction methods, it cannot deal with the reduction of large-scale data sets effectively because of memory overflow. The attribute reduction mechanism based on IEP can effectively solve the problem of space consumption analyzed in this paper. During the reduction process, IEP effectively prunes the subdivisions with cardinality 1, which speeds up the calculation of equivalence class division. IEP has better time and space effects in reduction, but it only adapts to the environment of static data sets. Considering the constant updating of data, in reality, IEPAO and IEPDO are proposed on the basis of IEP to deal with the reduction of adding data objects and deleting data objects respectively. As to IEP, the entire data set has to be reduced again and consumed a lot of time with the data changes. IEPAO and IEPDO only compute the changed part data and combine the previous reduction results, which can obtain the data set with fewer redundancies and better outcomes.

Of course, the algorithm proposed has some shortcomings in this paper. For example, (1) The IEP method can only reduce integer or character data, but cannot adapt to process other types of data. (2) The incremental update algorithm proposed in this paper does not consider the changes in attributes and values.

In the future, we will conduct the research from the following aspects: design an increment algorithm adapting to different types of data; develop a reduction method regarding the change values of data objects; propose an incremental mechanism with adding and deleting some attributes. Additionally, those approaches should adapt to an incomplete decision system.

Appendix

The proof of Theorem 5

Proof. From Definition 8, we have

Data Availability

All the data included in this study are available upon request by contact with the corresponding author.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was partially supported by the Natural Science Foundation of China (61836016), the Quality Improvement Project of Chaohu University on Discipline Construction (kj21gczx03), Special Support Plan for Innovation and Entrepreneurship Leaders in Anhui Province, the Provincial Natural Science Research Program of Higher Education Institutions of Anhui Province (KJ2021A1030), and the Key Subject Subprojects of Chaohu University ZDXK-201815.