Abstract

Multigranulation rough set is an extension of classical rough set, and optimistic multigranulation and pessimistic multigranulation are two special cases of it. β multigranulation rough set is a more generalized multigranulation rough set. In this paper, we first introduce fuzzy rough theory into β multigranulation rough set to construct a β multigranulation fuzzy rough set, which can be used to deal with continuous data; then some properties are discussed. Reduction is an important issue of multigranulation rough set, and an algorithm of granular space reduction to β multigranulation fuzzy rough set for preserving positive region is proposed. To test the algorithm, experiments are taken on five UCI data sets with different values of β. The results show the effectiveness of the proposed algorithm.

1. Introduction

Qian et al. [13] proposed a multigranulation rough set, which is constructed on a family of granular structures and is different from Pawlak’s rough set [47]. Qian’s multigranulation rough set can be used to approximate an unknown concept through a family of binary relations; each binary relation can generate a granulation space, which may be partition [1], covering [8, 9] or even neighborhood system [1012] on the universe of discourse.

Qian’s rough set includes two basic models, one is optimistic multigranulation rough set and the other is pessimistic multigranulation rough set. The word “optimistic” means that at least one of the granulation spaces can be used for approximating while the word “pessimistic” means that all of the granulation spaces should be used for approximating. In these two models, all of the binary relations, or granulation spaces, are presented simultaneously; therefore, optimistic and pessimistic are two special cases of multigranulation rough set. To get a more suitable model for practical application, Xu et al. [13] proposed a more generalized multigranulation rough set, called β multigranulation rough set that designed by a threshold β for controlling the number of the equivalence classes, which are contained in the target.

In recent years, the multigranulation approach has attracted many researchers’ attention [1418]. Xu et al. generalized multigranulation fuzzy rough sets to tolerance approximation space to construct optimistic and pessimistic multigranulation fuzzy rough sets models [14]. Qian et al. further generalized their optimistic multigranulation rough set into incomplete information system [15]. In [16] Yang et al. introduced fuzzy theory into multigranulation rough set, which employed the T-similarity relations (reflexive, symmetric, and T-transitive) to construct the multigranulation fuzzy rough sets. Therefore, how to generalize multigranulation rough set is an important research field, as we all know how to introduce the fuzzy case to rough set model plays an important role in the development of rough set theory; fuzzy rough set [19] has attracted increasing attention from the domains of machine learning and intelligence data analysis. So, it is not difficult to introduce fuzzy rough set theory into β multigranulation rough set.

Granular space reduction is an important issue of multigranulation rough set and it is recently researched by many scholars [2024]. In this paper, we focus on the problem to deal with β multigranulation rough set. Hu et al. in [25] proposed a fuzzy-rough attribute reduction. Motivated by this idea, we will introduce fuzzy rough set into β multigranulation rough set to construct β multigranulation fuzzy rough set model and design an algorithm of granular space reduction to β multigranulation fuzzy rough set. The algorithm can be used to granular space reduction in multigranular structures for preserving positive region of β multigranulation fuzzy rough set and will be very useful in big continuous data.

The purpose of this paper is to further generalize β multigranulation rough set to fuzzy environment. To facilitate our discussion, we first present some basic knowledge of rough set in Section 2. In Section 3, β multigranulation fuzzy rough set will be constructed and the properties will be discussed. In Section 4, an algorithm of granular space reduction to β multigranulation fuzzy rough set will be proposed and experiments are taken on five UCI data sets. In Section 5, conclusion is made.

2. Preliminaries

2.1. Rough Sets

Formally, a decision system is an information system , in which is a nonempty finite set of objects called the universe of discourse and is a nonempty finite set of the condition attributes; is the set of the decision attributes and .

For all , let us denote by the value that holds on   . For an information system , one then can describe the relationship between objects through their attributes’ values. With respect to a subset of attributes such that , an indiscernibility relation may be defined as

The relation is reflexive, symmetric, and transitive; then is an equivalence relation.

Definition 1. Let be a knowledge base in which , for all , the lower and upper approximations of in terms of the equivalence relation are denoted by and , respectively: where is the equivalence class based on indiscernibility relation and is denoted as .
is referred to as Pawlak’s rough set.

2.2. Multigranulation Rough Set

Multigranulation rough set is different from Pawlak’s rough set. The former is constructed on a family of the equivalence relations, and the latter is constructed on an equivalence relation. In Qian et al.’s multigranulation rough set theory, two basic models were defined. The first one is the optimistic multigranulation rough set, and the second one is the pessimistic multigranulation rough set.

Definition 2. Let be an information system in which ;  for all , the optimistic multigranulation lower and upper approximations are denoted by and , respectively: where is the complementary set of .

Definition 3. Let be an information system in which ; for all , the pessimistic multigranulation lower and upper approximations are denoted by and , respectively: where is the complementary set of .

2.3. β Multigranulation Rough Set

Optimistic and pessimistic are two special cases of multigranulation rough set. Optimistic case is loose since if only one equivalence class of an object is contained in the target, then such object is included into lower approximation; pessimistic is strict since if all the equivalence classes of an object are contained in the target, then such object is included into lower approximation. To solve this problem, Xu et al. [13] proposed a more generalized multigranulation rough set, called β multigranulation rough set that are designed by a threshold β for controlling the number of the equivalence classes, which are contained in the target.

Definition 4. Let be a multigranulation decision system; for all and , the characteristic function is defined as where .

Definition 5. Let be a multigranulation decision system; for all , the β  multigranulation lower and upper approximations of are denoted by where . is the complementary set of .
is referred to as β  multigranulation rough set of .

2.4. Fuzzy Rough Set

Fuzzy rough set is a generalization of rough set. It can be used for decision information system to deal with continuous types of conditional attributes. Usually a fuzzy similarity relation is computed by conditional attributes and is employed to measure similarity between two objects, which then develop upper and lower approximations of fuzzy sets. Fuzzy rough set generalize the objects discussed in rough set to fuzzy set and turn the equivalence relation to fuzzy equivalence relation.

Definition 6. Let be a universe of discourse and a fuzzy similarity relation of ; for all , the fuzzy lower and upper approximations of are denoted by
In fuzzy rough set, measurement should be introduced to construct fuzzy similarity relation, such as the max-min method. Then fuzzy similarity matrix can be constructed by fuzzy similarity relation; after that, fuzzy equivalent matrix can be constructed in terms of fuzzy similarity matrix by transitive closure method: where is the relation value of and .
is a fuzzy equivalence relation if satisfies reflectivity, symmetry, and transitivity.

2.5. Multigranulation Fuzzy Rough Set

In [15], Qian et al. introduced the theory of fuzzy set into multigranulation rough set to construct the optimistic and pessimistic multigranulation fuzzy rough sets.

Definition 7. Let be a fuzzy decision information system, are   fuzzy subsets, and decision attribute; for all   , the optimistic multi-granulation fuzzy lower and upper approximations of   are denoted by where is the fuzzy equivalent class of . is the complementary set of .
is optimistic multigranulation fuzzy rough set.

Definition 8. Let be a fuzzy decision information system, are fuzzy subsets, and decision attribute; for all , the pessimistic multigranulation fuzzy lower and upper approximations of are denoted by where is pessimistic multigranulation fuzzy rough set. is the complementary set of .

3. β Multigranulation Fuzzy Rough Sets

In β multigranulation rough sets, by setting different values of β, we can get different reductions from which the most suitable reduction can be used in next research. The fuzzy rough set is very suitable for big continuous data set. So it is natural to introduce the theory of fuzzy set into β multigranulation rough sets to construct β multigranulation fuzzy rough sets model.

In this section, we will give some definitions of β multigranulation fuzzy rough sets model and discuss some properties of it.

Definition 9. Let be a fuzzy decision information system, are fuzzy subsets, and is decision attribute; for all and , the characteristic function is defined as where is the fuzzy equivalent class of . Then β multigranulation fuzzy rough set is defined as follows.

Definition 10. Let be a fuzzy decision information system, are fuzzy subsets, and decision attribute; for all , the  β multigranulation fuzzy lower and upper approximations of are denoted by where is β multigranulation fuzzy rough set. is the complementary set of .
Following Definition 10, we will employ the following denotations:positive region of : ;negative region of : ;boundary region of : .

Theorem 11. Let S be a fuzzy decision information system, are m fuzzy subsets, and decision attribute; for all ,

Proof. We only prove (15); others can be proven analogously.
For all , by (13), there exist ,  ; there must be such that , from which we can conclude that ,  .
For all , by (9), there exists such that . Therefore, by and , we can get such that  .
Theorem 11 shows that if , β multigranulation fuzzy rough set turns to optimistic multigranulation fuzzy rough set. If , β multigranulation fuzzy rough set turns to pessimistic multigranulation fuzzy rough set. Obviously, β multigranulation fuzzy rough set is an extension of optimistic multigranulation fuzzy rough set and pessimistic multigranulation fuzzy rough set.

Theorem 12. Let S be a fuzzy decision information system, are fuzzy subsets, and D decision attribute; for all , , we can get

Proof. For all , by (17), there exists ; by (13), there must be , such that .
For all , by (13), there exists ; there must be such that ; then ; by (15), there exists such that .
So .
Formula (20) can be proven analogously.

4. Reduction of β Multigranulation Fuzzy Rough Sets

In single granular fuzzy rough set, reduction is a minimal subset of the attributes, which is independent and has the same discernibility power as all of the attributes. The method of preserving the positive region is usually used for attribute reduction. In this paper, we consider each attribute as a granular space. It is natural to introduce this method into β multigranulation fuzzy rough set for granular space reduction.

Definition 13. Let be a fuzzy decision information system, are fuzzy subsets, is decision attribute, is the partition induced by a set of decision attributes , and approximation qualities of in terms of β multigranulation fuzzy rough set are defined as where, is the cardinal number of set .

4.1. Significance of Granulation

Definition 14. Let be a fuzzy decision information system, are fuzzy subsets, is decision attribute, and is a reduction if and only if(1);(2)for all .
By Definition 10, we can get a reduction of when preserving the approximation quality.

Definition 15. Let be a fuzzy decision information system, are fuzzy subsets, is decision attribute, and ;   for all , the significance of granulation of in terms of is defined as where represents the changes of the approximation quality if a set of attributes is eliminated from . Also, we can define for all ,   represents the changes of the approximation quality if a set of attributes is put in . These two significances can be used to forward granular structure selection algorithm, and can determine the significance of every granulation in terms of the approximation quality.

4.2. Granular Space Reduction Algorithm

See Algorithm 1.

Input: a fuzzy decision information system
    =
Output: a granular space reduction RED
Step  1. , compute fuzzy similarity matrix
Step  2. , compute fuzzy equivalence matrix in terms of the result of Step  1,
then get fuzzy equivalence class
Step  3.
Step  4. , compute
Step  5.
Step  6. , compute
  If
    
  Until
  End
Step  7.  return RED

4.3. Experiment

To demonstrate the above approach, we use 5 data sets gotten from UCI Repository of Machine Learning databases; the description of the selected data sets is listed in Table 1.

In this experiment, each feature is used to construct a granular structure. All features then correspond to multiple granular structures. For each data set, 5 different β are used; then the different results of granular selection of Table 1 under different values of β are listed in Table 2. In Table 2, “” means the number of features, “0.001” is the smallest value of β, “1” is the biggest value of β, and “” is bigger than 0.01 and smaller than 1.

Through Table 2, we get an interesting outcome that the result of granular selection of each data set is changed with different values of β. The granular selection results are increased with the increased value of β. Take for instance that when β = 0.001, we get the least number of features; when β = 1, we get the most number of features. Obviously, when β =, it actually represents the optimistic multigranulation fuzzy rough set; when β = 1, it actually represents pessimistic multigranulation fuzzy rough set, which is too strict for only when all of the granulation spaces satisfy the inclusion condition between the equivalence classes and the target that the object belongs to the lower approximation. in this experiment, there are no reductions can be gotten, which shows that it is difficult to get a satisfied reduction result under pessimistic condition.

In order to test the performance of the proposed algorithm and to get a proper β, we employ neural network and decision tree as the validation function. The results are listed in Tables 3 and 4.

We can find in Tables 3 and 4 that when  β = 1, the selected features are just the same as the original data sets, but the accuracy is not always the biggest, such as the result of “biodeg” in Table 4; when  β = 0.005, the accuracy is bigger than  β = 1, which shows that when   β is set a suitable value it cannot only reduce redundant granular space but also retain the most useful granular space so as to get better accuracy. The selection of granular space is crucial to the performance of the sequent learning, so the selection should reflex the structure of the data and patterns. Comparing the results of Table 3 with Table 4, we can see that the accuracy of Table 4 is bigger than Table 3; it shows that the accuracy is also related to the performance of classifier.

This experiment shows that the proposed algorithm is more flexible for selecting granular space than optimistic and pessimistic multigranulation fuzzy rough sets, which can use fewer features to get higher accuracy.

5. Conclusions

In this paper, a  β multigranulation fuzzy rough set model is proposed, and a corresponding algorithm is proposed; different from other methods, the proposed algorithm is constructed on multigranular spaces, our experiment shows that the granular selection results are increased with the increased value of  β, and the algorithm cannot only deal with continuous data but also when β is set properly, the reduction will be suitable to be classified to get a good result. The experiment result shows the effectiveness of our algorithm.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported by the Graduate Student Innovation Plan of Jiangsu Province (no. CXZZ12-0728), the National Natural Science Foundation of China (nos. 61203024, 61100116, and 71171100), the Natural Science Fundamental Research Project of Jiangsu Colleges and Universities (no. 12KJB120001), and Key Laboratory of Intelligent Perception and Systems for High-Dimensional Information (Nanjing University of Science and Technology), Ministry of Education (no. 30920130122005).