*Manuscript Click here to view linked References
1 2 3 Multigranulation decision-theoretic rough sets 4 5 6 Yuhua Qiana, Hu Zhangb, Yanli Sangb, Jiye Lianga,∗ 7 a 8 Key Laboratory of Computational Intelligence and Chinese Information Processing of Ministry of Education, Shanxi University, Taiyuan, 030006, Shanxi, China 9 bSchool of Computer and Information Technology, Shanxi University, Taiyuan, 030006, Shanxi, China 10 11 12 13 14 Abstract 15 16 The Bayesian decision-theoretic rough sets propose a framework for studying rough set approximations using 17 probabilistic theory, which can interprete the parameters from existing forms of probabilistic approaches to rough 18 sets. Exploring rough sets in the viewpoint of multigranulation is becoming one of desirable directions in rough 19 set theory, in which lower/upper approximations are approximated by granular structures induced by multiple binary 20 relations. Through combining these two ideas, the objective of this study is to develop a new multigranulation rough 21 set model, called a multigranulation decision-theoretic rough set. Many existing multigranulation rough set models 22 can be derived from the multigranulation decision-theoretic rough set framework. 23 24 Keywords: Decision-theoretic rough sets; Granular computing; Multigranulation; Bayesian decision theory 25 26 27 28 1. Introduction 29 30 Rough set theory,originated by Pawlak [24, 25], has becomea well-established theory for uncertainty management 31 in a wide variety of applications related to pattern recognition, image processing, feature selection, neural computing, 32 conflict analysis, decision support, data mining and knowledge discovery [3, 5, 10, 11, 15, 16, 28, 29, 30, 31, 34, 33 36, 41, 55]. In the past ten years, several extensions of the rough set model have been proposed in terms of various 34 requirements, such as the decision-theoretic rough set model (see [51]), the variable precision rough set (VPRS) model 35 (see [56, 58]), the rough set model based on tolerance relation (see [12, 13, 14]), the Bayesian rough set model (see 36 [37]), the Dominance-based rough set model (see [4]), game-theoretic rough set model (see [6, 7]), the fuzzy rough 37 set model and the rough fuzzy set model (see [2]). 38 Recently, the probabilistic rough sets have been paid close attentions [8, 45, 48, 50, 52]. A special issue on 39 probabilistic rough sets was set up in International Journal of Approximate Reasoning, in which six relative papers 40 were published [48]. Yao presented a new decision making method based on the decision-theoretic rough set, which is 41 constructed by positive region, boundary region and negativeregion, respectively [52]. In the literature [50], the author 42 further emphasized the superiority of three-way decisions in probabilistic rough set models. In fact, the probabilistic 43 rough sets are developed based on the Bayesian decision principle, in which its parameters can be learned from a given 44 decision table. Three-way decisions are most of superiorities of probabilistic rough set models. The decision-theoretic 45 46 rough sets can derive various existing rough set models through setting the thresholds α and β. Since the decision- 47 theoretic rough sets was proposed by Yao [49], it have attracted more and more concerns. Azam and Yao [1] proposed 48 a threshold configuration mechanism for reducing the overall uncertainty of probabilistic regions in the probabilistic 49 rough sets. Jia et al. [9] developed an optimization representation of decision-theoretic rough set model, and gave a 50 heuristic approach and a particle swarm optimization approach for searching an attribute reduction with a minimum 51 cost. Liu et al. [23] combined the logistic regression and the decision-theoretic rough set into a new classification 52 approach, which can effectively reduce the misclassification rate. Yu et al. [53] applied decision-theoretic rough set 53 model for automatically determining the number of clusters with much smaller time cost. 54 55 ∗Corresponding author. Tel./Fax: +86 0351 7018176. 56 Email addresses: [email protected] (Yuhua Qian), [email protected] (Hu Zhang), [email protected] (Yanli Sang), 57 [email protected] (Jiye Liang) 58 59 Preprint submitted to International Journal of Approximate Reasoning March 1, 2013 60 61 62 63 64 65 1 2 3 In the view of granular computing (proposed by Zadeh [54]), in existing rough set models, a general concept 4 described by a set is always characterized via the so-called upper and lower approximations under a single granula- 5 tion, i.e., the concept is depicted by known knowledge induced from a single relation (such as equivalence relation, 6 tolerance relation and reflexive relation) on the universe [17, 18, 51]. Conveniently, this kind of rough set models 7 is called single granulation rough sets, just SGRS. In many circumstances, we often need to describe concurrently a 8 9 target concept through multi binary relations according to a user’s requirements or targets of problem solving. Based 10 on this consideration, Qian et al. [26, 27, 28] introduced multigranulation rough set theory (MGRS) to more widely 11 apply rough set theory in practical applications, in which lower/upper approximations are approximated by granular 12 structures induced by multi binary relations. From the viewpoint of rough set’s applications, the multigranulation 13 rough set theory is very desirable in many real applications, such as multi-source data analysis, knowledge discovery 14 from data with high dimensions and distributive information systems. 15 Since the multigranulation rough set was proposed by Qian in 2006 [26], the theoretical framework have been 16 largely enriched, and many extended multigranulation rough set models and relative properties and applications have 17 also been proposed and studied [27, 28, 29, 30, 31, 32]. Wu and Leung [39] proposed a formal approach to granular 18 computing with multi-scale data measured at different levels of granulations, and studied theory and applications of 19 granular labelled partitions in multi-scale decision information systems. Tripathy et al. [38] developed an incomplete 20 multigranulation rough sets in the context of intuitionistic fuzzy rough sets and gave some important properties of 21 the new rough set model. Raghavan et al. [33] first researched the topological properties of multigranulation rough 22 sets. Based on the idea of multigranulation rough sets, Xu et al. [42, 43, 44] developed a variable multigranulation 23 rough set model, a fuzzy multigranulation rough set model and an ordered multigranulation rough set model. Wu [40] 24 extended classical multigranualtion rough sets to the version based on a fuzzy relation, and proposed a new multigran- 25 ulation fuzzy rough set (MGFRS). Zhang et al. [57] defined a variable precision multigranulation rough set, in which 26 the optimistic multigranulation rough sets and the pessimistic one can be regarded as two extreme cases. Through 27 introducing some membership parameters, this model becomes a multigranulation rough set with dynamic adaption 28 29 according to practical acquirements. Yang et al. [46, 47] examined the fuzzy multigranulation rough set theory, and 30 revealed the hierarchical structure properties of the multigranulation rough sets. Liu and Miao [21, 22] established 31 a multigranulation rough set approach in covering contexts. Liang et al. [19] presented a kind of efficient feature 32 selection algorithms for large scale data with a multigranulation strategy. She et al. [35] explored the topological 33 structures and the properties of multigranulation rough sets. Lin et al. [20] gave a neighborhood multigranulation 34 rough set model for multigranulation rough data analysis in the context of hybrid data. In the murigranulation rough 35 set theory, each of various binary relation determines a corresponding information granulation, which largely impacts 36 the commonality between each of the granulations and the fusion among all granulations. As one of very important 37 rough set models, the decision-theoretic rough sets (DTRS) are still not be researched in the context of multigranula- 38 tion, which limits its further applications in many problems, such as multi-source data analysis, knowledge discovery 39 from data with high dimensions and distributive information systems. 40 In what follows, besides those motivations mentioned in first multigranulation rough set paper (see Cases 1-3 in 41 the literature [29]), we further emphasize the specific interest of multigranulation rough sets, which can be illustranted 42 from the following three aspects. 43 44 • Multigranulation rough set theory is a kind of of information fusion strategies through single granulation rough 45 sets. Optimistic version and pessimistic version are only two simple methods in these information fusion approaches, 46 which are used to easily introduce multigranulation ideas to rough set theory. 47 • In fact, there are some other fusion strategies [20, 39-41, 43]. For instance, in the literature [39], Xu et al. 48 49 introduced a supporting characteristic function and a variable precision parameter β, called an information level, to 50 investigate a target concept under majority granulations. 51 • With regard to some special information systems, such as multi-source information systems, distributive infor- 52 mation systems and groups of intelligent agents, the classical rough sets can not be used to data mining from these 53 information systems, but multigranulation rough sets can. 54 55 In this study, our objective is to develop a new multigranulation rough decision theory through combining the 56 multigranulation idea and the Bayesian decisoin theory, called multigranulation decision-theoretic rough sets (MG- 57 DTRS). We mainly give three common forms, the mean multigranulation decision-theoretic rough sets, the optimistic 58 59 2 60 61 62 63 64 65 1 2 3 multigranulation decision-theoretic rough sets, and the pessimistic multigranulation decision-theoretic rough sets. 4 The study is organized as follows. Some basic concepts in classical rough sets and multigranulation rough sets 5 are briefly reviewed in Section 2. In Section 3, we first analyze the loss function and the entire decision risk in the 6 context of multigranulation. Then, we propose three multigranulation decision-theoretic rough set forms that include 7 the mean multigranulation decision-theoretic rough sets, the optimistic multigranulation decision-theoretic rough sets, 8 9 and the pessimistic multigranulation decision-theoretic rough sets. When the thresholds have a special constraint, the 10 multigranulation decision-theoretic rough sets will produce one of various variables of multigranulation rough sets. 11 In Section 4, we establish the relationships among multigranulation decision-theoretic rough sets (MG-DTRS), other 12 MGRS models, single granulation decision-theoretic rough sets (SG-DTRS) and other SGRS models. Finally, Section 13 5 concludes this paper by bringing some remarks and discussions. 14 15 2. Preliminary knowledge on rough sets 16 17 In this section, we review some basic concepts such as information system, Pawlak’s rough set, and optimistic 18 multigranulation rough set. Throughout this paper, we assume that the universe U is a finite non-empty set. 19 20 2.1. Pawlak's rough set 21 Formally, an information system can be considered as a pair I =< U, AT >, where 22 • U is a non–empty finite set of objects, it is called the universe; 23 24 • AT is a non–empty finite set of attributes, such that ∀a ∈ AT, Va is the domain of attribute a. 25 ∀x ∈ U, we denote the value of x under the attribute a (a ∈ AT) by a(x). Given A ⊆ AT, an indiscernibility 26 27 relation IND(A) can be defined as 28 IND(A) = {(x, y) ∈ U × U : a(x) = a(y), a ∈ A}. (1) 29 30 The relation IND(A) is reflexive, symmetric and transitive, then IND(A) is an equivalence relation. By the indis- 31 cernibility relation IND(A), one can derive the lower and upper approximations of an arbitrary subset X of U. They 32 are defined as 33 A(X) = {x ∈ U :[x]A ⊆ X} and A(X) = {x ∈ U :[x]A ∩ X , ∅} (2) 34 respectively, where [x] = {y ∈ U :(x, y) ∈ IND(A)} is the A–equivalence class containing x. Thepair [A(X), A(X)] is 35 A referred to as the Pawlak’s rough set of X with respect to the set of attributes A. 36 37 2.2. Multigranulation rough sets 38 The multigranulation rough set (MGRS) is different from Pawlak’s rough set model because the former is con- 39 40 structed on the basis of a family of indiscernibility relations instead of single indiscernibility relation. 41 In optimistic multigranulation rough set approach, the word “optimistic” is used to express the idea that in multi 42 independent granular structures, we need only at least one granular structure to satisfy with the inclusion condition 43 between equivalence class and the approximated target. The upper approximation of optimistic multigranulation 44 rough set is defined by the complement of the lower approximation. 45 Definition 1. [32] Let I be an information system in which A1, A2, ··· , Am ⊆ AT, then ∀X ⊆ U, the optimistic O 46 m O m 47 multigranulation lower and upper approximations are denoted by A (X) and A (X), respectively, 48 X i X i i=1 i=1 49 50 m O = { ∈ ⊆ ∨ ⊆ ∨···∨ ⊆ } 51 X Ai (X) x U :[x]A1 X [x]A2 X [x]Am X ; (3) 52 i=1 53 O m m O 54 A (X) = ∼ A (∼ X); (4) 55 X i X i i=1 i=1 56 57 where [x]Ai (1 ≤ i ≤ m) is the equivalence class of x in terms of set of attributes Ai, and ∼ X is the complement of X. 58 59 3 60 61 62 63 64 65 1 2 O 3 m O m 4 By the lower approximation X Ai (X) and upper approximation X Ai (X), the optimistic multigranulation 5 i=1 i=1 6 boundary region of X is O 7 m m O 8 O BN m A (X) = X Ai (X) − X Ai (X). (5) 9 Pi=1 i i=1 i=1 10
11 Proposition 1. Let I be an information system in which A1, A2, ··· , Am ⊆ AT, then ∀X ⊆ U, we have 12 O 13 m 14 = { ∈ ∩ , ∅ ∧ ∩ , ∅∧···∧ ∩ , ∅} X Ai (X) x U :[x]A1 X [x]A2 X [x]Am X . (6) 15 i=1 16 roof 17 P . By Definition 1, we have O 18 m m O 19 x ∈ A (X) ⇔ x < A (∼ X) 20 X i X i i=1 i=1 21
22 ⇔ [x]A1 * (∼ X) ∧ [x]A2 * (∼ X) ∧···∧ [x]Am * (∼ X) 23 , , , ⇔ [x]A1 ∩ X ∅ ∧ [x]A2 ∩ X ∅∧···∧ [x]Am ∩ X ∅. 24 25 26 27 From Proposition 1, it can be seen that though the optimistic multigranulation upper approximation is defined by 28 the complement of the optimistic multigranulation lower approximation, it can also be considered as a set in which 29 objects have non–empty intersection with the target in terms of each granular structure. 30 Based on the SCED strategy, the following definition gives the formal representation of lower/upper approximation 31 in the context of multi granular structures. 32 33 Definition 2. Let I be an information system in which A1, A2, ··· , Am ⊆ AT, then ∀X ⊆ U, the pessimistic multigran- P 34 m P m 35 ulation lower and upper approximations are denoted by X Ai (X) and X Ai (X), respectively, 36 i=1 i=1 37 P 38 m A (X) = {x ∈ U :[x] ⊆ X ∧ [x] ⊆ X ∧···∧ [x] ⊆ X}; (7) 39 X i A1 A2 Am 40 i=1 P 41 m m P 42 A (X) = ∼ A (∼ X). (8) X i X i 43 i=1 i=1 44 P 45 m P m 46 By the lower approximation X Ai (X) and upper approximation X Ai (X), the pessimistic multigranulation 47 i=1 i=1 48 boundary region of X is P 49 m m P 50 P BN m A (X) = X Ai (X) − X Ai (X). (9) 51 Pi=1 i i=1 i=1 52 53 Proposition 2. Let I be an information system in which A1, A2, ··· , Am ⊆ AT, then ∀X ⊆ U, we have 54 P 55 m = { ∈ ∩ , ∅ ∨ ∩ , ∅∨···∨ ∩ , ∅} 56 X Ai (X) x U :[x]A1 X [x]A2 X [x]Am X . (10) 57 i=1 58 59 4 60 61 62 63 64 65 1 2 3 Proof. By Definition 2, we have 4 P 5 m m P 6 x ∈ X Ai (X) ⇔ x < X Ai (∼ X) 7 i=1 i=1 8 ⇔ [x] (∼ X) ∨ [x] (∼ X) ∨···∨ [x] (∼ X) 9 A1 * A2 * Am * , , , 10 ⇔ [x]A1 ∩ X ∅ ∨ [x]A2 ∩ X ∅∨···∨ [x]Am ∩ X ∅. 11 12 13 14 Different from the upper approximation of optimistic multigranulation rough set, the upper approximation of 15 pessimistic multigranulation rough set is represented as a set in which objects have non–empty intersection with the 16 target in terms of at least one granular structure. 17 18 19 3. MG-DTRS: Multigranulation decision-theoretic rough sets 20 Probabilistic approaches to rough sets have many forms, such as the decision-theoretic rough set model (DTRS) 21 [49, 52, 50], the variable precision rough set model [58], the Bayesian rough set model [37], and other related studies. 22 23 Specially, the decision-theoretic rough sets proposed by Yao [49, 50, 52] has very strong theoretical basis and sound 24 semantic interpretation. Through giving special thresholds, the decision-theoretic rough set model can degenerate 25 into the classical Pawlak rough sets, the variable precision rough set, the 0.5-probabilistic rough set, and so on. In 26 many real applications such as multi-source data analysis, knowledge discovery from data with high dimensions and 27 distributive information systems, if one applies the decision-theoretic rough sets in these cases, the multigranulation 28 version of DTRS will be very desirable. In this section, we will establish a multigranulation decision-theoretic rough 29 set framework. 30 31 3.1. Decision-theoretic rough sets 32 In this subsection, we briefly review some basic concepts in decision-theoretic rough sets. 33 In the Bayesian decision procedure, a finite set of states can be written as Ω= {ω , ω , ··· , ω }, and a finite set of 34 1 2 s m possible actions can be denoted by A = {a , a , ··· , a }. Let P(ω |x) be the conditional probability of an object x 35 1 2 r j being in state ω given that the object is described by x. Let λ(a |ω ) denote the loss, or cost, for taking action a when 36 j i j i ω 37 the state is j, the expected loss associated with taking action ai is given by 38 s R(ai|x) = λ(ai|ω j)P(ω j|x). 39 jP=1 40 41 In classical rough set theory, the approximation operators partition the universe into three disjoint classes POS (A), 42 NEG(A), and BND(A). Through using the conditional probability P(X|[x]), the Bayesian decision precedure can 43 decide how to assign x into these three disjoint regions [50, 52]. The expected loss R(ai|[x]) associated with taking 44 the individual actions can be expressed as 45 c R(a1|[x]) = λ11P(X|[x]) + λ12P(X |[x]), 46 R(a |[x]) = λ P(X|[x]) + λ P(Xc|[x]), 47 2 21 22 R(a |[x]) = λ P(X|[x]) + λ P(Xc|[x]), 48 3 31 32 c 49 where λi1 = λ(ai|X), λi2 = λ(ai|X ), and i = 1, 2, 3. When λ11 ≤ λ31 < λ21 and λ22 ≤ λ32 < λ12, the Bayesian decision 50 procedure leads to the following minimum-risk decision rules: 51 52 (P) If P(X|[x]) ≥ γ and P(X|[x]) ≥ α, decision POS (X); 53 (N) If P(X|[x]) ≤ β and P(X|[x]) ≤ γ, decision NEG(X); 54 (B) If β ≤ P(X|[x]) ≤ α, decide BND(X); 55 where 56 α = λ12−λ32 , 57 (λ31−λ32)−(λ11−λ12) 58 59 5 60 61 62 63 64 65 1 2 3 γ = λ12−λ22 , 4 (λ21−λ22)−(λ11−λ12) β = λ32−λ22 . 5 (λ21−λ22)−(λ31−λ32) 6 If a loss function with λ11 ≤ λ31 < λ21 and λ22 ≤ λ32 < λ12 further satisfies the condition: 7 8 (λ12 − λ32)(λ21 − λ31) ≥ (λ31 − λ11)(λ32 − λ22), 9 then α ≥ γ ≥ β. 10 When α > β, we have α>γ>β. The decision-theoretic rough set has the decision rules: 11 12 (P1) If P(X|[x]) ≥ α, decide POS (X); 13 (N1) If P(X|[x]) ≤ β, decide NEG(X); 14 (B1) If β< P(X|[x]) < α, decide BND(X). 15 Using these three decision rules, we get the probabilistic approximation: 16 17 apr (X) = {x : P(X|[x]) ≥ α, x ∈ U}, 18 α aprβ(X) = {x : P(X|[x]) > β, x ∈ U}. 19 20 When α = β, we have α = γ = β. The decision-theoretic rough set has the following decision rules: 21 (P2) If P(X|[x]) > α, decide POS (X); 22 (N2) If P(X|[x]) < α, decide NEG(X); 23 (B2) If P(X|[x]) = α, decide BND(X). 24 25 Using the above three decision rules, we get the probabilistic approximation: 26 apr (X) = {x : P(X|[x]) > α, x ∈ U}, 27 α 28 aprα(X) = {x : P(X|[x]) ≥ α, x ∈ U}. 29 In the framework of decision-theoretic rough sets, the Pawlak rough set model, the variable precision rough set 30 model, the Bayesian rough set model and the 0.5-probabilisitc rough set model can be pooled together and studied 31 based on the notions of conditional functions. 32 33 34 3.2. Theoretical foundation in multigranulation decision-theoretic rough sets 35 The multigranulation rough set (MGRS) is different from Pawlak’s rough set model because the former is con- 36 structed on the basis of a family of indiscernibility relations instead of single indiscernibility relation. 37 Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U, the lower/upper approximation in a multigranulation 38 rough set can be formally represented as two fusion functions, respectively, 39 40 m 41 X Ri(X) = fl(R1, R2, ··· , Rm), 42 i=1 43 m 44 X Ri(X) = fu(R1, R2, ··· , Rm), 45 i=1 46 where f is called a lower fusion function, and f is called an upper fusion function. These two functions are used to 47 l u compute the lower/upper approximation of a multigranulation rough set through fusing m granular structures. 48 In practical applications of multigranulation rough sets, the fusion function has many forms according to various 49 | 50 semantics and requirements. Conveniently, let λk(ai ω j) denote the loss, or cost, for taking action ai when the state is 51 ω j by k-th granular structures. Let P(ω j|xk) be the conditional probability of an object x being in state ω j given that 52 the object is described by xk under k-th granular structures. The expected loss associated with taking action ai is given 53 by 54 55 m s 56 R(ai|x1, x2, ··· , xm) = X X λk(ai|ω j)P(ω j|xk). (11) 57 k=1 j=1 58 59 6 60 61 62 63 64 65 1 2 3 The expected loss R(ai|x1, x2, ··· , xm) is a conditional risk. τ(x1, x2, ··· , xm) specifies which action to take, and 4 its value is one of the actions a1, a2, ··· , ar. The overall risk R is the expected loss associated with the decision rule 5 τ(x , x , ··· , x ), the overall risk is defined by 6 1 2 m 7 8 R = X R(τ(x1, x2, ··· , xm)|x1, x2, ··· , xm)P(x1, x2, ··· , xm), (12) 9 x1,x2,···,xm 10 11 where P(x1, x2, ··· , xm) is a joint probability, which is calculated through fusing (P(x1), P(x2), ··· , P(xm)) induced by 12 m granular structures induced by the same universe. 13 Given multiple granular structures R1, R2, ··· , Rm ⊆ R, the multigranulation decision-theoretic rough sets aim 14 to select a series of actions for which the overall risk is as small as possible, in which the actions include deciding 15 positive region, deciding negative region and deciding boundary region. 16 In the multigranulation decision-theoretic rough sets, there are two kinds of assumptions. One assumes that the 17 values of λk(ai|ω j), k ≤ m, are all equal each other, and the other assumes that they are not equivalent, in which each 18 granular structure has its independent loss( or cost) functions itself. In order to introduce the idea of multigranulation 19 decision-theoretic rough sets, this paper only deals with the first assumption. Hence, the determined procedure of the 20 parameters α, β and γ is consistent with that of classical decision-theoretic rough sets, and the value of each parameter 21 in every granular structure is also equal each other. The multigranulation decision-theoretic rough sets for the second 22 assumption will be established in future work. 23 24 25 3.3. Three cases of multigranulation decision-theoretic rough sets 26 Given m granular structures R1, R2, ··· , Rm ⊆ R, when λk(ai|ω j) = λl(ai|ω j), k, l ∈ {1, 2, ··· , m}, the expected loss 27 associated with taking action ai can be given by 28 29 m s 30 R(ai|x1, x2, ··· , xm) = X X λ(ai|ω j)P(ω j|xk). (13) 31 k=1 j=1 32 33 In this case, the information fusion in multigranulationdecision-theoretic rough sets can be simplified as the fusion 34 of a set of probabilities under the same universe. In this subsection, we give three multigranulation decision-theoretic 35 rough set models, which are a mean multigranulation decision-theoretic rough set (MMG-DTRS), an optimistic multi- 36 granulation decision-theoretic rough set (OMG-DTRS) and a pessimistic multigranulation decision-theoretic rough 37 set (PMG-DTRS), respectively. 38 39 3.3.1. Mean multigranulation decision-theoretic rough sets 40 In multigranulation decision-theoretic rough sets, when the loss function is fixed, judging the conditional proba- 41 bility of an object x within a target concept in m granular structures can be computed by its mathematic expectation. 42 That is to say, 43 44 E(P(X|x)) = P(X|[x] ) + P(X|[x] ) + ··· + P(X|[x] ))/m. (14) 45 R1 R2 Rm 46 The joint probability is estimated by the mean value of m conditional probabilities. Based on this idea, hence we 47 give a kind of multigranulation decision-theoretic rough set, called mean multigranulation decision-theoretic rough 48 sets. Its formal definition is as follows. 49
50 Definition 3. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U, the mean multigranulation lower and M,β 51 m M, α m 52 upper approximations are denoted by X Ri (X) and X Ri (X), respectively, 53 i=1 i=1 54 55 m M, α 56 = { | + | + ··· + | ≥ ∈ } X Ri (X) x :(P(X [x]R1) P(X [x]R2 ) P(X [x]Rm ))/m α, x U ; (15) 57 i=1 58 59 7 60 61 62 63 64 65 1 2 M,β 3 m 4 = − { | + | + ··· + | ≤ ∈ } X Ri (X) U x :(P(X [x]R1) P(X [x]R2) P(X [x]Rm ))/m β, x U ; (16) 5 i=1 6
7 where [x]Ri (1 ≤ i ≤ m) is the equivalence class of x induced by Ri, P(X|[x]Ri) is the conditional probability of the
8 equivalent class [x]Ri with respect to X, and α, β are two probability constraints. 9 M,β 10 m M, α m 11 By the lower approximation X Ri (X) and upper approximation X Ri (X), the mean multigranulation 12 i=1 i=1 13 boundary region of X is M,β 14 m m M, α 15 M BN m R (X) = X Ri (X) − X Ri (X). (17) Pi=1 i 16 i=1 i=1 17 18 Similar to the classical decision-theoretic rough sets, when the thresholds α > β, we can obtain the decision rules 19 tie-broke: | + | + ··· + | ≥ 20 (MP1) If (P(X [x]R1) P(X [x]R2 ) P(X [x]Rm ))/m α, decide POS (X);
21 (MN1) If (P(X|[x]R1 ) + P(X|[x]R2 ) + ··· + P(X|[x]Rm ))/m ≤ β, decide NEG(X);
22 (MB1) If β< (P(X|[x]R1) + P(X|[x]R2) + ··· + P(X|[x]Rm ))/m < α, decide BND(X). 23 When α = β, we have α = γ = β. The mean multigranulation decision-theoretic rough set has the following 24 decision rules: 25
26 (MP2) If (P(X|[x]R1) + P(X|[x]R2 ) + ··· + P(X|[x]Rm ))/m > α, decide POS (X); 27 (MN2) If (P(X|[x]R1 ) + P(X|[x]R2 ) + ··· + P(X|[x]Rm ))/m < α, decide NEG(X); 28 (MB2) If (P(X|[x]R1) + P(X|[x]R2) + ··· + P(X|[x]Rm ))/m = α, decide BND(X). 29 30 3.3.2. Optimistic multigranulation decision-theoretic rough sets 31 In existing optimistic multigranulation rough set approaches, the word “optimistic” is used to express the idea that 32 in multi independent granular structures, its multigranulation lower approximation only needs at least one granular 33 structure to satisfy with the inclusion condition between an equivalence class and the approximated target. While 34 35 the upper approximation of an optimistic multigranulation rough set is defined by the complement of the lower ap- 36 proximation. Based on this idea, in this part, we develop an optimistic multigranulation decision-theoretic rough 37 set. 38 In this optimistic multigranulation decision-theoretic rough set, its lower approximation collects those objects in 39 which each object has at least one granular structure satisfying the probability constraint (≥ α) between its equivalence 40 class and the approximate target, while its upper approximation collects those objects in which each object has all 41 granular structures satisfying the probability constraint (≤ β) between its equivalence class and the approximate 42 target. 43 44 Definition 4. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U, the optimistic multigranulation lower O,β 45 m O, α m 46 and upper approximations are denoted by X Ri (X) and X Ri (X), respectively, 47 i=1 i=1 48 49 m O, α = { | ≥ ∨ | ≥ ∨···∨ | ≥ ∈ } 50 X Ri (X) x : P(X [x]R1 ) α P(X [x]R2) α P(X [x]Rm) α, x U ; (18) 51 i=1 O,β 52 m 53 R (X) = U − {x : P(X|[x] ) ≤ β ∧ P(X|[x] ) ≤ β ∧···∨ P(X|[x] ) ≤ β, x ∈ U}; (19) 54 X i R1 R2 Rm i=1 55
56 where [x]Ri (1 ≤ i ≤ m) is the equivalence class of x induced by Ri, P(X|[x]Ri) is the conditional probability of the 57 equivalent class [x]Ri with respect to X, and α, β are two probability constraints. 58 59 8 60 61 62 63 64 65 1 2 O,β 3 m O, α m 4 By the lower approximation X Ri (X) and upper approximation X Ri (X), the optimistic multigranulation 5 i=1 i=1 6 boundary region of X is O,β 7 m m O, α 8 O BN m R (X) = X Ri (X) − X Ri (X). (20) 9 Pi=1 i i=1 i=1 10 11 From the definition of optimistic multigranulation decision-theoretic rough sets, one can obtain the following three 12 propositions. 13 14 Proposition 3. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U. Then, the following properties hold 15 1) m R O, α(X) ⊇ R (X), i ≤ m; 16 Pi=1 i iα 17 m O,β 2) i=1 Ri (X) ⊆ Riβ(X), i ≤ m; 18 P where R (X) = {x : P(X|[x] ) ≥ α, x ∈ U}, and R (X) = {x : P(X|[x] ) ≥ β, x ∈ U}. 19 iα Ri iβ Ri 20 21 Proposition 4. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U. Then, the following properties hold 22 1) m R O, α(X) = m R (X); Pi=1 i Si=1 iα 23 O,β 24 2) m R (X) = m R (X); Pi=1 i Ti=1 iβ 25 where R (X) = {x : P(X|[x] ) ≥ α, x ∈ U}, and R (X) = {x : P(X|[x] ) ≥ β, x ∈ U}. 26 iα Ri iβ Ri 27 28 Proposition 5. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X1 ⊆ X2 ⊆ U. Then, the following properties 29 hold 30 1) m R O, α(X ) ⊆ m R O, α(X ); Pi=1 i 1 Pi=1 i 2 31 O,β O,β 2) m R (X ) ⊆ m R (X ). 32 Pi=1 i 1 Pi=1 i 2 33 34 Similar to the classical decision-theoretic rough sets, when the thresholds α > β, we can obtain the decision rules 35 tie-broke:
36 (OP1) If ∃i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) ≥ α, decide POS (X);
37 (ON1) If ∀i ∈ {1, 2, ··· , m} such that P(X|[x]Ri) ≤ β, decide NEG(X); 38 (OB1) Otherwise, decide BND(X). 39 = = = 40 When α β, we have α γ β. The optimisitc multigranulation decision-theoretic rough set has the following 41 decision rules: 42 (OP2) If ∃i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) > α, decide POS (X); 43 (ON2) If ∀i ∈ {1, 2, ··· , m} such that P(X|[x]Ri) < α, decide NEG(X); 44 (OB2) Otherwise, decide BND(X). 45 46 3.3.3. Pessimistic multigranulation decision-theoretic rough sets 47 In decision making analysis, “Seeking common ground while eliminating differences” (SCED) is one of usual 48 49 decision strategies. This strategy argues that one reserves common decisions while deleting inconsistent decisions, 50 which can be seen as a conservative decision strategy. Based on this consideration, Qian et al. [31] proposed a so- 51 called pessimistic multigranulation rough set. In this subsection, we will combine pessimistic multigranulation rough 52 set and decision-theoretic rough set into an entire decision framework together. 53 In the pessimistic multigranulation decision-theoretic rough sets, its lower approximation collects those objects 54 in which its equivalence class from all granular structures satisfying the probability constraint (≥ α) between its 55 equivalence class and the approximate target, while its upper approximation collects those objects in which each 56 object has at least one granular structure satisfying the probability constraint (≤ β) between its equivalence class and 57 the approximate target. 58 59 9 60 61 62 63 64 65 1 2 3 Definition 5. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U, the pessimistic multigranulation lower P,β 4 m P, α m 5 and upper approximations are denoted by R (X) and R (X), respectively, 6 X i X i i=1 i=1 7 8 m P, α 9 = { | ≥ ∧ | ≥ ∧···∧ | ≥ ∈ } X Ri (X) x : P(X [x]R1 ) α P(X [x]R2 ) α P(X [x]Rm ) α, x U ; (21) 10 i=1 11 P,β 12 m = − { | ≤ ∨ | ≤ ∨···∨ | ≤ ∈ } 13 X Ri (X) U x : P(X [x]R1) β P(X [x]R2 ) β P(X [x]Rm ) β, x U ; (22) 14 i=1 15 where [x] (1 ≤ i ≤ m) is the equivalence class of x induced by R , P(X|[x] ) is the conditional probability of the 16 Ri i Ri equivalent class [x] with respect to X, and α, β are two probability constraints. 17 Ri P,β 18 m P, α m 19 By the lower approximation A (X) and upper approximation A (X), the pessimistic multigranulation 20 X i X i i=1 i=1 21 boundary region of X is 22 P,β m m P, α 23 P BN m R (X) = X Ri (X) − X Ri (X). (23) 24 Pi=1 i 25 i=1 i=1 26 From the definition of pessimistic multigranulation decision-theoretic rough set, the following three propositions 27 can be easily induced. 28 29 Proposition 6. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U. Then, the following properties hold 30 1) m R P, α(X) ⊆ R (X), i ≤ m; 31 Pi=1 i iα 32 m P,β 2) = Ri (X) ⊇ Riβ(X), i ≤ m; 33 Pi 1 where R (X) = {x : P(X|[x] ) ≥ α, x ∈ U}, and R (X) = {x : P(X|[x] ) ≥ β, x ∈ U}. 34 iα Ri iβ Ri 35 36 Proposition 7. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U. Then, the following properties hold 37 1) m R P, α(X) = m R (X); 38 Pi=1 i Ti=1 iα m P,β m 39 2) i=1 Ri (X) = i=1 Riβ(X); 40 P S where R (X) = {x : P(X|[x] ) ≥ α, x ∈ U}, and R (X) = {x : P(X|[x] ) ≥ β, x ∈ U}. 41 iα Ri iβ Ri 42 ··· ⊆ ∀ ⊆ ⊆ 43 Proposition 8. Given R1, R2, , Rm R m granular structures and X1 X2 U. Then, the following properties 44 hold 45 1) m R P, α(X ) ⊆ m R P, α(X ); Pi=1 i 1 Pi=1 i 2 46 P,β P,β 2) m R (X ) ⊆ m R (X ). 47 Pi=1 i 1 Pi=1 i 2 48 49 Similar to the classical decision-theoretic rough sets, when the thresholds α > β, we can obtain the decision rules 50 tie-broke:
51 (PP1) If ∀i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) ≥ α, decide POS (X);
52 (PN1) If ∃i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) ≤ β, decide NEG(X); 53 (PB1) Otherwise, decide BND(X). 54 When α = β, we have α = γ = β. The pessimistic multigranulation decision-theoretic rough set has the following 55 decision rules: 56 57 (PP2) If ∀i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) > α, decide POS (X); 58 59 10 60 61 62 63 64 65 1 2 3 (PN2) If ∃i ∈ {1, 2, ··· , m} such that P(X|[x]Ri ) < α, decide NEG(X); 4 (PB2) Otherwise, decide BND(X). 5 6 The following proposition establishes the relationships among the mean multigranulation decision-theoretic rough 7 sets, the optimistic multigranulation decision-theoretic rough sets, and the pessimistic multigranulation decision- 8 theoretic rough sets. 9 10 Proposition 9. Given R1, R2, ··· , Rm ⊆ R m granular structures and ∀X ⊆ U. Then, the following properties hold 11 1) m R P, α(X) ⊆ m R M, α(X) ⊆ m R O, α(X); 12 Pi=1 i Pi=1 i Pi=1 i m P,β m M,β m O,β 13 2) i=1 Ri (X) ⊇ i=1 Ri (X) ⊇ i=1 Ri (X). 14 P P P 15 16 4. Relationships between MG-DTRS and other MGRS models 17 4.1. Classical multigranulation rough sets 18 19 In the decision-theoretic rough sets, the probability value, the thresholds α and β decide its detailed form of rough 20 sets. From Yao’s work [50, 52], it follows that when α = 1 and β = 0, the decision-theoretic rough sets will degenerate 21 into the standard rough sets. In this case, we have that
22 |[x]R∩X| P(X|[x]R) = = 1 ⇔ [x]R ⊆ X, 23 |[x]R| |[x]R∩X| P(X|[x]R) = = 0 ⇔ [x]R ∩ X = Ø. 24 |[x]R| 25 Hence, 26 m R O, α(X) = {x : P(X|[x] ) ≥ 1 ∨ P(X|[x] ) ≥ 1 ∨···∨ P(X|[x] ) ≥ 1, x ∈ U} 27 Pi=1 i R1 R2 Rm 28 m O, α ⇒ i=1 Ri (X) = {x :[x]R1 ⊆ X ∨ [x]R2 ⊆ X ∨···∨ [x]Rm ⊆ X, x ∈ U}, 29 P 30 and 31 O,β m R (X) = U − {x : P(X|[x] ) ≤ 0 ∧ P(X|[x] ) ≤ 0 ∧···∧ P(X|[x] ) ≤ 0, x ∈ U} 32 Pi=1 i R1 R2 Rm m O,β 33 ⇒ Ri (X) = {x :[x]R ∩ X , Ø ∧ [x]R ∩ X , Ø ∧···∧ [x]R ∩ X , Ø, x ∈ U}. Pi=1 1 2 m 34 The multigranulation lower approximation and multigranulation upper approximation are consistent with those 35 in classical optimistic multigranulation rough sets (OMGRS) [32]. Hence, when α = 1 and β = 0, the optimistic 36 37 multigranulation decision-theoretic rough sets (OMG-DTRS) will degenerate into the optimistic multigranulation 38 rough sets (OMGRS). 39 Similarly, one has that 40 m P, α = Ri (X) = {x : P(X|[x]R1 ) ≥ 1 ∧ P(X|[x]R2 ) ≥ 1 ∧···∧ P(X|[x]Rm ) ≥ 1, x ∈ U} 41 Pi 1 ⇒ m R P, α(X) = {x :[x] ⊆ X ∧ [x] ⊆ X ∧···∧ [x] ⊆ X, x ∈ U}, 42 Pi=1 i R1 R2 Rm 43 and 44 m P,β 45 i=1 Ri (X) = U − {x : P(X|[x]R1) ≤ 0 ∨ P(X|[x]R2) ≤ 0 ∨···∨ P(X|[x]Rm) ≤ 0, x ∈ U} P P,β 46 ⇒ m R (X) = {x :[x] ∩ X , Ø ∨ [x] ∩ X , Ø ∨···∨ [x] ∩ X , Ø, x ∈ U}. 47 Pi=1 i R1 R2 Rm 48 The multigranulation lower approximation and multigranulation upper approximation are equivalent to those 49 in classical pessimistic multigranulation rough sets (PMGRS) [31]. Thus, pessimistic multigranulation decision- 50 theoretic rough sets (PMG-DTRS) will degenerate into the pessimistic multigranulation rough sets (PMGRS). 51 52 4.2. Variable multigranulation rough sets 53 When α + β = 1 and 0 ≤ β ≤ 0.5 < α ≤ 1, the decision-theoretic rough sets become the variable precision rough 54 sets. The condition 0 ≤ β ≤ 0.5 < α ≤ 1 follows that the lower approximation is a subset of the upper approximation. 55 Hence, 56 m O, α 57 i=1 Ri (X) = {x : P(X|[x]R1) ≥ α ∨ P(X|[x]R2 ) ≥ α ∨···∨ P(X|[x]Rm ) ≥ α, x ∈ U}, 58 P 59 11 60 61 62 63 64 65 1 2
3
G
T R S M G DT
M RS 4 D
5
G G
R S M MG PR S
S SG
R