Next Article in Journal
An Entropy-Based Weighted Concept Lattice for Merging Multi-Source Geo-Ontologies
Previous Article in Journal
Entropy-Based Fast Largest Coding Unit Partition Algorithm in High-Efficiency Video Coding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Granulation Entropy and Its Applications

School of Computer Science & Engineering, University of Electronic Science and Technology of China, Sichuan, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(6), 2288-2302; https://doi.org/10.3390/e15062288
Submission received: 2 April 2013 / Revised: 22 May 2013 / Accepted: 30 May 2013 / Published: 6 June 2013

Abstract

:
In the view of granular computing, some general uncertainty measures are proposed through single-granulation by generalizing Shannon’s entropy. However, in the practical environment we need to describe concurrently a target concept through multiple binary relations. In this paper, we extend the classical information entropy model to a multi-granulation entropy model (MGE) by using a series of general binary relations. Two types of MGE are discussed. Moreover, a number of theorems are obtained. It can be concluded that the single-granulation entropy is the special instance of MGE. We employ the proposed model to evaluate the significance of the attributes for classification. A forward greedy search algorithm for feature selection is constructed. The experimental results show that the proposed method presents an effective solution for feature analysis.

1. Introduction

Uncertainty analysis represents one of the most significant challenging tasks in intelligent computation. Since Shannon introduced the information entropy to measure the uncertainty of the system, a series of measures were proposed for machine learning, data mining and pattern recognition, etc. [1,2,3].
In the field of granular computing, Yu et al. introduced the fuzzy entropy for attribute reduction [4]. Hu et al. presented kernel entropy by extended Yu’s work [5]. In [6], the authors defined neighborhood entropy by using a neighborhood relation. In the view of granular computing, there are two modules in the entropy methodology mentioned above: (1) granulation of data (samples) into a set of information granules according to the relation of objects; (2) calculating the sum of the uncertainty quantity of all the information granules. We will give an example to illustrate this two-step process in detail in Section 2. It shows that granulation plays a key role in these entropy models. However, the classical information entropy theory utilizes solely the granularity structure of the given data, which is expressed by one suitable binary relation. The neighborhood entropy is only based on the neighborhood granulation; the fuzzy entropy on the fuzzy granulation; and the kernel entropy on the kernel granulation. In [7], Qian at el. proposed that there is a contradiction between two different binary relations in some data analysis issues. In other words, the decision or the view of each of decision makers may be independent for the same object in the process of some decision making. Accordingly, Qian et al. proposed multi-granulation rough set (MGRS) according to a user’s different requirements or targets of problem solving. Since then, many researchers have extended the classical MGRS by using various generalized binary relations. Lin et al. [8] proposed a covering-based pessimistic multi-granulation rough set. Xu et al. [9] proposed another generalized version, called variable precision multi-granulation rough set. There are two essential problems to be addressed when employing the rough sets model to real-world applications as similar as the information entropy model: (1) information granulation [10,11]; (2) approximate classification realized in the presence of such induced information granules [12,13]. The idea of multi-granulation is expressed through the approximation classification realizing. For example, one of the contributions in MGRS is to describe the lower and upper approximations by the multiple equivalence relations instead of the single equivalence relation. As a matter of fact, we can construct the multi-granulation structure in the process of the information granulation. Based on this idea, the contribution of this paper includes: (1) we extend the classical information entropy model to a multi-granulation entropy model (MGE) by using a series of general binary relations; (2) moreover, a number of theorems are obtained; (3) furthermore, we employ the proposed model to evaluate the significance of the attributes for classification. A forward greedy search algorithm for feature selection is constructed. The experimental results show that the proposed method presents an effective solution for feature analysis.
The paper is organized as follows: in Section 2, some basic concepts about entropy in the view of granular computing are briefly reviewed. In Section 3, the MGE model is proposed. A series of theorems about MGE is discussed. Section 4 shows the applications of MGE to feature evaluating and feature selection. Numeric experiments are reported in Section 5. Finally, Section 6 concludes the paper.

2. Entropy in the View of Granular Computing

Knowledge representation is realized via the information system ( I S ) which is a tabular form, similar to databases. An information system is pair I S = ( U , A ) , where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects, A is a nonempty finite set of attributes, and f a : U V a is a mapping for any a A , where V a is called the value set of a .
Relations, as a fundamental concept in mathematics, represent the connections of a set elements in the domain. A binary relation on U can be represented as a matrix. The matrix M R = ( r i j ) n × n is called the relation matrix of R on the universe U . The matrix M R = ( r i j ) n × n is denoted as:
M R = ( r 11 r 12 r 1 n r 21 r 22 r 2 n r n 1 r n 2 r n n )
In the classical set theory, the relations take values in the set { 0 , 1 } . In this case the relation matrix is a Boolean matrix. In the fuzzy set theory, the relations take values in the interval [ 0 , 1 ] . In this paper, we use a general binary relation R to denote any instantiated relation, such as fuzzy relation and kernel relation, etc. The fuzziness of relations is the essential characteristic in these cases. Therefore, r i j [ 0 , 1 ] in our study.
Given an information system I S = ( U , A ) and a binary relation R , U = { x 1 , x 2 , , x n } , x i U the information granule [ x i ] R is defined as:
[ x i ] R = r 1 i x 1 + r 2 i x 2 + + r n i x n
With each sample, we express the information granule in the form of fuzzy sets. Here, we give an instance about kernel entropy to illustrate the information entropy model in the view of kernel granulation [5]:
Example 1. Given an information system I S = ( U , A ) as follows:
Table 1. I S description.
Table 1. I S description.
a 1 a 2
x 1 0.10.4
x 2 0.20.3
where U = { x 1 , x 2 } and A = { a 1 , a 2 } , the two modules in the kernel entropy methodology are as follows, respectively:
(1) Information granulation:
The kernel relation is computed with Gaussian kernel as follows, where x i x j is the Euclidean distance between samples x i and x j :
r i j = exp ( x i x j 2 2 σ 2 )
Hence, we have r 12 = r 21 0.37 and r 11 = r 22 = 1 if σ is set to 0.1. The kernel granules can be constructed according Equation (2).
(2) Calculating the kernel entropy:
The cardinality of [ x i ] R is computed in the form of | [ x i ] R | = j = 1 n r i j . Thus, the expected cardinality of [ x i ] R is computed as follows, where | U | is cardinality of set U .
C a r d ¯ ( [ x i ] R ) = | [ x i ] R | | U |
The kernel entropy is defined as follows:
K H ( A ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] R )
Then, the kernel entropy of this I S is K H ( A ) = 1 2 ( log 2 1.37 2 + log 2 1.37 2 ) 0.55
Remark. To deal with nominal attributes and numerical attributes, which are common in practice, we use a extended Euclidean distance as the method introduced in literature [13]. This distance function is computed as follows:
x i x j = l = 1 N d a l ( x i , x j ) 2
d a l ( x i , x j ) = { n o m _ d i f f a l ( x i , x j )     if  a l  is a nominal attribute n u m _ d i f f a l ( x i , x j )     if  a l  is a numerical attribute
where n o m _ d i f f a l ( x i , x j ) = { 0     if  x i l = x j l 1     if  x i l x j l and n u m _ d i f f a l ( x i , x j ) = | x i l x j l | .
In a real environment, we often need to concurrently describe a target concept through multiple binary relations (e.g., neighborhood relation, kernel relation, and fuzzy relation) according to a user’s requirements or targets of problem solving. Therefore, we will study the multi-granulation entropy model in the next section.

3. Multi-Granulation Entropy

In this section, two types of multi-granulation entropy (MGE) are introduced to measure the uncertainty of knowledge in information systems. Then, the joint entropy and conditional entropy are presented in the view of multi-granulation. A number of theorems will be discussed in detail.

3.1. Two Types of MGE

Definition 1. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , the optimistic granule [ x i ] O R is computed as follows, where “ ” means “max”:
[ x i ] O R = j = 1 t [ x i ] R j
The granule [ x i ] R j is defined in forms of fuzzy set for the “max” operation. The word “optimistic” is used to express the idea that the information granulation seeks common ground while reversing difference among these general binary relations.
Definition 2. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , the pessimistic granule [ x i ] O R is computed as follows, where “ ” means “min”:
[ x i ] P R = j = 1 t [ x i ] R j
The granule [ x i ] R j is defined in forms of fuzzy set for the “min” operation. The word “pessimistic” is used to express the idea that the information granulation seeks common ground while rejection difference among these general binary relations.
Then, the expected cardinality of [ x i ] O R and [ x i ] P R are computed as follows, respectively:
C a r d ¯ ( [ x i ] O R ) = | [ x i ] O R | | U |
C a r d ¯ ( [ x i ] P R ) = | [ x i ] P R | | U |
Here, we give the definition about the two types of MGE.
Definition 3. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , B A , the first type of MGE, called optimistic multi-granulation entropy (OMGE), is denoted by:
O H ( B ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] O R )
Definition 4. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , B A , the second type of MGE, called pessimistic multi-granulation entropy (PMGE), is denoted by:
P H ( B ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] P R )
The following example will illustrate the two types of MGE in detail.
Example 2. Given a nonempty finite set of objects U = { x 1 , x 2 , x 3 , x 4 , x 5 } . Two relation matrixes about R 1 and R 2 are denoted as:
M R 1 = ( 1 0.4 0.5 0.6 0.8 0.2 1 0.4 0.5 0.1 0.6 0.9 1 0.1 0.2 0.3 0.5 0.8 1 0.2 0.4 0.5 0.6 0.1 1 ) , M R 2 = ( 1 0.6 0.1 0.2 0.1 0.2 1 0.8 0.5 0.1 0.6 0.9 1 0.2 0.2 0.3 0.7 0.6 1 0.3 0.4 0.5 0.8 0.1 1 )
The optimistic and pessimistic relation matrixes are denoted by M O and M P respectively as follows:
M O = ( 1 0.6 0.5 0.6 0.8 0.2 1 0.8 0.5 0.1 0.6 0.9 1 0.2 0.2 0.3 0.7 0.8 1 0.3 0.4 0.5 0.8 0.1 1 ) , M P = ( 1 0.4 0.1 0.2 0.1 0.2 1 0.4 0.5 0.1 0.6 0.9 1 0.1 0.2 0.3 0.5 0.6 1 0.2 0.4 0.5 0.6 0.1 1 )
Every row of the matrixes denotes the information granule (e.g., [ x 1 ] O R = 1 x 1 + 0.6 x 2 + 0.5 x 3 + 0.6 x 4 + 0.8 x 5 ).
OMGE and PMGE are computed according to Equations (12) and (13):
O H ( A ) { R 1 , R 2 } = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] O R ) 0.54
P H ( A ) { R 1 , R 2 } = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] P R ) 0.49
Definition 5. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , B 1 , B 2 A ,the optimistic information granules [ x i ] P R 1 and [ x i ] P R 2 are induced by B 1 and B 2 . The optimistic joint entropy is expressed as:
O H ( B 1 B 2 ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] O R 1 [ x i ] O R 2 )
where “ ” means “min”.
The pessimistic joint entropy is defined as follows:
P H ( B 1 B 2 ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] P R 1 [ x i ] P R 2 )
Definition 6. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , B 1 , B 2 A , the optimistic conditional entropy of B 2 to B 1 is expressed as:
O H ( B 2 | B 1 ) = O H ( B 2 B 1 ) O H ( B 1 )
Similarly, we have the pessimistic conditional entropy:
P H ( B 2 | B 1 ) = P H ( B 2 B 1 ) P H ( B 1 )
The conditional entropy reflects the uncertainty of B 2 if B 1 is given.

3.2. Some Theorems about MGE

Theorem 1. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given only one equivalence binary relation = { R e } , we have:
O H ( A ) = P H ( A ) = H ( A )
where H ( A ) is Shannon’s entropy.
Proof O H ( A ) = P H ( A ) is straightforward. The equivalence binary relation is computed as:
r i j = { 1           i f    x i = x j 0            i f    x i x j
The samples are divided into disjoint X 1 , X 2 , X m , where x i = x j X k . Assumed there are w k samples in X k , then H ( A ) = i = 1 w w k | U | log 2 w k | U | . c a r d ¯ ( [ x i ] R ) = | X k | | U | = w k | U | if x i X k , we have:
O H ( A ) = P H ( A ) = 1 | U | i = 1 n log 2 C a r d ¯ ( [ x i ] R ) = 1 | U | x X 1 log 2 w 1 | U | + + ( 1 | U | x X k log 2 w k | U | ) = w 1 | U | log 2 w 1 | U | + + ( w k | U | log 2 w k | U | ) = H ( A )
It is shown that the MGE is a natural generalization of the Shannon’s entropy in the view of granulation by the proof above. In [14,15], the authors generalized Shannon’s entropy to fuzzy entropy, kernel entropy and neighborhood entropy, respectively. These entropy models utilize solely the granularity structure of the given data, which is expressed by one suitable binary relation. The neighborhood entropy is only based on the neighborhood granulation; the fuzzy entropy on the fuzzy granulation; and the kernel entropy on the kernel granulation. Hence, it also can be concluded that the single-granulation entropy, such as neighborhood entropy, kernel entropy, fuzzy entropy, etc., is the special instance of MGE.
Theorem 2. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , 1 2 , we have:
O H ( A ) 1 O H ( A ) 2
P H ( A ) 1 P H ( A ) 2
Proof x i U , 1 2 , we have [ x i ] O 1 [ x i ] O 2 . Therefore, c a r d ¯ ( [ x i ] O 1 ) c a r d ¯ ( [ x i ] O 2 ) . Obviously, O H ( A ) 1 O H ( A ) 2 . Similarly, [ x i ] P 1 [ x i ] P 2 , P H ( A ) 1 P H ( A ) 2 .
For convenience, the monotonicity of entropy value induced by the set of relations is called the granulation monotonicity.
Corollary 1. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , we have:
O H ( A ) P H ( A )
Proof x i U we have j = 1 t [ x i ] R j = [ x i ] P R [ x i ] O R = j = 1 t [ x i ] R j . Therefore, c a r d ¯ ( [ x i ] P R ) c a r d ¯ ( [ x i ] O R ) . Obviously, O H ( A ) P H ( A ) .
Corollary 2. Let I S = ( U , A ) be an information system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , 1 2 , B 1 , B 2 A , we have:
O H ( B 2 | B 1 ) 2 O H ( B 2 | B 1 ) 2
P H ( B 2 | B 1 ) 1 P H ( B 2 | B 1 ) 2
Proof According to Lemma 4.1 in Ref. [16], we know that the combination of information granules by “ ” operator will increase the conditional entropy monotonously. Similarly, it can be concluded that the conditional entropy will decrease through combining information granules by “ ” operator. QED

4. Feature Selection Based on MGE

One of the most important applications of information entropy theory is to evaluate the classification power of the attributes in a decision system by computing the significance of the condition attributes for the resulting decision. This entropy-based model was widely used in feature selection algorithms for categorical data [17]. However, classical entropy models cannot be used to express multi-granulation which represents the different points of view for describing one concept. Here, we show a feature selection technique based on MGE.
If the set of samples is assigned with a decision attribute D , we call this information system I S = ( U , C , D ) a decision system, where C are conditional attributes. Therefore, as we explain in Definition 6 that multi-granulation conditional entropy O H ( D | C ) ( P H ( D | C ) ) is the uncertainty of D if condition attributes C are given, conditional entropy reflects the relevance between condition attributes and decision.
Definition 7. Let I S = ( U , C , D ) be a decision system, where U = { x 1 , x 2 , , x n } is a nonempty finite set of objects. Given a set of general binary relations = { R 1 , R 2 , R t } , B C , we thus define significance of attribute subset B in the multi-granulation of view:
O S I G ( B , D ) = O H ( D ) O ( D | B ) = O H ( D ) + O H ( B ) O H ( D B )
P S I G ( B , D ) = P H ( D ) P ( D | B ) = P H ( D ) + P H ( B ) P H ( D B )
O S I G ( B , D ) is used to evaluate the significance of attribute subset B by the optimistic multi-granulation. Similar to O S I G ( B , D ) , P S I G ( B , D ) is another evaluation measure of the attributes. The pessimistic granules, which are formed by the binary relations , are used to compute P S I G ( B , D ) . It is easy to observe that O S I G ( B , D ) ( P S I G ( B , D ) ) becomes a symmetric uncertainty measure. In fact this is mutual information of B and D defined in Shannon’s information theory if B and D generate Boolean equivalence relations according to Equation (23) [18]. As it is well-known, mutual information is widely applied in evaluating features and constructing decision trees [19,20], the classical definition of mutual information can just be used to deal with only one granulation. The multi-granulation significance defined here can be used to express lots of views with a series of binary relations. Equations (30) and (31) can be used to find the significant features for classification. Actually, it is impractical to get the optimal subset of features from 2 n 1 candidates through exhaustive search, where n is the number of features. The greedy search guided by some heuristics is usually more efficient than the plain brute-force exhaustive search. In a forward greedy search, one starts with an empty set of attributes, and keeps adding features to the subset of selected attributes one by one. Each selected attribute maximizes the increment of significance of the current subset. A forward search algorithm for feature selection based on MGE is written as follows. Here, O S I G and P S I G are denoted as S I G uniformly.
Algorithm 1. Feature selection based on MGE(OMGE or PMGE)
Input: decision system I S = ( U , C , D ) , binary relations = { R 1 , R 2 , R t } and stopping threshold ε .
Output: selected features r e d .
  • r e d φ
  • while r e d C
  •  for each a i ( C r e d )
  •   compute s i g i = S I G ( a i r e d , D )
  •  end for
  •  find the maximal s i g i and the corresponding attribute a i
  •  if s i g i S I G ( r e d , D ) > ε
  •    r e d r e d a i
  •  else
  •   exit while
  •  end if
  • end while
  • return r e d
The time complexity of the algorithm is O ( n 2 m log m ) , where n and m are the numbers of features and samples, respectively. It is worth noting that the proposed measures of mutual information can be incorporated with other search strategies used in other feature selection algorithms, such as ABB (Automatic Branch and Bound), probabilistic search [21] and GP (Genetic programming) [22]. In this study, we are not going to compare the influence of search strategies on the results of feature selection. Here we focus on the comparison of the proposed method when dealing with different evaluation measures.

5. Experimental Analysis

In this section, we compare the effectiveness of MGE in evaluating feature quality. The data sets are downloaded from the UCI Machine Learning Repository. They are described in Table 2. The numerical attributes of the samples are linearly normalized as follows:
x = ( x x min ) / ( x max x min )
where x min and x max are the bounds of the given attribute. Three popular leaning algorithms such as CART, liner SVM and RBF SVM are introduced to evaluate the quality of selected features. The experiments were run in a 10-fold cross validation mode. The parameters of the linear SVM and RBF SVM are taken as the default values (the use of the MATLAB toolkit osu_svm3.00).
Table 2. Data description.
Table 2. Data description.
IDDataSamplesFeaturesClass
1wine178133
2wdbc569312
3iono351342
4heart270132
5glass21497
6wpbc198332
7sonar208602
In the experiment, we employ three symmetric membership functions for multi-granulation. One is the kernel relation defined as Equation (3) in Example 1; the other two are computed as follows, respectively:
r i j = { 1           i f    x i x j δ 0           i f    x i x j > δ
r i j = { 1 4 × x i x j           , i f    x i x j 0.25 0           , i f    x i x j > 0.25
Equation (33), called neighborhood relation, is used to compute neighborhood entropy (NE) in Ref. [6] where the threshold δ 0 . According to this definition, the samples in a neighborhood granule have the distance is less than the threshold δ . Literature [6] has explained that the result is optimum if threshold δ is set between 0.1 and 0.2. In the following, if not specified, δ = 0.15 . Similarly, the fuzzy entropy (FE) is proposed based on the fuzzy relation according to Equation (34) [4]. We compare MGE with kernel entropy (KE), NE and FE, where the compared methods are the typical single-granulation entropy. The parameters of the KE and NE are kept consistent in Ref. [5] and Ref. [6]. We compute the significance of single feature with five evaluation functions, such as OMGE, PMGE, KE, NE and FE. At the same time, we reported the classification accuracies of the each feature based on the use of the linear SVM and RBF SVM.
Two data sets wine and glass are used in the experiment. There are 13 features in the wine and nine features in the glass dataset. The results are given in Figure 1 and Figure 2. As to the wine data, the features 1, 6, 7, 10, 11, 12, 13 produce higher values of all evaluation functions, as shown in Figure 1a; at the same time, we can also find that the classification accuracies of these features are better than others (again shown in Figure 1b). As to the glass data, features 2, 3, 4, 8 are better than others in terms of the five evaluating functions, corresponding the classification accuracies of features 2, 3, 4, 8 are also higher than the other features. These results show that all five evaluating functions can produce good estimates of classification ability of the features. It can be concluded that OMGE and PMGE are competent with other entropy models.
Figure 1. Significance and accuracy of single feature (wine). (a) Significance of a single feature computed with different evaluating. (b) Classification accuracies obtained for single features when using linear SVM and RBF SVM.
Figure 1. Significance and accuracy of single feature (wine). (a) Significance of a single feature computed with different evaluating. (b) Classification accuracies obtained for single features when using linear SVM and RBF SVM.
Entropy 15 02288 g001aEntropy 15 02288 g001b
Figure 2. Significance and accuracy of single feature(glass). (a) Significance of a single feature computed with different evaluating. (b) Classification accuracies obtained for single features when using linear SVM and RBF SVM.
Figure 2. Significance and accuracy of single feature(glass). (a) Significance of a single feature computed with different evaluating. (b) Classification accuracies obtained for single features when using linear SVM and RBF SVM.
Entropy 15 02288 g002
The above results show MGE can be used to evaluate single attributes. Now, we show the effectiveness in attribute reduction. The selected features with different algorithms are presented in Table 3 and Table 4, respectively. Regarding OMGE, PMGE, FE, NE and KE, the orders of the features presented in the tables are the orders that the features are kept being added to the feature space. These orders reflect the relative significance of features in terms of the corresponding measures. Some results can be derived from the selected attributes. First, whatever attribute selection techniques have been used, most of the attributes in all datasets can be deleted. The reduction rate is high to 90% for some datasets, such as sonar and wpbc. Second, some selected attributes are slightly different. Especially, some of the selected features are the subset of attributes selected by other models.
Table 3. Subsets of features selected with OMGE and PMGE.
Table 3. Subsets of features selected with OMGE and PMGE.
DataOMGEPMGE
wine7,1,10,137,1,11,4
wdbc29,22,23,12,924,29,23,30,26,9,13,10,28,3,27
iono5,6,8,25,28,24,10,215,6,34,29,8,23
heart13,12,3,11,1,7,413,12,3,1,10,4
glass3,7,4,9,53,7,4,9,5,1
wpbc34,2,13,14,72,34,13,7,23
sonar12,27,21,37,32,30,5412,16,26,40,48
Table 4. Subsets of features selected with FE, NE and KE.
Table 4. Subsets of features selected with FE, NE and KE.
DataFENEKE
wine7, 1, 10,137,1,11,47,1,10,13
wdbc29,22,23,12,924,29,23,30,8,27,26,13,10,3,1929,22,23,9,12
iono5,6,8,25,28,24,34,75,6,34,29,8,235,6,8,25,28,24,34,7
heart13,12,3,10,1,7,11,2,8,413,12,3,10,1,4,513,12,3,10,1,7,11
glass3,7,4,9,53,7,4,9,5,13,7,4,9,5
wpbc34,2,13,14,72,34,13,7,2334,2,13,14,7
sonar12,27,21,37,32,30,5412,16,26,40,4812,16,26,37,22,32,28
As we know, we consider the ranking of features in feature selection, sometimes, a little difference in feature qualities may lead to completely different ranking. Therefore, the great difference between these selected features is the difference between the qualities of features computed with diverse granularities. In other words, there is a inconsistent relationship between its values under one-granularity and those under the another granularity. In [7], the authors give a tentative study that multi-granulation model will display its advantage for rule extraction when two granularities process a contradiction relationship. We will test this idea by the following experiment. We build classification models with the selected features and test their classification performance based on 10-fold cross validation. The average value and standard deviation are used to measure the classification performance. We compare the raw data, MGE, FE, NE and KE in Table 5, Table 6 and Table 7, where learning algorithms CART, linear SVM and RBF SVM are introduced to evaluate the selected features.
Table 5. Classification accuracies based on CART (%).
Table 5. Classification accuracies based on CART (%).
DataRaw dataOMGEPMGEFENEKE
wine86.4 ± 7.992.2 ± 7.589.9 ± 8.592.2 ± 7.589.9 ± 8.592.2 ± 7.5
wdbc90.3 ± 6.093.0 ± 3.893.5 ± 3.993.0 ± 3.894.0 ± 3.293.0 ± 3.8
iono86.4 ± 7.287.5 ± 5.688.6 ± 6.588.1 ± 6.088.6 ± 6.588.1 ± 6.0
heart77.0 ± 5.578.5 ± 7.380.4 ± 9.075.2 ± 9.280.0 ± 7.777.8 ± 9.2
glass69.2 ± 13.265.7 ± 12.965.1 ± 14.665.7 ± 12.965.1 ± 14.665.7 ± 12.9
wpbc70.2 ± 5.470.7 ± 10.371.1 ± 11.970.7 ± 10.371.1 ± 11.970.7 ± 10.3
sonar57.7 ± 9.273.1 ± 12.672.2 ± 15.373.1 ± 12.672.2 ± 15.362.6 ± 13.2
Table 6. Classification accuracies based on liner SVM (%).
Table 6. Classification accuracies based on liner SVM (%).
DataRaw dataOMGEPMGEFENEKE
wine98.3 ± 2.797.2 ± 3.994.4 ± 5.297.2 ± 3.994.4 ± 5.297.2 ± 3.9
wdbc98.0 ± 1.996.1 ± 2.196.3 ± 2.196.1 ± 2.195.9 ± 2.196.1 ± 2.1
iono87.5 ± 6.483.4 ± 5.385.0 ± 5.985.0 ± 5.385.0 ± 5.985.0 ± 5.3
heart84.1 ± 9.383.0 ± 8.981.9 ± 7.282.9 ± 9.482.2 ± 6.082.6 ± 8.6
glass55.7 ± 7.760.4 ± 9.157.1 ± 8.760.4 ± 9.157.1 ± 8.760.4 ± 9.1
wpbc77.3 ± 5.776.3 ± 3.076.3 ± 3.076.3 ± 3.076.3 ± 3.076.3 ± 3.0
sonar64.4 ± 15.764.9 ± 11.667.8 ± 15.764.9 ± 11.667.8 ± 15.765.5 ± 13.5
Table 7. Classification accuracies based on RBF SVM (%).
Table 7. Classification accuracies based on RBF SVM (%).
DataRaw dataOMGEPMGEFENEKE
wine97.8 ± 2.996.7 ± 3.995.0 ± 4.196.7 ± 3.995.0 ± 4.196.7 ± 3.9
wdbc97.0 ± 2.696.7 ± 2.197.2 ± 2.396.7 ± 2.197.5 ± 2.596.7 ± 2.1
iono94.0 ± 4.293.7 ± 4.892.6 ± 5.794.0 ± 4.692.6 ± 5.794.0 ± 4.6
heart79.2 ± 9.681.1 ± 7.983.3 ± 7.682.6 ± 9.484.0 ± 7.881.5 ± 9.2
glass68.3 ± 12.162.7 ± 11.964.1 ± 11.262.7 ± 11.964.1 ± 11.262.7 ± 11.9
wpbc78.9 ± 6.074.3 ± 7.475.3 ± 7.774.3 ± 7.475.3 ± 7.774.3 ± 7.4
sonar58.5 ± 16.065.9 ± 14.964.1 ± 11.265.9 ± 14.964.1 ± 11.261.1 ± 7.2
Comparing the performance of raw data and granulation-based selection, we can find although most of features have been removed, most of the classification accuracies derived from the reduced data sets do not decrease, but increase. It shows there are redundant and irrelevant attributes in the raw data.
The experimental results show that no matter which classification algorithms are used, MGE is better than or equivalent to KE. Table 6 shows that MGE outperforms FE and NE with respect to liner SVM. As to CART learning algorithm in Table 5, MGE is better than or equivalent to NE for six of the seven databases. It can be concluded that MGE is a better choice for the diverse granularities. Actually, the different decision makers have different granulation points of view. Therefore, it is necessary to take diverse factors into consideration for granular computing in the real world.

6. Conclusions

In this paper, the classical single-granulation entropy theory has been extended. As a result of this extension, a multi-granulation entropy model (MGE) has been developed. The uncertainty of the information system is defined by using multiple relations on the universe. These relations can be chosen according to a user’s requirements or targets of problem solving.
In MGE model, we introduce OMGE and PMGE to describe the relations between different granularities. Based on the mutual information defined through MGE, we proposed the forward greed features selection algorithms, which will be helpful for applying this theory to practical issues. MGE provides an effective approach in the context of multiple granulations. We conclude that the single-granulation entropy is the special instance of MGE. The experimental result shows that MGE will display its advantage for rule extraction and knowledge discovery when the different granularities in information systems possess a contradiction or inconsistent relationship.
The future work could move along two directions. First, the existing feature selection algorithms based entropy sometimes might not be robust enough for real-world applications. How to improve it is an important issue. Second, we will continue to construct MGE models with various binary relations for discussing the common properties of this kind of entropy model.

Acknowledgments

This work is supported by the Fundation of Science & Technology Department of Sichuan Province, China (Grant No. 2012GZ0061).

Conflict of Interest

The authors declare no conflict of interest.

References

  1. Knuth, K.H. Lattice duality: The origin of probability and entropy. Neurocomputing 2005, 67, 245–274. [Google Scholar] [CrossRef]
  2. Harremoës, P.; Vajda, I. On the bahadur-efficient testing of uniformity by means of the entropy. IEEE Trans. Inform. Theory 2008, 54, 321–331. [Google Scholar] [CrossRef]
  3. Santhanam, N.P.; Modha, D. Lossy Lempel-Ziv like compression algorithms for memoryless sources. In Proceedings of 49th Annual Allerton Conference, Monticello, IL, USA, 28–30 September 2011; pp. 1751–1756.
  4. Yu, D.R.; Hu, Q.H.; Wu, C.X. Uncertainty measures for fuzzy relations and their applications. Appl. Soft Comput. 2007, 7, 1135–1143. [Google Scholar] [CrossRef]
  5. Hu, Q.H.; Zhang, L.; Chen, D.G.; Pedrycz, W.; Yu, D.R. Gaussian kernel based fuzzy rough sets: model, uncertainty measures and applications. Int. J. Approx. Reason. 2010, 51, 453–471. [Google Scholar] [CrossRef]
  6. Hu, Q.H.; Zhang, L.; Zhang, D.; Pan, W.; An, S.; Pedrycz, W. Measuring relevance between discrete and continuous features based on neighborhood mutual information. Expert Syst. Appl. 2011, 38, 10737–10750. [Google Scholar] [CrossRef]
  7. Qian, Y.H.; Liang, J.Y.; Yao, Y.Y.; Dang, C.Y. MGRS: A multi-granulation rough set. Inform. Sci. 2010, 180, 949–970. [Google Scholar] [CrossRef]
  8. Lin, G.P.; Li, J.J. A covering-based pessimistic multigranulation rough set. In Proceedings of International Conference on Intelligent Computing, Zhengzhou, China, 11–14 August, 2011; pp. 673–680.
  9. Xu, W.H.; Zhang, X.T.; Wang, Q.R. A generalized multi-granulation rough set approach. In Proceedings of International Conference on Intelligent Computing, Zhengzhou, China, 11–14 August, 2011; pp. 681–689.
  10. Qin, K.Y.; Yang, J.L.; Pei, Z. Generalized rough sets based on reflexive and transitive relations. Inform. Sci. 2008, 178, 4138–4141. [Google Scholar] [CrossRef]
  11. Zhu, W.; Wang, S.P. Rough matroids based on relations. Inform. Sci. 2013, 232, 241–252. [Google Scholar] [CrossRef]
  12. Tang, J.G.; She, K.; Min, F.; Zhu, W. A matroidal approach to rough set theory. Theor. Comput. Sci. 2013, 471, 1–11. [Google Scholar] [CrossRef]
  13. Jing, S.Y.; She, K.; Ali, S. A Universal neighbourhood rough sets model for knowledge discovering from incomplete heterogeneous data. Expert Syst. 2013, 30, 89–96. [Google Scholar] [CrossRef]
  14. Al-Sharhan, S.; Karray, F.; Gueaieb, W.; Basir, O. Fuzzy entropy: a brief survey. In The 10th IEEE International Conference on Fuzzy Systems; IEEE: Melbourne, Australia, 2001; Volume 3, pp. 1135–1139. [Google Scholar]
  15. Hu, Q.H.; Yu, D.R. Neighborhood Entropy. In Proceedings of the Eighth International Conference on Machine Learning and Cybernetics, Baoding, China, 12–15 July 2009; Volume 3, pp. 1776–1782.
  16. Wang, G.Y. Rough reduction in algebra view and information view. Int. J. Intell. Syst. 2003, 18, 679–688. [Google Scholar] [CrossRef]
  17. Slezak, D. Approximate entropy reducts. Fund. Informat. 2002, 53, 365–390. [Google Scholar]
  18. Hu, Q.H.; Yu, D.R.; Xie, Z.X.; Liu, J.F. Fuzzy probabilistic approximation spaces and their information measures. IEEE Trans. Fuzzy Syst. 2006, 14, 191–201. [Google Scholar]
  19. Yu, L.; Liu, H. Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res. 2004, 5, 1205–1224. [Google Scholar]
  20. Battiti, R. Using mutual information for selecting features in supervised neural net learning. IEEE T. Neural Networ. 1994, 5, 537–550. [Google Scholar] [CrossRef] [PubMed]
  21. Dash, M.; Liu, H. Consistency-based search in feature selection. Artif. Intell. 2003, 151, 155–176. [Google Scholar] [CrossRef]
  22. Muni, D.P.; Pal, N.R.; Das, J. Genetic programming for simultaneous feature selection and classifier design. IEEE T. Syst. Man. Cy. B 2006, 36, 106–117. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Zeng, K.; She, K.; Niu, X. Multi-Granulation Entropy and Its Applications. Entropy 2013, 15, 2288-2302. https://doi.org/10.3390/e15062288

AMA Style

Zeng K, She K, Niu X. Multi-Granulation Entropy and Its Applications. Entropy. 2013; 15(6):2288-2302. https://doi.org/10.3390/e15062288

Chicago/Turabian Style

Zeng, Kai, Kun She, and Xinzheng Niu. 2013. "Multi-Granulation Entropy and Its Applications" Entropy 15, no. 6: 2288-2302. https://doi.org/10.3390/e15062288

Article Metrics

Back to TopTop