Next Article in Journal
Constrained Inference When the Sampled and Target Populations Differ
Previous Article in Journal
A Cross-Entropy-Based Admission Control Optimization Approach for Heterogeneous Virtual Machine Placement in Public Clouds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Preference Inconsistence-Based Entropy

1
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Computer School, China West Normal University, Nanchong 637009, China
*
Author to whom correspondence should be addressed.
Entropy 2016, 18(3), 96; https://doi.org/10.3390/e18030096
Submission received: 12 November 2015 / Revised: 3 March 2016 / Accepted: 9 March 2016 / Published: 15 March 2016

Abstract

:
Preference analysis is a class of important issues in ordinal decision making. As available information is usually obtained from different evaluation criteria or experts, the derived preference decisions may be inconsistent and uncertain. Shannon entropy is a suitable measurement of uncertainty. This work proposes the concepts of preference inconsistence set and preference inconsistence degree. Then preference inconsistence entropy is introduced by combining preference inconsistence degree and Shannon entropy. A number of properties and theorems as well as two applications are discussed. Feature selection is used for attribute reduction and sample condensation aims to obtain a consistent preference system. Forward feature selection algorithm, backward feature selection algorithm and sample condensation algorithm are developed. The experimental results show that the proposed model represents an effective solution for preference analysis.

1. Introduction

Multiple attribute decision making refers to making preference decisions between available alternatives characterized by multiple, usually conflicting, attributes [1]. A multiple attributes decision making system can be depicted by using the alternative performance matrix, where element x i j is the rating of alternative i with respect to attribute j. A weight vector is designed to indicate the significance of every attribute. Various methods for finding weights can be found in the literature, such as the analytic hierarchy process (AHP) method [2], weighted least squares method [3], Delphi method [4], the entropy method, multiple objective programming [5,6], principal element analysis [6], etc. Ordinal decision making is a class of important issues in multiple attribute decision making where the conditional attributes and the decision are all ordinal. There are numerous applications associated with an assignment of objects evaluated by a set of criteria to pre-defined preference-ordinal decision classes, such as credit approval, stock risk estimation, and teaching evaluations [7]. The alternative performance matrix and weights vector method cannot reflect the ordinal nature of decisions based on conditional attributes, therefore, is not a suitable tool for ordinal decision making.
Preference analysis is a class of important issues in ordinal decision making. Preference relations are very useful in expressing a decision maker’s preference information in decision problems in various fields. During the past years, the use of preference relations has received increasing attention, and a number of studies have focused on this issue, and various types of preference relations have been developed, such as, multiplicative preference relation introduced by Saaty [2], incomplete multiplicative preference relation introduced by Harker [8], interval multiplicative preference relation introduced by Saaty and Vargas [9], incomplete interval multiplicative preference relation introduced by Xu [10], triangular fuzzy multiplicative preference relation introduced by Van Laarhoven and Pedrycz [11], fuzzy preference relation introduced by Luce and Suppes [12], incomplete fuzzy preference relation introduced by Alonso et al. [13], interval fuzzy preference relation introduced by Xu and Herrera et al., and linguistic preference relation introduced by Herrera et al. [14,15,16].
One task of preference analysis is to predict a decision maker’s preference according to available information. A lot of methods for preference analysis can be found in the literature. Greco et al. introduced a dominance rough set model that is suitable for preference analysis [17,18]. They extracted dominance relations from multiple criteria, and constructed similarity relations and equivalence relations from numerical attributes and nominal features respectively. In [19], Hu et al. developed a rough set model based on fuzzy preference by integrating fuzzy preference relations with fuzzy rough set model, and generalized the dependency to compute relevance between the criteria and decision and obtain attribute reductions. In ordinal decision making, there exists a basic assumption that better conditions should result in better decisions, which usually holds for a single criterion. For multi-criteria ordinal decision system, however, the decision is made considering various sources of information which is usually obtained from different evaluation criteria or experts and thus the decision preference might be inconsistent with some criteria.
Preference inconsistence leads to decision uncertainty. One of the decision uncertainty measures which have been proposed by researchers is the Shannon entropy concept. The entropy concept is used in various scientific fields. In transportation models, entropy acts as a measure of dispersal of trips between origin and destinations. In physics, the word entropy has important physical implications as the amount of “disorder” of a system. Also the entropy associated with an event is a measure of the degree of randomness in the event. The concept of Shannon’s entropy has an important role in information theory and is used to refer to a general measure of uncertainty [1]. In preference analysis, some methods based on entropy have been introduced in the literature. Abbas presented an optimal question-selection algorithm to elicit von Neumann and Morgenstern utility values by using information theory and entropy-coding principles for a set of ordered prospects of a decision situation [20]. Lotfi extended the Shannon entropy method to interval data for obtaining the weights of criteria of a multiple attributes decision making problem [1]. Yang and Qiu proposed the expected utility-entropy (EU-E) measure of risk and a decision making model about preference orderings among gambles based on expected utility and entropy of an action involving risk [21]. Abbas introduced an analogy between probability and utility through the notion of a utility density function and illustrate the application of this analogy to the maximum entropy principle [22]. A normative method was presented to assign unbiased utility values when only incomplete preference information was available about the decision maker [23]. In this research, Shannon entropy has been expanded to preference inconsistence entropy and then is used to measure the uncertainty of preference decision to conditional attributes.
Based on this idea, the contributions of this work include: (1) this paper defines a preference inconsistence set and preference inconsistence degree, and some properties are given; (2) based on preference inconsistence degree and Shannon entropy, the notion of preference inconsistence entropy is proposed; (3) relative attribute significance is defined, feature selection is investigated, and meanwhile, a forward feature selection algorithm and backward feature selection algorithm are developed; (4) a sample condensation algorithm is also given; (5) finally, some experiments are completed to verify the proposed approach. The experimental results show that relative attribute significance based on preference inconsistence entropy can reflect the significance of features in a preference decision and the developed feature selection and sample condensation algorithms are helpful to make the data set preference consistent.
The remainder of the paper is organized as follows: Section 2 provides a brief overview of preference relations and Shannon entropy, and some basic formulas that will be used in the remaining sections of this paper are given. In Section 3, the concepts of preference inconsistence degree and preference inconsistence entropy (PIE) are proposed. Some properties of PIE are discussed and a series of theorems are put forward. Section 4 presents two applications of PIE: feature selection and sample condensation. Relative attribute significance is defined. A forward feature selection algorithm and backward feature selection algorithm as well as sample condensation algorithm are developed. Numerical experiments are reported in Section 5. Finally, Section 6 wraps up with conclusions.

2. Preliminary Knowledge

In this section, we will review some basic concepts about preference relation and information entropy, which have been addressed in [24,25,26,27,28,29,30,31].

2.1. Information Entropy

In 1948, Claude Shannon introduced the entropy term, nowadays named Shannon entropy. Shannon entropy is a simple quantitative measure of uncertainty. The following concepts are taken from Shannon’s original paper [30]. Let X be a random variable, taking a finite number of possible values x 1 , x 2 ,   x 3 ,   ,   x n with respective probabilities p i 0 for i = 1, 2, …, n and i = 1 n p i = 1 . The Shannon entropy H(X) is defined as:
H ( X ) = i = 1 n p i l o g p i  
Suppose there are two events, X and Y, in question with m possibilities for the first and n for the second. Let p(i, j) be the probability of the joint occurrence of i for the first and j for the second. The entropy of the joint event is:
H ( X , Y ) = i , j p ( i , j ) l o g p ( i , j )
For any particular value i for X, we can assume there is a conditional probability p ( j | i ) where Y has the value i. The conditional entropy of Y, H ( Y | X ) , is defined as the average of entropy of Y for each value of X weighted according to the probability of getting that particular X. That is:
H ( Y | X ) = i , j p ( i , j ) l o g p ( j | i ) = i , j p ( i , j ) l o g p ( i , j ) + i , j p ( i , j ) l o g j p ( i , j ) = H ( X , Y ) H ( X )
The conditional entropy of Y when X is known is the joint event X, Y minus the entropy of X. We will use this property to calculate the relative attribute significance for our feature selection algorithm.

2.2. Preference Relation

A decision system is a tuple D S = < U , A D , f > , where U is a nonempty finite set of objects, A = { a 1 , a 2 , , a m } is a set of conditional attributes to characterize the objects, D = { d 1 , d 2 , , d n } is the decision, f a : U V a is a mapping for any a A , where V a is called the value set of attribute a. The decision system DS is referred to as ordinal decision system, if d 1 d 2 d n , and a i 1 a i 2 a i k   holds   for any   a i A , where { a i 1 , a i 2 , , a i k } is the value set of conditional attribute a i ( 1 i m ) .
Let D S = < U , A D , f > is an ordinal decision system. For preference analysis, the following relations are defined. C A , x , y U :
R C = { ( x , y ) U × U   : f ( x , a ) f ( y , a ) ,   a C }
R C = { ( x , y ) U × U   : f ( x , a ) f ( y , a ) ,   a C }
R D = { ( x , y ) U × U   : f ( x , D ) f ( y , D ) }
R D = { ( x , y ) U × U   : f ( x , D ) f ( y , D ) }
where f ( x , a ) is attribute value of object x with respect to attribute a, and f ( x , D ) is the decision of object x.
R C and R C are referred to as conditional domination relation. R D   and   R D are referred to as decision domination relation.
If ( x , y ) R C , we say that object x is greater than or equal to object y with respect to C. In other word, object x is not less than object y with respect to C .
If ( x , y ) R C , we say that object x is less than or equal to object y with respect to C. In other word, object x is not greater than object y with respect to C .
If ( x , y ) R D , we say that the decision object x is better than or equal to object y. In other word, the decision of object x is not worse than that of y.
If ( x , y ) R D , we say that the decision object x is worse than or equal to object y. In other word, the decision of object x is not better than that of y.
In preference analysis, x’s decision should not be worse than y’s if x is better than y in terms of criterion a; otherwise, the decision is inconsistent [14]. Similarly, x’s decision should not be better than y’s if x is less than y in terms of criterion a; otherwise, the decision is inconsistent. R and R stands only for relation of not less than and not greater than, respectively, but cannot represent relations strictly greater than and strictly less than. We extend the domination relation to strict preference relation:
R C > = { ( x , y ) U × U   : f ( x , a ) > f ( y , a ) ,   a C }
R C < = { ( x , y ) U × U   : f ( x , a ) < f ( y , a ) ,   a C }
R D > = { ( x , y ) U × U   : f ( x , D ) > f ( y , D ) }
R D < = { ( x , y ) U × U   : f ( x , D ) < f ( y , D ) }
If ( x , y ) R C > , we say object x is better than object y with respect to C . If ( x , y ) R C , we say object x is less than object x with respect to C .
If ( x , y ) R D > , we say the decision of object x is better than that of y. If ( x , y ) R D , we say the decision of object x is less than that of y.
In this work, R and R > are all named as upward domination relations, and R and R < are all named as downward domination relations. The preference information granulation can be represented as:
[ x ] C = { y U   :   ( y , x ) R C }
[ x ] C = { y U   :   ( y , x ) R C }
[ x ] C > = { y U   :   ( y , x ) R C > }
[ x ] C < = { y U   :   ( y , x ) R C < }
Let P S = < U , C D , f > be an ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . x U , a C , if y [ x ] a > , f ( y , D ) f ( x , D ) holds, we say object x is upward preference consistent in terms of attribute a. Otherwise, x is upward preference inconsistent in terms of a. Similarly, if y [ x ] a < , f ( y , D ) f ( x , D ) holds, we say object x is downward preference consistent in terms of attribute a, whereas, x is downward preference inconsistent in terms of a.
A lot of research has been done regarding preference consistent decision systems. In a real environment, however, we often need to face preference inconsistent cases. Preference inconsistence leads to preference analysis uncertainty. Those achievements for preference consistent system cannot be used to analyze preference inconsistent cases. Shannon entropy is a very good measure of uncertainty. Therefore, we will study the entropy based on preference inconsistence to deal with preference inconsistent issues.

3. Preference Inconsistence Based Entropy

In this section, two types of preference inconsistence entropy (PIE)—upward preference inconsistence entropy (UPIE) and downward preference inconsistence entropy (DPIE)—are introduced to measure the preference uncertainty. Some relative concepts and theorems will be discussed in detail.
Definition 1. 
Given a preference inconsistent ordinal decision system P S = < U ,   C D , f > . x U , a C , the preference inconsistent set (PIS) of x in terms of a can be denoted by:
  • upward preference inconsistent set (UPIS):
    u i c o n S ( x , a ) = { y U | y [ x ] a >     f ( y , D ) < f ( x , D ) }
  • downward preference inconsistent set (DPIS):
    d i c o n S ( x , a ) = { y U | y [ x ] a <     f ( y , D ) > f ( x , D ) }
For a set of attributes B C , we have:
u i c o n S ( x , B ) = { y U | y [ x ] b >     f ( y , D ) < f ( x , D ) , b B }
d i c o n S ( x , B ) = { y U | y [ x ] b <     f ( y , D ) > f ( x , D ) ,   b B }
For a set of attributes B C , object x is upward preference consistent in terms of B if b B , object x is upward preference consistent in terms of b, whereas, x is upward preference inconsistent in terms of B unless x is upward preference inconsistent in terms of each attribute in B. As to downward preference, a similar case occurs. If b B , object x is downward preference consistent in terms of b, x is downward preference consistent in terms of B. Otherwise, x is downward preference inconsistent in terms of B.
Theorem 1. 
Let P S = < U ,   C D , f > be a preference inconsistent ordinal decision system, U is nonempty finite set of objects, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } ,   d 1 d 2 d N . Given two attributes a 1 , a 2 C , we have:
u i c o n S ( x , a b ) = u i c o n S ( x , a ) u i c o n S ( x , b )
d i c o n S ( x , a b ) = d i c o n S ( x , a ) d i c o n S ( x , b )
Proof:
u i c o n S ( x , a b ) = { y U | y [ x ] a >     y [ x ] b >     f ( y , D ) < f ( x , D ) } = { y U | y [ x ] a >     y   f ( y , D ) < f ( x , D ) } = { y U | y [ x ] b >     y   f ( y , D ) < f ( x , D ) } = u i c o n S ( x , a ) u i c o n S ( x , b )
d i c o n S ( x , a b ) = { y U | y [ x ] a <     y [ x ] b <     f ( y , D ) > f ( x , D ) } = { y U | y [ x ] a <     y   f ( y , D ) > f ( x , D ) } = { y U | y [ x ] b <     y   f ( y , D ) > f ( x , D ) } = d i c o n S ( x , a ) d i c o n S ( x , b )
We can easily expand this theorem to a multiple attributes case:
u i c o n S ( x , i = 1 n a i   ) = i = 1 n u i c o n S ( x , a i )  
d i c o n S ( x , i = 1 n a i   ) = i = 1 n d i c o n S ( x , a i )  
Definition 2. 
Let P S = < U , C D , f > be a reference inconsistent ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . Given A C , the preference inconsistent degree (PID) can be defined as:
  • upward preference inconsistent degree (UPID):
    u p i c d ( x , A ) = | u i c o n S ( x , A ) | | U |
  • downward preference inconsistent degree (DPID):
    d p i c d ( x , A ) = | d i c o n S ( x , A ) | | U |
    where u i c o n S   and   d i c o n S are the set of upward and downward preference inconsistent set, respectively.
Theorem 2. 
Let P S = < U ,   C D , f > be a preference inconsistent ordinal decision system, U is a nonempty finite set of objects, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } ,   d 1 d 2 d N . Given two attributes a 1 , a 2 C , we have:
u p i c d ( x , a 1 ) u i c o n S ( x , a 1 a 2 )
u p i c d ( x , a 2 ) u i c o n S ( x , a 1 a 2 )
d p i c d ( x , a 1 ) d i c o n S ( x , a 1 a 2 )
d p i c d ( x , a 2 ) d i c o n S ( x , a 1 a 2 )
Proof. 
It can be proven easily by Definitions 2–4 and Theorem 1. □
The preference inconsistent information granule can be represented as follows:
[ A ] = u p i c d ( x 1 , A ) x 1   +   u p i c d ( x 2 , A ) x 2   +   u p i c d ( x 3 , A ) x 3   + +   u p i c d ( x 4 , A ) x 4
[ A ] = d p i c d ( x 1 , A ) x 1   +   d p i c d ( x 2 , A ) x 2   +   d p i c d ( x 3 , A ) x 3   + +   d p i c d ( x 4 , A ) x n
Definition 3. 
Let P S = < U , C D , f > be a preference inconsistent ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . The preference inconsistent entropy(PIE) can be defined as:
  • upward preference inconsistent entropy (UPIE):
    U H ( A ) = 1 | U | x i U ,   u p i c d ( x i , A ) 0 log 2 ( 1 u p i c d ( x i , A ) )
  • downward preference inconsistent entropy (DPIE):
    D H ( A ) = 1 | U | x i U , d p i c d ( x i , A ) 0 log 2 ( 1 d p i c d ( x i , A ) )  
    where u p i c d and d p i c d are the upward preference inconsistence degree and downward preference inconsistence degree, respectively. As for preference consistent ordinal decision system, the preference decisions are certain, and U H ( A ) and D H ( A ) are all equal to 0, which is consistent with Shannon entropy.
Definition 4. 
Let P S = < U , C D , f > be a preference inconsistent ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . Given two attributes A 1 , A 2 C , the preference inconsistent joint entropy (PIJE) is expressed as:
  • upward preference inconsistent joint entropy (UPIJE):
    U H ( A 1 A 2 ) = 1 | U | x i U , u p i c d ( x i , A 1 A 2 ) 0 log 2 ( 1 u p i c d ( x i , A 1 A 2 ) )  
  • downward preference inconsistent joint entropy (DPIJE):
    D H ( A 1 A 2 ) = 1 | U | x i U , u p i c d ( x i , A 1 A 2 ) 0 log 2 ( 1 d p i c d ( x i , A 1 A 2 ) )  
Theorem 3. 
Let P S = < U ,   C D , f > be a preference inconsistent ordinal decision system, U is a nonempty finite set of objects, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } ,   d 1 d 2 d N . Given two attributes A 1 , A 2 C , we have:
U H ( A 1 ) U H ( A 1 A 2 )
U H ( A 2 ) U H ( A 1 A 2 )
D H ( A 1 ) D H ( A 1 A 2 )
D H ( A 2 ) D H ( A 1 A 2 )
Proof. 
It can be proven by Definition 2–4 and Theorem 2. ☐
Theorems 1–3 show that in a preference inconsistence set with less elements, smaller preference inconsistence degree and smaller preference inconsistence entropy can be obtain by adding conditional attributes. Obviously, a conditional attribute is redundant if it cannot decrease the preference inconsistence entropy.
Definition 5. 
Let P S = < U , C D , f > be a preference inconsistent ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . Given two attributes A 1 , A 2 C , the preference inconsistent conditional entropy of A 1 to A 2 is expressed as:
  • upward preference inconsistent conditional entropy:
    U H ( A 1 | A 2 ) = U H ( A 1 A 2 ) U H ( A 2 )
  • downward preference inconsistent conditional entropy:
    D H ( A 1 | A 2 ) = D H ( A 1 A 2 ) D H ( A 2 )
The conditional entropy reflects the uncertainty of A 1 if A 2 is given. Here, we give an instance to illustrate the information entropy model about preference inconsistence ordinal decision system.
Example 1. 
Assume that we have 10 candidates x1, x2, …, x10 with two conditional attributes A1, A2, as shown in Table 1. D = {1, 2, 3} is the rank set of candidates.
We can take the preference inconsistent information granules and preference inconsistence entropy as follows:
  • upward preference inconsistent information granules:
    [ A 1 ] = 0 x 1   +   0 x 2   +   0 x 3   + 0.1 x 4   + 0.1 x 5   + 0.1 x 6   + 0 x 7   + 0.1 x 8   + 0 x 9   + 0 x 10
    [ A 2 ] = 0 x 1   +   0 x 2   +   0 x 3   + 0.1 x 4   + 0 x 5   + 0 x 6   + 0 x 7   + 0.1 x 8   + 0 x 9   + 0 x 10  
    [ A 1 A 2 ] = 0 x 1   +   0 x 2   +   0 x 3   + 0 x 4   + 0 x 5   + 0 x 6   + 0 x 7   + 0 x 8   + 0 x 9   + 0 x 10
  • downward preference inconsistent information granules:
    [ A 1 ] = 0 x 1   +   0 x 2   +   0.3 x 3   + 0 x 4   + 0 x 5   + 0 x 6   + 0.1 x 7   + 0 x 8   + 0 x 9   + 0 x 10  
    [ A 2 ] = 0 x 1   +   0.1 x 2   +   0 x 3   + 0 x 4   + 0 x 5   + 0.1 x 6   + 0 x 7   + 0 x 8   + 0 x 9   + 0 x 10  
    [ A 1 A 2 ] = 0 x 1   +   0 x 2   +   0 x 3   + 0 x 4   + 0 x 5   + 0 x 6   + 0 x 7   + 0 x 8   + 0 x 9   + 0 x 10  
  • upward reference inconsistent information entropy:
    U H ( A 1 ) = 1 10 × ( log 2 0.1 + log 2 0.1 + log 2 0.1 + log 2 0.1 ) = 1.3288  
    U H ( A 2 ) = 1 10 × ( log 2 0.1 + log 2 0.1 ) = 0.6644
    U H ( A 1 A 2 ) = 0
  • downward preference inconsistent information entropy:
    D H ( A 1 ) = 1 10 × ( log 2 0.3 + log 2 0.1 ) = 0.5059
    D H ( A 2 ) = 1 10 × ( log 2 0.1 + log 2 0.1 ) = 0.6644
    D H ( A 1 A 2 ) = 0
  • upward reference inconsistent conditional entropy:
    U H ( A 1 | A 2 ) = U H ( A 1 A 2 ) U H ( A 2 ) = 0.6644
    U H ( A 2 | A 1 ) = U H ( A 1 A 2 ) U H ( A 1 ) = 1.3288
  • downward reference inconsistent conditional entropy:
    D H ( A 1 | A 2 ) = D H ( A 1 A 2 ) D H ( A 2 ) = 0.6644
    D H ( A 2 | A 1 ) = D H ( A 1 A 2 ) D H ( A 1 ) = 0.5059
As is shown in the example above, upward reference inconsistence information entropy and downward preference inconsistence information entropy are all greater than or equal to 0, which is consistent with Theorem 2. As for preference inconsistence conditional entropy, however, it is less than or equal to 0. This can be explained easily by Theorem 3 and Definition 5. If the conditional entropy is equal to 0, attribute A 1 can be eliminated while not increasing the preference inconsistency degree. In next section, we will illustrate the usefulness of preference inconsistence entropy through selecting features and condensing samples.

4. The Application of PIE

4.1. Feature Selection

One of the most important applications of preference analysis is to calculate the reduction of attributes. We are usually required to analyze the relevance between criteria and decisions, and then select those features which are most significant for decision-making and eliminate those redundant or irrelevant features. Plenty of perfect models and algorithms are proposed. However, almost all current work which focuses on the preference consistent cases cannot be used in inconsistent preference cases. Here, we show a feature selection technique based on preference inconsistence entropy while not decreasing the preference consistent degree.
Definition 7. 
Let P S = < U , C D , f > be a preference inconsistent ordinal decision system, C is the set of conditional attributes and D is the decision, D = { d 1 , d 2 , , d N } , d 1 d 2 d N . Given two attributes A 1 , A 2 C , the relative significance of attribute subset A 1   to   A 2 can be defined as:
  • upward relative significance:
    u S i g ( A 1 , A 2 ) = U H ( A 1 | A 2 ) = U H ( A 2 ) U H ( A 1 A 2 )
  • downward relative significance:
    d S i g ( A 1 , A 2 ) = D H ( A 1 | A 2 ) = D H ( A 2 ) D H ( A 1 A 2 )
The greater the inconsistence entropy is, the smaller is the relevance of the decision to an attribute, that is, the attribute with greater inconsistence entropy has less significance. u S i g ( A 1 , A 2 ) is used to evaluate the relative significance of attribute subset A 1 to attribute subset A 2 for upward preference. Similar to u S i g ( A 1 , A 2 ) , d S i g ( A 1 , A 2 ) is the evaluation measure of attribute significance for downward preference. Equations (60) and (61) can be used to find those important features which are indispensable to hold the preference consistence of conditional attributes to a decision. As a matter of fact, it is impractical to get the optimal feature subset from 2 n 1 candidates through exhaustive search, where n is the number of features. The greedy search guided by some heuristics is usually more efficient than the plain brute-force exhaustive search [25]. A forward search algorithm (Algorithm 1) and a backward algorithm (Algorithm 2) for feature selection are expressed as follows, respectively. Here, UPIE and DPIE are denoted as PIE uniformly, UH and DH are denoted as H uniformly, and uSig and dSig are denoted as Sig uniformly.
Algorithm 1 Forward feature selection (FFS) based on PIE
Input: the preference inconsistent ordinal decision system P S = < U , C D > ;
Output: selected feature subset s f .
1. foreach a C
2.   if f ( x , a ) A
3.      C = C { a }
4.   end if
5. end for
6. foreach a C
7.   compute H ( a )
8. end for
9. find the minimal H ( a ) and the corresponding attribute a
10.  s f a
11. while s f C
12.   for each a C s f
13.     compute s i g a = u S i g ( a , s f )
14.   end for
16.   find the maximal s i g a and the corresponding attribute a
17.   if s i g a > 0
18.      s f s f a
19.   else
20.     exit while
21.   end if
22. end while
23. return s f
Algorithm 2 Backward feature selection (BFS) based on PIE
Input: the preference inconsistent ordinal decision system P S = < U , C D > ;
Output: selected feature subset s f .
1. foreach a C
2.   if f ( x , a ) A
3.      C = C { a }
4.    end if
5. end for
6.  s f C
7. for each a s f
8.   compute s i g a = S i g ( a , s f { a } )
9.   if s i g a = 0
10.      s f s f { a }
11.   end if
12. end for
13. return s f
The time complexities of algorithm FFS and BFS are O ( m 2 n l o g n ) and O ( m n ) , respectively, where m is the number of features and n is the number of samples, respectively. Replacing P I E with U P I E , we can get the algorithms for upward preference. Replacing P I E with D P I E , we can get the algorithms for downward preference.
If x U , f ( x , a ) A , we can get U H ( a ) = 0 ,   D H ( a ) = 0   and u S i g ( a ) ,   d S i g ( a ) . Obviously, however, feature a has minimal inconsistence entropy and maximal significance but has no effect on the decision. For the forward algorithm, if features like a exist, the selected feature subset contains only one element like a. For the backward algorithm, we could get a similar result, so before starting the algorithms, we should eliminate all the features like a.

4.2. Sample Condensation

If an ordinal decision system is upward preference inconsistent, there exist some objects with greater attribute values and worse decision. Similarly, if an ordinal decision is downward preference inconsistent, there exist some samples with less attribute values and better decision. These samples are named as exceptional samples. Obviously, in this view, the inconsistent decision system would become preference consistent once the exceptional samples are all eliminated. A backward search algorithm (Algorithm 3) is written as follows: like feature selection algorithms, UPIE and DPIE are denoted as PIE uniformly, upicd and dpicd are denoted as picd, and UH and DH are denoted as H uniformly. The time complexity of this algorithm is O(mn), where m is the number of features and n is the number of samples.
Algorithm 3 Sample Condensation based on PIE (BFS)
Input: the preference inconsistent ordinal decision system P S = < U , C D > ;
Output: the preference consistent sample subset.
1. for each a C
2.    s a ( a ) U
3.   compute the preference inconsistence entropy H ( s a ( a ) , a )
4.   while H ( s a ( a ) ,     a ) > 0
5.     for each   x s a ( a )
6.       Compute p i c d x = p i c d ( x , a )
7.     end for
8.     find the maximum p i c d x and the corresponding sample x
9.       s a ( a ) s a ( a ) { x }
10.      compute the preference inconsistence entropy H ( s a ( a ) , a )
11.  end while
12. end for
13.  s a s = ;
14. for each a C
15.   s a s s a ( a ) s a s ;
16. end for
17. return s a s

5. Experimental Analysis

In this section, we present experimental results obtained by using two real data sets named Pasture Production and Squash Harvest downloaded from WEKA (http://www.cs.waikato.ac.nz/ml/weka/). Three experiments are done for every data set. In the first experiment, preference inconsistence entropy and significance of single attribute are computed. In the second experiment, some features are selected based on the algorithms proposed in this work. Then we condense the data set and eliminate some samples causing inconsistence in the last experiment.
The Pasture Production data set was collected by Dave Barker from areas of grazed North Island hill country with recorded different management history during the period from 1793 to 1994. The objective was to analyze the influence of biophysical factors on pasture production. The Production of Pasture data set is composed of 36 tuples associated with 36 paddocks. Each tuple has 22 attributes, of which 19 are vegetation, and the others are soil chemical, physical and biological, and soil water variables, respectively. All the samples are partitioned into three classes, LO, MED and HI, representing low, medium and high production, respectively. Table 2 is the description of the features. For use in this work, we replaced these symbol values by integer 1, 2 and 3, respectively. Accordingly, the enumerated value (LL, LN, HN and HH) of attribute fertilizer are replaced by 1, 2, 3 and 4. Figure 1 shows the result of Experiment 1, Table 3 presents the selected features in the experiment, and the sample condensation result of Experiment 3 is listed in Table 4.
Figure 1a shows the inconsistence entropy comparison of single features based on upward and downward preference. Figure 1b shows the comparison of signification of single features based on upward and downward preference. The relation of inconsistence entropy and attribute significance is presented in Figure 1c.
The inconsistence entropy of every feature of the decision is different, which shows that those features are not of the same importance in predicting pasture production. The inconsistence entropy monotonically decreases with increasing feature significance. The greater the inconsistence entropy is, the less the feature helps predict the pasture production, but in some cases, exceptions also exist.
We arrange these features in descending order by upward significance {4, 16, 22, 10, 9, 5, 7, 17, 1, 6, 11, 13, 3, 20, 8, 15, 2, 19, 21, 12, 18, 14}, and arrange the features in ascending order by upward inconsistence entropy {14, 4, 16, 22, 10, 9, 5, 7, 17, 1, 6, 11, 13, 3, 20, 8, 15, 2, 19, 21, 12, 18}. Except for feature 14, those attributes at the head of the first series have smaller inconsistence entropy and greater significance in predicting pasture production, such as 4, 16, 22, and so on. Feature 14 is an exception. The inconsistence entropy of feature 14 is the smallest value 0, but its significance also has the smallest value of 0. We can find that all distances have the same value 0 in feature 14, so its inconsistence entropy is 0. But at the same time, feature 14 is not different for all samples and cannot reflect any preference, so the significance should be 0. This is consistent with our common knowledge. The forward feature selection algorithm output the same features for upward preference and downward preference. As for the backward method, a similar case occurs.
In addition, from Figure 1, we can see that there is little difference between upward inconsistence entropy and downward inconsistence entropy. A similar case occurs for the significance of a single feature. In order to test how much the order of features impacts the output, we generated four different data sets by arranging the features in different order and executing the feature selection algorithms in all the data sets. All the features are labeled as 1, 2, ..., 22. The original data set is the first one. By arranging those features in a descending label order, we get the second data set. We compute the significance of every feature and get the third and the fourth data sets by arranging the features in the order of ascending and descending attribute significance, respectively. The forward algorithm is performed on data sets 1, 2 and 4, and the backward algorithm is executed on data sets 1, 2 and 3. Table 3 shows the experiment results.
For the forward algorithm, features 4, 11 and 1 are selected from data set 1, features 4, 11 and 1 are selected from data set 2, and features 4, 11 and 22 are selected from data set 4. Features 4 and 11 are included in all feature subset. The top two of three selected features are the same and only the last is different for all data sets. As for the backward algorithm, however, we have an entirely different situation. Features 17, 18, 20 and 22 are selected from data set 1, features 1, 3, 4 and 5 are selected from data set 2, and features 4, 10 and 16 are selected from data set 3. The selected feature subsets are totally different. Obviously, we can conclude that the forward algorithm is more stable and the backward algorithm is affected more by the order of features.
In the forward feature selection algorithm, the relative significance is computed in turn in the order of increasing feature label. As for the backward method, we compute the relative significance of features in the same turn. However, there exists an essential difference between the algorithms. For the forward algorithm, after the significances of all features are computed, the feature with maximal relative significance is selected, but for the backward method, if the relative significance is 0, the feature is immediately deleted from the features set. From Table 3, we can find that we can get similar feature subsets if we sort the data set by attribute significance. The backward algorithm selects features 4, 16 and 10 from data set 3 and the forward algorithm selects features 4, 16 and 11. We can see that features 4 and 16 appear in both the subsets. This shows that these features are important in term of different selection methods.
Table 4 shows the sample condensation results based on selected feature subsets in previous experiments. Figure 2 illustrates the distribution, histogram and pie of selected samples for upward preference. Figure 3 illustrates the distribution, histogram and pie of selected samples for downward preference.
The Squash Harvest data set, collected by Winna Harvey Crop & Food Research (Christchurch, New Zealand), aims to determine which pre-harvest variables are important to good tasting squash after different periods of storage time so as to pinpoint the best time to give the best quality at the marketplace. This is determined by whether a measure of acceptability found by classifying each squash as either unacceptable, acceptable or excellent. The name and the corresponding descriptions of those features are described in Table 5. There are 52 instances with 24 attributes. For use in this work, we replaced these class symbol values with integer 1, 2 and 3, respectively. We do the same tests on this data set.
We compute the inconsistence entropy and attribute significance of single features for upward preference and downward preference, as shown in Figure 4. Features 1, 2, 6, 11, 16 and 24 have greater single attribute significance, indicating that, as far as single features are concerned, site, number of days after flowering, maturity of fruit at harvest, the heat accumulation from flowering to harvesting, total number of glucose, fructose and sucrose and the amount of heat input before flowering are very important for good taste.
We also extract the preference consistence feature subset. Similar to Pasture Production, we generate four data sets from Squash Harvest. All the features are labeled as 1, 2, ..., 24. The original data set is the first. By arranging those features in the order of descending label, we get the second data set.
We compute the significance of every feature and get the third and the fourth data sets by arranging the features in the order of ascending and descending attribute significance, respectively. The forward algorithm is performed on data sets 1, 2 and 4, and the backward algorithm is executed on data sets 1, 2 and 3. Table 6 shows the experiment results. Finally, we condense the samples with selected features. The results are shown in Table 7. Figure 5 and Figure 6 show the distribution and histograms of selected samples.
For the forward algorithm, we get the same feature subset 1, 2 and 14. The attribute significance of features 1 is 0.98. Feature 2 has an attribute significance of 0.63 and 0.55 for upward preference and downward preference, respectively. Features 1 and 2 are important for good taste, but the significance of feature 14 is as low at 0.26 and 0.22 for upward preference and downward preference, respectively. This indicates that, more glucose can give better taste at harvest if the plant site is suitable and the number of days after flowering is enough.

6. Conclusions

In multiple attribute ordinal decision, the derived decisions may be inconsistent and uncertain since the available information is usually obtained from different evaluation criteria or experts. Shannon entropy is a suitable measurement of uncertainty. In this research, Shannon entropy is extended to preference inconsistence entropy for preference inconsistent ordinal decision systems. Two applications of preference inconsistence entropy, feature selection and sample condensation, are addressed and some algorithms are developed. The experimental result shows that preference inconsistence entropy is a suitable measurement of uncertainty of preference decision, and the proposed algorithms are effective solutions for preference analysis. Classification is another interesting issue for preference analysis, and further investigations on classification methods and algorithms based on preference inconsistence entropy are worthwhile.

Acknowledgments

The reported research was supported in part by an operating grant from the Natural Science Foundation of the Education Department of Sichuan Province (No. 12ZA178), Key Technology Support Program of Sichuan Province (Grant No. 2015GZ0102), and Foundation of Visual Computing and Virtual Reality Key Laboratory of Sichuan Province (Grant No. KJ201406). The authors would like to thank the anonymous reviewers for their useful comments. The authors would also like to thank the editor of the journal Entropy for his useful guidance and suggestions that improved the paper.

Author Contributions

Kun She conceived and designed the experiments; Pengyuan Wei performed the experiments and Wei Pan wrote the paper. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lotfi, F.H.; Fallahnejad, R. Imprecise Shannon’s entropy and multi attribute decision making. Entropy 2010, 12, 53–62. [Google Scholar] [CrossRef]
  2. Saaty, T.L. The Analytic Hierarchy Process; McGraw-Hill: New York, NY, USA, 1980. [Google Scholar]
  3. Chu, A.T.W.; Kalaba, R.E.; Spingarn, K. A comparison of two methods for determining the weights of belonging to fuzzy sets. J. Optim. Theory Appl. 1979, 27, 531–538. [Google Scholar] [CrossRef]
  4. Hwang, C.-L.; Lin, M.-J. Group Decision Making under Multiple Criteria: Methods and Applications; Springer: Berlin/Heidelberg, Germany, 1987. [Google Scholar]
  5. Choo, E.U.; Wedley, W.C. Optimal criterion weights in repetitive multicriteria decision-making. J. Oper. Res. Soc. 1985, 36, 983–992. [Google Scholar] [CrossRef]
  6. Fan, Z.P. Complicated Multiple Attribute Decision Making: Theory and Applications. Ph.D. Thesis, Northeastern University, Shenyang, China, 1996. [Google Scholar]
  7. Moshkovich, H.M.; Mechitov, A.I.; Olson, D.L. Rule induction in data mining: Effect of ordinal scales. Expert Syst. Appl. 2002, 22, 303–311. [Google Scholar] [CrossRef]
  8. Harker, P.T. Alternative modes of questioning in the analytic hierarchy process. Math. Model. 1987, 9, 353–360. [Google Scholar] [CrossRef]
  9. Saaty, T.L.; Vargas, L.G. Uncertainty and rank order in the analytic hierarchy process. Eur. J. Oper. Res. 1987, 32, 107–117. [Google Scholar] [CrossRef]
  10. Xu, Z.S. Method for group decision making with various types of incomplete judgment matrices. Control Decis. 2006, 21, 28–33. [Google Scholar]
  11. Van Laarhoven, P.J.M.; Pedrycz, W. A fuzzy extension of Saaty’s priority theory. Fuzzy Sets Syst. 1983, 11, 199–227. [Google Scholar] [CrossRef]
  12. Luce, R.D.; Suppes, P. Preferences Utility and Subject Probability. In Handbook of Mathematical Psychology; Luce, R.D., Bush, R.R., Galanter, E.H., Eds.; Wiley: New York, NY, USA, 1965; Volume 3, pp. 249–410. [Google Scholar]
  13. Alonso, S.; Chiclana, F.; Herrera, F.; Herrera-Viedma, E. A Learning Procedure to Estimate Missing Values in Fuzzy Preference Relations Based on Additive Consistency. In Modeling Decisions for Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2004; pp. 227–238. [Google Scholar]
  14. Xu, Z.S. A practical method for priority of interval number complementary judgment matrix. Oper. Res. Manag. Sci. 2002, 10, 16–19. [Google Scholar]
  15. Herrera, F.; Herrera-Viedma, E.; Verdegay, J.L. A model of consensus in group decision making under linguistic assessments. Fuzzy Sets Syst. 1996, 78, 73–87. [Google Scholar] [CrossRef]
  16. Herrera, F.; Herrera-Viedma, E.; Verdegay, J.L. Direct approach processes in group decision making using linguistic OWA operators. Fuzzy Sets Syst. 1996, 79, 175–190. [Google Scholar] [CrossRef]
  17. Greco, S.; Matarazzo, B.; Slowinski, R. Rough approximation of a preference relation by dominance relations. Eur. J. Oper. Res. 1999, 117, 63–83. [Google Scholar] [CrossRef]
  18. Greco, S.; Matarazzo, B.; Slowinski, R. Rough approximation by dominance relations. Int. J. Intell. Syst. 2002, 17, 153–171. [Google Scholar] [CrossRef]
  19. Hu, Q.H.; Yu, D.R.; Guo, M.Z. Fuzzy preference based rough sets. Inf. Sci. 2010, 180, 2003–2022. [Google Scholar] [CrossRef]
  20. Abbas, A.E. Entropy methods for adaptive utility elicitation. IEEE Trans. Syst. Man Cybern. 2004, 34, 169–178. [Google Scholar] [CrossRef]
  21. Yang, J.; Qiu, W. A measure of risk and a decision-making model based on expected utility and entropy. Eur. J. Oper. Res. 2005, 164, 792–799. [Google Scholar] [CrossRef]
  22. Abbas, A.E. Maximum Entropy Utility. Oper. Res. 2006, 54, 277–290. [Google Scholar] [CrossRef]
  23. Abbas, A. An Entropy Approach for Utility Assignment in Decision Analysis. In Proceedings of the 22nd International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Moscow, Russian Federation, 3–9 August 2002; pp. 328–338.
  24. Hu, Q.H.; Yu, D.R. Neighborhood Entropy. In Proceedings of the 2009 International Conference on Machine Learning and Cybernetics, Baoding, China, 12–15 July 2009; Volume 3, pp. 1776–1782.
  25. Jaynes, E.T. Information theory and statistical mechanics II. Phys. Rev. 1957, 108, 171–190. [Google Scholar] [CrossRef]
  26. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  27. Kullback, S. Information Theory and Statistics; Wiley: New York, NY, USA, 1959. [Google Scholar]
  28. Hu, Q.H.; Che, X.J.; Zhang, L.D.; Guo, M.Z.; Yu, D.R. Rank entropy based decision trees for monotonic classification. IEEE Trans. Knowl. Data Eng. 2012, 24, 2052–2064. [Google Scholar] [CrossRef]
  29. Jaynes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  30. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
  31. Zeng, K.; Kun, S.; Niu, X.Z. Multi-Granulation Entropy and Its Application. Entropy 2013, 15, 2288–2302. [Google Scholar] [CrossRef]
Figure 1. Preference inconsistence entropy and attribute significance of single features (Pasture production). (a) Inconsistent entropy of single features; (b) Significance of single features; (c) Relation of inconsistence entropy and attribute significance.
Figure 1. Preference inconsistence entropy and attribute significance of single features (Pasture production). (a) Inconsistent entropy of single features; (b) Significance of single features; (c) Relation of inconsistence entropy and attribute significance.
Entropy 18 00096 g001
Figure 2. Distribution, histogram and pie of selected samples for upward preference (Pasture Production). (a) Distribution of selected samples; (b) Histogram of selected samples; (c) Pie of selected samples.
Figure 2. Distribution, histogram and pie of selected samples for upward preference (Pasture Production). (a) Distribution of selected samples; (b) Histogram of selected samples; (c) Pie of selected samples.
Entropy 18 00096 g002aEntropy 18 00096 g002b
Figure 3. Distribution, histogram and pie of selected samples for downward preference (Pasture Production). (a) Distribution of selected samples; (b) Histogram of selected samples; (c) Pie of selected samples.
Figure 3. Distribution, histogram and pie of selected samples for downward preference (Pasture Production). (a) Distribution of selected samples; (b) Histogram of selected samples; (c) Pie of selected samples.
Entropy 18 00096 g003
Figure 4. Preference inconsistence entropy and attribute significance of single features (Squash Harvest). (a) Inconsistence entropy of single features; (b) Significance of single features; (c) Relation of attribute significance and inconsistence entropy.
Figure 4. Preference inconsistence entropy and attribute significance of single features (Squash Harvest). (a) Inconsistence entropy of single features; (b) Significance of single features; (c) Relation of attribute significance and inconsistence entropy.
Entropy 18 00096 g004
Figure 5. Distribution, histogram and pie of selected samples for upward preference (Squash Harvest). (a) Distribution of selected samples; (b) Histogram of selected samples.
Figure 5. Distribution, histogram and pie of selected samples for upward preference (Squash Harvest). (a) Distribution of selected samples; (b) Histogram of selected samples.
Entropy 18 00096 g005aEntropy 18 00096 g005b
Figure 6. Distribution and histogram of selected samples for downward preference (Squash Harvest). (a) Distribution of selected samples; (b) Histogram of selected samples.
Figure 6. Distribution and histogram of selected samples for downward preference (Squash Harvest). (a) Distribution of selected samples; (b) Histogram of selected samples.
Entropy 18 00096 g006
Table 1. Ordinal decision system DS1.
Table 1. Ordinal decision system DS1.
x1x2x3x4x5x6x7x8x9x10
A10.280.250.600.480.420.550.780.750.830.85
A20.280.310.420.470.510.660.550.650.800.91
D1112222333
Table 2. Features of Pasture Production.
Table 2. Features of Pasture Production.
SequenceAttributeDescription
1fertiliser{LL, LN, HN, HH}enumerated
2slopeSlope of the paddocknumeric
3aspect-dev-NWThe deviation from the north-westnumeric
4olsenP numeric
5minN numeric
6TS numeric
7Ca-MgCalcium magnesium rationnumeric
8LOMSoil lom (g/100g)numeric
9NFIX-meanA mean calculationnumeric
10Eworms-main-3Main 3 spp earth worms per g/m2numeric
11Eworms-No-speciesNumber of sppnumeric
12KUnSatmm/hnumeric
13OM numeric
14Air-Perm numeric
15Porosity numeric
16HFRG-pct-meanMean percentnumeric
17legume-yieldkgDM/hanumeric
18OSPP-pct-meanMean percentnumeric
19Jan-Mar-mean-TDR numeric
20Annual-Mean-RunoffIn mmnumeric
21root-surface-areaM2/m3numeric
22Leaf-PPpmnumeric
ClassPasture prod-class{Low, Median, High}enumerated
Table 3. Selected features for Pasture Production (original data set).
Table 3. Selected features for Pasture Production (original data set).
Data SetAlgorithmPreferenceSelected Features SubsetSignificance of Feature SubsetInconsistence Entropy of Feature Subset
Data set 1Forward Feature Selectionupward4, 11, 10.97, 0.25, 0.290
downward4, 11, 10.97, 0.24, 0.310
Backward Feature Selectionupward17, 18, 20, 220.4, 0.10, 0.19, 0.970
downward17, 18, 20, 220.48, 0.10, 0.17, 0.970
Data set 2Forward Feature Selectionupward4, 11, 220.36, 0.27, 0.970
downward4, 11, 220.32, 0.26, 0.970
Backward Feature Selectionupward5, 4, 3, 10.10, 0.15, 0.19, 0.970
downward5, 4, 3, 10.10, 0.13, 0.17, 0.970
Data set 3Backward Feature Selectionupward10, 16, 40.55, 0.92, 0.970
downward10, 16, 40.52, 0.92, 0.970
Data set 4Forward Feature Selectionupward4, 11, 160.97, 0.25, 0.920
downward4, 11, 160.97, 0.24, 0.920
Table 4. Selected samples (Pasture Production).
Table 4. Selected samples (Pasture Production).
Features SubsetUpward PreferenceDownward Preference
All features2,4,7,8,9,10,11,12,13,14,15,16,
18,19,20,21,22,26,27,28,33,35
1,2,3,4,5,6,7,8,12,17,22,23,
24,25,28,29,30,31,32,34,35,36
4,11,12,9,10,11,12,13,14,15,16,18,19,20,
21,22,26,27,28,29,30,31,32,33,34,35,36
1,2,3,4,5,6,7,8,23,24,25,
28,29,30,31,32,34,35,36
4,11,222,4,7,8,9,10,11,12,13,14,15,16,
18,19,20,21,22,26,27,28,33,35
1,2,3,4,5,6,7,8,12,17,22,23,
24,25,28,29,30,31,32,34,35,36
17,18,20,221,2,4,7,8,9,10,11,12,13,14,15,
16,18,19,20,21,22,26,27,28,33,35
1,2,3,4,5,6,7,8,12,17,22,23,
24,25,28,29,30,31,32,34,35,36
1,3,4,55,9,10,11,12,13,14,15,16,18,
20,21,24,25,26,31,32,33,34,35,36
1,2,4,6,11,12,19,20,21,22,23,
24,25,28,29,30,31,32,34,35,36
10,16,42,3,4,8,9,10,11,12,13,14,15,16,
18,20,21,26,27,29,31,32,33,34,35,36
1,3,5,6,7,17,18,19,22,23,24,
25,28,29,30,31,32,34,35,36
4,11,162,5,6,7,9,10,11,12,13,14,15,16,18,20,
21,22,23,24,25,26,28,29,30,31,32,34,35,36
1,2,3,4,5,6,7,8,10,14,17,19,22,23,
24,25,27,28,29,30,31,32,33,34,35,36
Table 5. Features of Squash Harvest.
Table 5. Features of Squash Harvest.
SequenceAttributeDescription
1sitewhere fruit is locatedenumerated
2dafnumber of days after floweringenumerated
3fruitindividual number of the fruitenumerated
4weightweight of whole fruit in gramsnumeric
5storewtweight of fruit after storagenumeric
6penepenetrometer indicates maturity of fruit at harvestnumeric
7solids_%a test for dry matternumeric
8brixa refactometer measurement used to indicate sweetness or ripeness of the fruitnumeric
9a*the a* coordinate of the HunterLab L*, a*, b* notation of color measurementnumeric
10egddthe heat accumulation above a base of 8c from emergence of the plant to harvest of the fruitnumeric
11fgddthe heat accumulation above a base of 8c from flowering to harvestingnumeric
12ground spotthe number indicating color of skin where the fruit rested on the groundnumeric
13glucosemeasured in mg/100g of fresh weightnumeric
14fructosemeasured in mg/100g of fresh weightnumeric
15sucrosemeasured in mg/100g of fresh weightnumeric
16totalmeasured in mg/100g of fresh weightnumeric
17Glucose + fructosmeasured in mg/100g of fresh weightnumeric
18starchmeasured in mg/100g of fresh weightnumeric
19sweetnessthe mean of eight taste panel scoresnumeric
20flavourthe mean of eight taste panel scoresnumeric
21dry/moistthe mean of eight taste panel scoresnumeric
22fibrethe mean of eight taste panel scoresnumeric
23heat_input_emergthe amount of heat emergence after harvestnumeric
24heat_input_flowerthe amount of heat input before floweringnumeric
25acceptabilitythe acceptability of the fruitenumarated
Table 6. Selected features for Squash Harvest (original data set).
Table 6. Selected features for Squash Harvest (original data set).
Data SetAlgorithmPreferenceSelected Features SubsetSignificance of Feature SubsetInconsistent Entropy of Feature Subset
Original dataForward Feature Selectionupward1, 2, 140.98, 0.63, 0.260
downward1, 2, 140.98, 0.55, 0.220
Backward Feature Selectionupward15, 17, 18, 22, 240.66, 0.38, 0.35, 0.33, 0.980
downward15, 17, 18, 22, 240.66, 0.38, 0.35, 0.33, 0.980
Sorted by feature label descendingForward Feature Selectionupward1, 2, 140.98, 0.63, 0.260
downward1, 2, 140.98, 0.55, 0.220
Backward Feature Selectionupward9, 3, 2, 10.98, 0.32, 0.72, 0.940
downward9, 3, 2, 10.82, 0.30, 0.75, 0.980
Sorted by significance descendingForward Feature Selectionupward1, 2, 140.98, 0.63, 0.260
downward1, 2, 140.98, 0.55, 0.220
Sorted by significance ascendingBackward Feature Selectionupward8, 12, 2, 16, 10.49, 0.62, 0.63, 0.71, 0.980
downward8, 12, 2, 16, 10.37, 0.53, 0.55, 0.57, 0.980
Table 7. Selected samples (Squash Harvest).
Table 7. Selected samples (Squash Harvest).
Features subsetUpward PreferenceDownward Preference
All features1,5,8,10,11,12,13,14,15,16,21,23,
25,32,33,34,35,36,42,44,45,46,47,48
1,2,6,7,9,12,13,14,15,16,22,24,26,27,28,29,
30,31,33,34,35,36,37,38,39,40,49,50,51,52
1,2,141,3,4,15,33,34,35,36,42,44,46,48,521,2,6,7,9,12,13,14,15,16,22,24,
26,27,28,29,30,31,40,49,50,51,52
15,17,18,22,245,8,10,11,12,13,14,15,16,21,23,
25,32,33,34,35,36,42,44,45,46,47,48
1,2,6,7,9,12,13,14,15,16,22,24,26,27,28,
29,30,31,33,34,35,36,37,38,39,40,49,50,51,52
9,3,2,11,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,
17,18,19,20,21,23,25,32,33,34,35,36,
37,38,39,41,42,43,44,45,46,47,48,51
1,2,6,7,9,12,13,14,15,16,22,24,26,27,
28,29,30,31,33,34,35,36,37,38,39,40,
41,42,43,44,45,46,47,48,49,50,51,52
8,12,2,16,11,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,
17,18,19,20,21,23,25,29,32,33,34,35,
36,37,38,39,41,42,43,44,45,46,47,48
1,2,6,7,9,12,13,14,15,16,22,24,26,27,
28,29,30,31,33,34,35,36,37,38,39,40,
41,42,43,44,45,46,47,48,49,50,51,52

Share and Cite

MDPI and ACS Style

Pan, W.; She, K.; Wei, P. Preference Inconsistence-Based Entropy. Entropy 2016, 18, 96. https://doi.org/10.3390/e18030096

AMA Style

Pan W, She K, Wei P. Preference Inconsistence-Based Entropy. Entropy. 2016; 18(3):96. https://doi.org/10.3390/e18030096

Chicago/Turabian Style

Pan, Wei, Kun She, and Pengyuan Wei. 2016. "Preference Inconsistence-Based Entropy" Entropy 18, no. 3: 96. https://doi.org/10.3390/e18030096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop