Next Article in Journal
An Exact Model-Based Method for Near-Field Sources Localization with Bistatic MIMO System
Next Article in Special Issue
Zero-Sum Matrix Game with Payoffs of Dempster-Shafer Belief Structures and Its Applications on Sensors
Previous Article in Journal
Posture Detection Based on Smart Cushion for Wheelchair Users
Previous Article in Special Issue
Testing a Firefly-Inspired Synchronization Algorithm in a Complex Wireless Sensor Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sensing Attribute Weights: A Novel Basic Belief Assignment Method

School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(4), 721; https://doi.org/10.3390/s17040721
Submission received: 20 December 2016 / Revised: 25 March 2017 / Accepted: 27 March 2017 / Published: 30 March 2017
(This article belongs to the Special Issue Soft Sensors and Intelligent Algorithms for Data Fusion)

Abstract

:
Dempster–Shafer evidence theory is widely used in many soft sensors data fusion systems on account of its good performance for handling the uncertainty information of soft sensors. However, how to determine basic belief assignment (BBA) is still an open issue. The existing methods to determine BBA do not consider the reliability of each attribute; at the same time, they cannot effectively determine BBA in the open world. In this paper, based on attribute weights, a novel method to determine BBA is proposed not only in the closed world, but also in the open world. The Gaussian model of each attribute is built using the training samples firstly. Second, the similarity between the test sample and the attribute model is measured based on the Gaussian membership functions. Then, the attribute weights are generated using the overlap degree among the classes. Finally, BBA is determined according to the sensed attribute weights. Several examples with small datasets show the validity of the proposed method.

1. Introduction

Data fusion technology of soft sensors [1,2,3] which integrates multi-source information to obtain a more objective result, is widely used in many application systems [4,5,6]. Some algorithms, such as fuzzy sets theory [7,8,9,10], Dempster–Shafer evidence theory (D-S evidence theory) [11,12], Z numbers [13,14], D numbers [15,16,17] and neural networks [18,19,20,21,22,23,24], are important tools in the data fusion of soft sensors. As one of them, D-S evidence theory gives a convenient mathematical framework to handle uncertainty information, so it is widely used in many fields [25,26,27]. However, determination of basic belief assignment (BBA) is the first and key step for applying D-S evidence theory. In general, there is no fixed model to obtain BBA. The method is usually designed according to the practical application.
Many scholars have investigated different methods to address the problem of obtaining BBA. Yager [28] proposed the D-S belief structure to determine BBA using the whole class of fuzzy measures. Zhu et al. [29] used fuzzy c-means to obtain BBA. Dubois et al. [30] proposed a probability–possibility transformation to gain BBA. Baudrit and Dubois [31] proposed practical representation methods for incomplete probabilistic information, based on formal links existing between possibility theory, imprecise probability and belief functions. Wang et al. [32] derived BBA function from the common multivariate data spaces. Masson and Denoeux [33] constructed a possibility distribution from a discrete empirical frequency distribution. Antoine et al. [34] constrained evidential clustering with instance-level constraints to determine BBA. Campos and Huete [35] determined BBA by considering the problem of assessing numerical values of possibility distributions.
Although various methods are proposed, these methods mainly focus on the incompleteness or the uncertainty of information itself, which does not simultaneously consider the importance and reliability of information sources. In the real world, information is not only incomplete and fuzzy, but also partly reliable. These are all the embodiment of the uncertainty of the information. We cannot only consider the ambiguity of information without considering the reliability of the information itself. For instance, in multiple attribute decision making, the importance of each attribute is different, so the weight of BBAs which are generated by different attributes is different. Many methods are proposed for measuring the weight of BBAs in D-S evidence theory [36,37]. The methods to determine these BBAs mainly depend on the support degree among BBAs, which are driven by data and relatively objective, however, these methods are questionable. The importance and reliability of these BBAs are not measured from the source of BBAs. Traditionally, the reliability of the information source is mainly evaluated based on the subjective judgement of the domain experts. However, the knowledge of experts is limited, which would lead to unreasonable results before some effective analysis is done. So far, the existing methods to determine BBA cannot address this issue, both in the closed world and in the open world. To address this issue, a method to determine BBA is proposed based on the attribute weights. The reliability of the information source is taken as a factor to correct the traditional BBA, and then BBA is gained which contains the reliability of the information source. The advantage of this method is that the reliability of BBA is measured by the attribute weights, which are driven by data and more reasonable.
Many studies have shown that unreasonable results may appear when using the classical D-S evidence theory in the open world. Based on this issue, Smets and Kennes [38] proposed the transferable belief model (TBM), and firstly introduced the concepts of the closed world and the open world, which has been widely applied in many fields [39,40,41]. The advantage of the TBM is that it holds belief at two levels, namely the credal level and the pignistic level. However, regardless of the credal level or pignistic level, the empty set is not assigned belief value in the BBA generation phase. Instead, the empty set is seen as an alarm or intermediary to show that conflict exists which may be caused by an incomplete framework of discernment in the process of evidential reasoning. As a result, the belief is assigned to the empty set in the process of evidential reasoning if the open world assumption is held, but not at the stage of BBA generation. Hence, Deng [42] presented the generalized evidence theory (GET), where generalized basic belief assignment (GBBA) and generalized combination rule (GCR) are proposed. GET assumes the decision-making environment is in an open world at the initial phase, so a generalized BBA that allows the beliefs to be assigned to the empty set is generated at the stage of data collection and modelling, which is more reasonable. GBBA will degenerate to classical BBA when m ( ) = 0 . Therefore, in this paper, a new method to determine BBA in the open world is presented based on GET. However, the GCR in GET still has some problems. For example, the method of obtaining m ( ) in the GCR is unreasonable. Based on this issue, a modified generalized combination rule (mGCR) in the framework of GET is presented [43]. The mGCR satisfies all properties of the GCR and is more reasonable than the GCR.
In this paper, based on the attribute weights, BBA is determined not only in the closed world, but also in the open world. First, the Gaussian distribution is used to model the Gaussian membership function of each attribute of each class. Second, the similarity between the test sample and the attribute model is measured based on these Gaussian membership functions. Then, the attribute weights are generated using the overlap degree among the classes. Given an attribute, the larger the overlap degree, the smaller the attribute weight; conversely, the smaller the overlap degree, the larger the attribute weight. Finally, BBA is determined after modifying the similarity by the sensed attribute weights. This BBA contains not only the information of each class, but also the reliability of each attribute, which is more objective and reasonable.
The rest of this paper is organized as follows. D-S evidence theory and some necessarily related concepts are briefly presented in Section 2. In Section 3, a novel method of determining BBA based on attribute weights is proposed. Several examples are illustrated in Section 4 to show the efficiency of this method. The conclusion is presented in Section 5.

2. Preliminaries

2.1. Dempster–Shafer Evidence Theory [11,12]

D-S evidence theory was introduced by Dempster. Then Shafer developed it by defining the belief function and plausibility function. It has the ability to deal with uncertain information without a prior probability. Thus, it is flexible and more effective than probability theory. Due to these advantages, it is widely applied in many fields [44,45,46].

2.1.1. Frame of Discernment (FoD) and Mass Function

In D-S evidence theory, the frame of discernment (FoD) is defined as Ω = { H 1 , H 2 , , H N } , which is composed of N exhaustive and exclusive hypotheses. Denote P ( Ω ) , the power set which is composed with 2 Ω propositions as:
P ( Ω ) = { , { H 1 } , { H 2 } , , { H N } , { H 1 H 2 } , { H 1 H 3 } , , Ω } ,
where the propositions { H 1 } , { H 2 } , , { H N } have only one element, so they are defined as the singleton subsect proposition; the propositions { H 1 H 2 } , { H 1 H 3 } , , Ω have more than one element, so they are defined as the compound subsect proposition.
A mass function m is a mapping from 2 Ω to [ 0 , 1 ] , formally defined as
m : 2 Ω [ 0 , 1 ] ,
which satisfies the following conditions:
A 2 Ω m ( A ) = 1 m ( ) = 0 .
The mass function m is also called BBA function. Any subset A of Ω such that m ( A ) > 0 is called a focal element.

2.1.2. Dempster’s Combination Rule

Suppose m 1 and m 2 are two mass functions in the same FoD Ω . Dempster’s combination rule, noted by m = m 1 m 2 , is defined as follows:
m ( A ) = B C = A m 1 ( B ) m 2 ( C ) 1 k , A 0 , A =
where
k = B C = m 1 ( B ) m 2 ( C ) .
Here, k is regarded as a measure of conflict between m 1 and m 2 . The larger the value of k, the more conflict between the evidence.

2.2. Generalized Evidence Theory

Generalized evidence theory (GET) is presented by Deng, which can handle the uncertain information in the open world [42].

2.2.1. Generalized Basic Belief Assignment

Suppose that U is a FoD in the open world. Its power set, 2 G U , is composed of 2 U propositions. A mass function is a mapping m G : 2 G U [ 0 , 1 ] that satisfies
A 2 G U m G ( A ) = 1 ,
where m G is GBBA of the U.
The difference between GBBA and classical BBA is the restriction of ∅. Note that m G ( ) = 0 is not necessary for GBBA, namely, the empty set can also be a focal element. GBBA will degenerate to classical BBA when m ( ) = 0 .

2.2.2. Generalized Combination Rule

In GET, 1 2 = is assigned to conflict coefficient K. Given two GBBAs ( m 1 and m 2 ), the GCR is defined as follows:
m ( A ) = ( 1 m ( ) ) B C = A m 1 ( B ) m 2 ( C ) 1 K , K = B C = m 1 ( B ) m 2 ( C ) , m ( ) = m 1 ( ) · m 2 ( ) , m ( ) = 1 i f a n d o n l y i f K = 1 .

2.3. Modified Generalized Combination Rule

The mGCR is presented by Jiang et al., where the intersection = 1 2 is considered as support to ∅, namely, the orthogonal sum of m 1 ( ) and m 2 ( ) is normalized like other focal elements. Given two GBBAs, the mGCR is defined as follows:
m ( A ) = B C = A m 1 ( B ) m 2 ( C ) 1 K m ( ) = m 1 ( ) m 2 ( ) 1 K
with
K = B C = B C m 1 ( B ) m 2 ( C ) m ( ) = 1 i f K = 1 o r A m ( A ) = 0 .
The mGCR inherits all the advantages and properties of the original GCR.

3. The Proposed Method

To obtain an objective classification, the key step is how to gain an effective BBA. As can be seen in Figure 1, a new method to determine BBA is presented. First, the dataset is divided into two parts, where a part is taken as the training set and another part is taken as the test set. Then the Gaussian models of k attributes are built using the training set, and the attribute models are tested by the test set to produce the similarity. Finally, the weight of each attribute is constructed based on the overlap degree of each class, which is used to modify the similarity and then to determine BBA.

3.1. The Modeling of Each Attribute

The Gaussian distribution is the common probability distribution in statistics, which is easy to analyze. Therefore, the membership function of Gaussian type is adopted to build the attribute models in this paper.
Suppose in the frame of discernment (FoD) Θ = { θ 1 , θ 2 , , θ n } , each class θ i ( i = 1 , 2 , , n ) has j ( j = 1 , 2 , , k ) attributes. The mean value X ¯ i j and the standard deviation σ i j of all the training samples in class θ i are calculated as follows:
X ¯ i j = 1 N l = 1 N x i j l , σ i j = 1 N 1 l = 1 N ( x i j l X ¯ i j ) 2 ,
where x i j l is the sample value of the jth attribute from the lth training sample in class θ i .
Hence, the corresponding attribute model of Gaussian type is obtained as follows:
μ A ( x ) = e x p ( ( x X ¯ i j ) 2 2 σ i j 2 )
The real-world sensor data usually has stochastic nature. To make the results more objective and credible, the training samples from each class are randomly selected to build the attribute model. For each attribute, n membership functions of Gaussian type are obtained as models of different class in the specific attribute.

3.1.1. The Construction of the Singleton Subsect and Compound Subsect

In D-S evidence theory, BBA has two forms. One is the singleton subsect, such as { A } , which represents a certain proposition. The other is the compound subsect, such as { A B } , which represents a kind of uncertain proposition. Namely, it indicates that { A } proposition or { B } proposition occurs, and it is unknown how to assign belief in { A } and { B } . For instance, in the case of motor rotor fault diagnosis, there are three faults: A, B and C, which represent the unbalance, misalignment and pedestal looseness respectively. BBA from a sensor is m ( { A } ) = 0 . 55 , m ( { A C } ) = 0 . 45 . Where { A } represents that the fault of the motor rotor is the unbalance, which is a singleton subsect proposition; { A C } represents that the fault of the motor rotor is the unbalance or pedestal looseness, which is a compound subsect proposition.
In this paper, the proposed method can automatically generate BBA of the singleton subsect and the compound subsect. This section mainly introduces how to construct the singleton subsect proposition and the compound subsect proposition. As shown in Figure 2, an attribute model of Gaussian type is denoted as a singleton subsect proposition, where μ A ( x ) and μ B ( x ) are all singleton subsect propositions. The compound subsect proposition is constructed by the overlap area of some membership functions of Gaussian type. The compound subsect proposition { A B } is noted as follows:
μ A B ( x ) = m i n ( μ A ( x ) , μ B ( x ) )
where w = s u p m i n ( μ A ( x ) , μ B ( x ) ) .

3.2. The Measurement of Similarity

In this paper, a nested structure similarity function s i m is defined to represent the matching degree between the test sample and the attribute model. This structure can avoid the high conflict, to some extent. The similarity is measured using the following regulations:
(1) If there is only one intersection between the test sample and the attribute models, then this intersection is assigned to the corresponding singleton subset as the similarity between this test sample and the corresponding singleton subset model;
(2) If there is more than one intersection between the test sample and the attribute models, then the top value of the intersections is assigned to the corresponding singleton subset as the similarity; the low value of the intersections is assigned to the corresponding compound subset as the similarity.
In the actual applications, the test model is covered by a different number of test data. For example, in the dataset, the test model is usually covered by one test data which, on account of its attribute value, is a fixed value, such as Iris. So this test model is the discrete value, as is the case for the test samples x 0 and x 1 which are shown in Figure 3. However, in a complex system, the attribute value is not a fixed value, for example, sensor data is not constant in sensor networks and is thus easily affected by the external environment. In such a case, the attribute is measured repeatedly by the sensor and the sensor data will be a set of values. Thus, this test model is covered by multiple test data, and is fuzzed into the membership function of Gaussian type, as is the case for the test models μ x 0 and μ x 1 which are shown in Figure 4. As shown in the analysis above, under the different application backgrounds, the measurement of the similarity is slightly different.
If the test model is covered by one test data, as shown in Figure 3a, for the test sample x 0 , the similarity is measured by the above regulations (1):
s i m ( B ) = μ B ( x 0 ) = a 0 ,
where μ B is the membership functions of the singleton subsect propositions { B } . x 0 has no intersection with μ A and μ A B , so s i m ( A ) = s i m ( A B ) = 0 .
As shown in Figure 3b, for the test sample x 1 , the similarity is measured by the above regulations (2):
s i m ( B ) = μ B ( x 1 ) = a 1 , s i m ( A B ) = μ A B ( x 1 ) = a 2 ,
where μ B is the membership functions of the singleton subsect propositions { B } , and μ A B is the membership functions of the compound subsect propositions { A B } .
Although x 1 has an intersection with μ A , this intersection is assigned to the compound subsect { A B } since it simultaneously locates in the overlapping portions between μ A and μ B . Based on the above regulations (2), s i m ( A ) = 0 and s i m ( A B ) = a 2 .
If the test model is covered by multiple test data, then the model is first constructed using the membership function of Gaussian type, which is similar to the model of training sets. Then, as shown in Figure 4a, for the test model μ x 0 , the similarity is measured by the above regulations (1):
s i m ( B ) = a 0 ,
where μ x 0 has no intersection with μ A and μ A B , so s i m ( B ) = s i m ( A B ) = 0 .
As shown in Figure 4b, for the test model μ x 1 , the similarity is measured by the above regulations (2):
s i m ( A ) = a 1 , s i m ( A B ) = a 2 ,
where although μ x 1 has an intersection with μ B ˜ , this intersection is assigned to the compound subsect { A B } since it simultaneously locates in the overlapping portions between μ A and μ B . Based on the regulations (2), s i m ( B ) = 0 and s i m ( A B ) = a 2 .

3.3. The Construction of Attribute Weights

In the multiclass classification, a comprehensive evaluation of each attribute needs to be made to obtain an objective result [47,48,49]. Given an attribute, if the similarity of some classes is high, that is, the overlap degree of the attribute models of these classes is large, then the ability to discriminate the difference of these classes of this attribute is weak. In this case, a false classification may easily occur based on this attribute, so the reliability of this attribute is low. Further, BBA generated from this attribute has a smaller contribution in multiclass classification. On the contrary, if the similarity of some classes is low, then the ability to discriminate the difference of these classes of this attribute is fine, so the reliability of this attribute is high. Further, BBA generated from this attribute has a larger contribution in multiclass classification. Thus, the effect of the attributes which provide a larger contribution should be enhanced and the effect of the attributes which provide a smaller contribution should be weakened to obtain more objective results. As analyzed above, the attribute weights are proposed and taken into account in this paper.
Suppose μ i j ( i = 1 , 2 , , n ; j = 1 , 2 , , k ) is the membership function of the jth attribute of the ith class; μ r j ( r = 1 , 2 , , n 2 n 2 ; j = 1 , 2 , , k ) is the generalized triangular fuzzy number model of the rth compound subsect proposition { A B } in the jth attribute; S ( x ) is the area of the x. Then the attribute weight w j ( j = 1 , 2 , , k ) is proposed as follows:
w j = 1 r = 1 n 2 n 2 S ( μ r j ) i = 1 n S ( μ i j ) r = 1 n 2 n 2 S ( μ r j )
where w j reveals that the larger the overlap degree, the larger the similarity, and the smaller the attribute weight.

3.4. The Construction of BBA

In D-S evidence theory, the closed world means that the FoD is complete; the open world means that the FoD is incomplete, namely, unknown propositions potentially exist outside of the given FoD. In the closed world, the FoD is complete, so the redundant BBA should be assigned to the universal set Θ , which is gained after using the attribute weights to correct the similarity s i m . However, in the open world, an unknown proposition exists. In GET, BBA of the unknown proposition is determined based on the known propositions. Then, the attribute weights are also constructed by the model of the known propositions and used to correct the similarity of the known propositions. Finally, the redundant BBA is assigned to the universal set Θ . According to the analysis above, the determination of BBA in the closed world and the open world is proposed respectively.

3.4.1. The Determination of BBA in the Closed World

In the closed world, BBA is determined by three steps. The similarity s i m is generated in the first step, which is described in Section 3.2. In the second step, if s i m > 1 , the similarity is normalized; if s i m < 1 , the similarity is not normalized. The FoD is complete in the closed world, so where s i m ( ) = 0 is necessary. In the third step, BBA of each proposition is obtained as follows:
m ( ) = 0 m ( A ) = w j × s i m j ( A ) , ( A Θ a n d A ) m ( Θ ) = 1 j = 1 k w j s i m j ( A )
where w j is denoted as the jth attribute’s weight, and s i m j is denoted as the similarity function of the jth attribute.
For example, in the frame of discernment Θ = { a , b , c } , the similarity function s i m 2 of attribute 2 is s i m 2 ( a ) = 0 . 6 , s i m 2 ( a b ) = 0 . 2 and the attribute weigh w 2 = 0 . 7 , then BBA is determined as follows:
m 2 ( a ) = w 2 × s i m 2 ( a ) = 0 . 7 × 0 . 6 = 0 . 42 , m 2 ( a b ) = w 2 × s i m 2 ( a b ) = 0 . 7 × 0 . 2 = 0 . 14 , m 2 ( Θ ) = m 2 ( a b c ) = 1 m 2 ( a ) m 2 ( a b ) = 1 0 . 42 0 . 14 = 0 . 44 , m 2 ( b ) = m 2 ( c ) = m 2 ( a c ) = m 2 ( b c ) = 0 .
After determining BBA, BBAs of k attributes are combined k 1 times with Dempster’s combination rule and the final recognized result can be gained.

3.4.2. The Determination of BBA in the Open World

In the open world, BBA is also determined by three steps. The similarity s i m is generated in the first step, which is described in Section 3.2. The FoD is incomplete in the open world, so where s i m ( ) = 0 is not necessary. Hence, in the second step, if s i m > 1 , then the similarity is normalized and the similarity of the empty set s i m ( ) = 0 ; if s i m < 1 , then the similarity is not normalized and the similarity of the empty set s i m ( ) = 1 s i m . In the third step, BBA of each proposition is obtained as follows:
m ( ) = s i m ( ) m ( A ) = w j × s i m j ( A ) , ( A Θ a n d A ) m ( Θ ) = 1 j = 1 k w j s i m j ( A ) m ( )
where w j is denoted as the jth attribute’s weight, and s i m j is denoted as the similarity function of the jth attribute.
For example, in the frame of discernment Θ = { a , b } , the similarity function s i m 1 of attribute 1 is s i m 1 ( a ) = 0 . 6 , s i m 1 ( a b ) = 0 . 2 and the attribute weigh w 1 = 0 . 8 , then BBA is determined as follows. Firstly, the similarity of the empty set is obtained as:
s i m 1 ( ) = 1 ( s i m 1 ( a ) + s i m 1 ( a b ) ) = 1 ( 0 . 6 + 0 . 2 ) = 0 . 2 .
Then BBA of each proposition is obtained as:
m 1 ( ) = s i m 1 ( ) = 0 . 2 , m 1 ( a ) = w 1 × s i m 1 ( a ) = 0 . 8 × 0 . 6 = 0 . 48 , m 1 ( b ) = w 1 × s i m 1 ( b ) = 0 . 8 × 0 = 0 , m 1 ( Θ ) = m 1 ( a b ) = 1 m 1 ( ) m 1 ( a ) m 1 ( b ) = 1 0 . 2 0 . 48 0 = 0 . 32 .
After determining BBA, BBAs of k attributes are combined k 1 times with the mGCR and the final recognized result can be gained.

4. Application Example

In this section, several experiments are performed both in the closed world and the open world respectively, to evaluate the validity and reasonability of the presented method. To better evaluate this method, an engineering example of fault diagnosis of a motor rotor is carried out. Note that the numbers from the following calculations are all unified as the four significant digits.
In addition, since the real-world data usually has stochastic nature, it is less advisable that the data with the highest recognition rate is selected for the experiment. Hence, in the following experiment, the training set and the test set are randomly selected, which makes the results more objective and credible.

4.1. An Example of Iris

The Iris dataset [50,51] contains three classes: Iris Setosa(S), Iris Versicolour(E) and Iris Virginica(V). Each class has 50 samples, and each sample contains four attributes: sepal length(SL), sepal width(SW), petal length(PL) and petal width(PW), respectively. Each of the four attributes is treated as an information source, and correspondingly there are three training sets and three test sets which are the same for the sample number. The data is obtained from the UCI repository of machine learning databases. (UCI Machine Learning Repository: http://archive.ics.uci.edu/ml/datasets/Iris.)

4.1.1. Experiment in the Closed World

Thirty samples are randomly selected as training samples from three classes respectively, and the remaining 20 samples of each of the three classes are taken as the test samples. Then, the experiment is carried out as follows.
Step 1: According to Section 3.1, the attribute models of each attribute of the training samples are obtained and shown in Figure 5.
Step 2: A test sample ( 4 . 6 , 3 . 1 , 1 . 5 , 0 . 2 ) is selected from the test set of S. Then, based on Section 3.2, the similarity between this test sample and each attribute model is obtained, which is shown in Figure 5 and Table 1.
From Table 1, we can see that the similarity of the SW attribute does not coincide with the reality. This is because the SW attributes of the three classes are mutually overlapping, which is not easy to distinguish in this case.
Step 3: According to Equation (13), the attribute weights of the four attributes are obtained as follows:
w S L = 1 1 . 0247 3 . 6677 1 . 0247 = 0 . 6123 , w S W = 1 0 . 7934 2 . 2250 0 . 7934 = 0 . 4458 , w P L = 1 0 . 2980 3 . 0420 0 . 2980 = 0 . 8914 , w P W = 1 0 . 0787 1 . 4263 0 . 0787 = 0 . 9416 .
Step 4: Based on Section 3.4, BBAs are determined as shown in Table 2.
Finally, four BBAs are combined three times with Dempster’s combination rule (Equation(4)), and the results are shown as follows:
m ( S ) = 0 . 9714 , m ( E ) = 0 , m ( V ) = 0 , m ( S E ) = 0 . 0021 , m ( S V ) = 0 . 0009 , m ( E V ) = 0 , m ( S E V ) = 0 . 0256 .
From the final BBA, it is clear that the test sample is classified into S class which coincides with the reality.

4.1.2. Experiment in the Open World

To verify the presented method’s performance in the open world, the training samples are randomly selected only from two Iris classes which are randomly selected from the three classes, and the remaining class is taken as the unknown class. The test samples are randomly selected from each of the three classes. In the Iris dataset, each class has 50 samples. In this paper, 30 samples of two classes are randomly selected as training samples from S and E respectively; the remaining 20 samples of each of these two classes and the 20 samples which are randomly selected from V, are taken as the test samples. Then, the experiment is carried out as follows.
Step 1: According to Section 3.1, the attribute model of each attribute of the training samples is obtained and shown in Figure 6.
Step 2: A test sample ( 5 . 5 , 4 . 2 , 1 . 4 , 0 . 2 ) is selected from the test set of S. Then, based on Section 3.2, the similarity between this test sample and each attribute model is obtained, which is shown in Figure 6 and Table 3.
From Table 3, we can see that the similarity of the PL attribute and the PW attribute coincides with the reality, but the similarity of the SL attribute and the SW attribute does not coincide with the reality. This is because the classes are not easy to distinguish if the overlap degree among the classes in the SL attribute and the SW attribute is large. Hence, the attribute weights are necessary when determining BBA.
Step 3: According to Equation (13), the attribute weights of four attributes are obtained as follows:
w S L = 1 0 . 3605 2 . 3381 0 . 3605 = 0 . 8177 , w S W = 1 0 . 2504 1 . 6729 0 . 2504 = 0 . 8240 , w P L = 1 0 1 . 6847 0 = 1 , w P W = 1 0 0 . 7362 0 = 1 .
Step 4: Based on Section 3.4, BBAs are determined as shown in Table 4.
Finally, four BBAs are combined three times with the mGCR (Equation(8)), and the results are shown as follows:
m ( S ) = 0 . 9980 , m ( E ) = 0 . 0012 , m ( S E ) = 0 . 0007 , m ( ) = 0 . 0001 .
From the final BBA, it is clear that the test sample is classified into S class which coincides with the reality.
Similarly, a test sample (6.8,3.0,5.5,2.1) is selected from the test set of unknown class (V), then the final BBA is obtained as follows:
m ( S ) = 0 , m ( E ) = 0 , m ( S E ) = 0 , m ( ) = 1 ,
where m ( ) = 1 indicates that the possibility of this test sample outside the FoD is very large, which coincides with the reality.

4.2. Experiments on Three Datasets: Five-Fold Cross-Validation

To further evaluate the proposed method, this method is compared with three well-known classifiers: support vector machine with radial basis function (SVM-RBF), decision tree learner (REPTree) and Naive Bayesian (NB) both in the closed world and the open world.
Three datasets are experimented in this section, which are obtained from the UCI repository of machine learning databases [52]. (UCI Machine Learning Repository: http://archive.ics.uci.edu/ml/datasets.) Within this, the Iris dataset is perhaps the best known database found in the pattern recognition literature. The dataset contains three classes of 50 instances each, where each class refers to a type of iris plant. Each class contains four attributes. The Seeds dataset contains three different varieties of wheat: Kama (K), Rosa (R) and Canadian (C). Each variety has 70 samples, and each sample contains seven attributes. The Wine dataset was the result of a chemical analysis of wines grown in the same region in Italy but derived from three different cultivars. The analysis determined the quantities of 13 constituents found in each of the three types of wine. A summary of these datasets is shown in Table 5.
The comparison results of five-fold cross-validation are shown in Table 6, Table 7 and Table 8 respectively.
From the above experimental results, it is found that:
(1) The results in the closed world
Both the proposed method and the classical machine learning algorithms are efficient in the closed world. From the first four lines of Table 6, Table 7 and Table 8, we can find that the proposed method obtains competitive performances with respect to machine learning algorithms. For example, in the first four lines of Table 7, the average recognition rate of Seeds of our method is 90.57%, NB is 78.09%, REPTree is 89.49% and SVM-RBF is 90.21%.
(2) The results in the open world
From the latter 12 lines of Table 6, Table 7 and Table 8, we can find that the average recognition rate of the empty set ∅ is 0 by the classical machine learning classifiers. It means that the “unknown” class is definitely misclassified by the classical machine learning classifiers. These standard machine learning classification algorithms cannot work in an open world environment.
In contrast, the “unknown” class in an open world environment can be classified by the proposed method. For example, in Table 8, the average recognition rate of the empty set ∅ of wine is 89.33%, 93.78% and 91.62% respectively in the frame of discernment of { A , B } , { A , C } and { B , C } . Therefore, our method is still effective in an open world.
By summarizing (1) the results in the closed world and (2) the results in the open world, it can be found that both the proposed method and the classical machine learning algorithms are efficient in the closed world. However, the classical machine learning algorithms cannot work in an open world. In contrast, our method is still effective in an open world, which is the advantage of our method compared with the classical machine learning algorithms.

4.3. An Example of Fault Diagnosis

Suppose there are three types of fault in a motor rotor, which are noted as F = { F 1 , F 2 , F 3 } = { r o t o r u n b a l a n c e , r o t o r m i s a l i g n m e n t , P e d e s t a l l o o s e n e s s } . Three vibration acceleration sensors and a vibration displacement sensor are placed in different installation positions to collect the vibration signal. Vibration displacement and acceleration vibration frequency amplitudes at the frequencies of 1X, 2X and 3X are taken as the fault feature variables. At the same time interval, each fault feature of each fault is continuously observed 40 times which is taken as a group of observations. In this paper, a total of five groups are measured in each fault feature of each fault. For example, five groups of observations of the fault feature variables of 1X of the rotor unbalance are shown in Table 9.
The example is carried out by the proposed method and three well-known classifiers: support vector machine with radial basis function (SVM-RBF), decision tree learner (REPTree) and Naive Bayesian (NB), respectively. The comparison results of the five-fold cross-validation are shown in Table 10.
The above experimental results demonstrate the validity of the proposed method. The advantages of the proposed method are discussed and concluded as follows:
  • The weights of the attributes have been considered in the proposed BBA generation method. Because of that, the generated BBA could reflect the difference of the importance between different attributes which have different classification ability. If the classification ability of each attribute is similar, then the result obtained by using the proposed method is basically identical with such a method that does not consider attributes’ weights. However, if attributes have different classification ability, then the proposed method has better performance since it has considered the weight of each attribute.
  • The weight of each attribute is totally determined based on the discrimination of the attribute, but without relying on other information. In many existing approaches, the weights of attributes are either from an externally given credibility of sensors or mutual support degree of different attributes. In this paper, the weight of every attribute is determined just by the attribute’s discrimination, before the amendment of the generated BBA is implemented. This approach is completely a data-driven solution, which makes the results more objective and credible.
  • The proposed BBA generation method is efficient simultaneously in the closed world and open world. At present, many existing BBA generation approaches only consider the situation of the closed world. In contrast, by introducing generalized evidence theory, the incomplete information about the frame of discernment is taken into consideration in the initial phase of the modelling of uncertain information. Therefore, it is more reasonable and realistic.
  • The proposed method is simple and can be easily used in many practical applications. In addition, it is based on the normal distribution assumption and can be easily changed to other forms to reflect the feature of training data much more realistically. Therefore, the proposed method is flexible and easily extensible.
However, compared to the traditional BBA method, namely the method that does not consider the attribute weights, the approach of weighted BBA has complicated calculations. However, considering that the performance is better after using the attribute weights, the increased computational complexity is acceptable to some extent.

5. Conclusions

In the application of soft sensors data fusion, based on evidence theory, how to determine BBA is an open issue. In this paper, a new method is presented to determine BBA in a closed or open world environment. Within the proposed BBA determination method, all data are separated into a training set and test set, and the attributes of each sample are seen as soft sensors. At first, according to the training samples, each attribute’s Gaussian model is built. Then, for every test sample, by considering the similarity and the attribute weights, a BBA is derived to express the test sample. Finally, according to the derived BBA, the class of the corresponding test sample is determined.
The advantages of the proposed BBA determination method include the following aspects. At first, the weights of attributes, which reflect the reliability of information sources, have been considered in the process of BBA generation. Moreover, it is compatible with the method which does not consider the attribute weights. Second, since the generalized evidence theory has been used in this work, the proposed method is efficient not only in the closed world but also in the open world. Many previous BBA determination methods have only been able to deal with the situation of the closed world. Third, the whole process is completely data-driven, which makes the results more objective. The proposed method can be easily used in many practical applications, for example, classification and fault diagnosis. A number of simulation experiments show that the proposed method is effective. However, some shortcomings still exist in the proposed method. Mainly, the amount of computation of the proposed method is increased since the weights of attributes have been considered. Besides, the proposed method relies on a proper training set, which requires a certain size of data samples.
In the future, we will research further the following problems. Firstly, we will try to apply the proposed method to more practical applications, and import a pre-judging procedure to determine whether the attribute weights need to be considered to reduce the computational complexity. Furthermore, the possible dependencies among the attributes of the actual targets will be considered to improved the proposed method.

Acknowledgments

The work is partially supported by National Natural Science Foundation of China (Grant No. 61671384), Natural Science Basic Research Plan in Shaanxi Province of China (Program No. 2016JM6018), Aviation Science Foundation (Program No. 20165553036), the Fund of SAST (Program No. SAST2016083), the Seed Foundation of Innovation and Creation for Graduate Students in Northwestern Polytechnical University (Program No. Z2017141).

Author Contributions

Wen Jiang designed all the research. Wen Jiang and Miaoyan Zhuang analyzed the data and wrote the manuscript. Miaoyan Zhuang and Chunhe Xie performed the experiments. Jun Wu analyzed the experiment results. All authors read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, L.; Wu, X.; Zhu, H.; AbouRizk, S.M. Perceiving safety risk of buildings adjacent to tunneling excavation: An information fusion approach. Autom. Constr. 2017, 73, 88–101. [Google Scholar] [CrossRef]
  2. Ma, J.; Liu, W.; Miller, P.; Zhou, H. An evidential fusion approach for gender profiling. Inf. Sci. 2016, 333, 10–20. [Google Scholar] [CrossRef]
  3. Deng, Y. Deng entropy. Chaos Solitons Fractals 2016, 91, 549–553. [Google Scholar] [CrossRef]
  4. Fortuna, L.; Graziani, S.; Xibilia, M.G. Comparison of soft-sensor design methods for industrial plants using small data sets. IEEE Trans. Instrum. Meas. 2009, 58, 2444–2451. [Google Scholar] [CrossRef]
  5. McKeown, M.R.; Shaw, D.L.; Fu, H.; Liu, S.; Xu, X.; Marineau, J.J.; Huang, Y.; Zhang, X.; Buckley, D.L.; Kadam, A.; et al. Biased multicomponent reactions to develop novel bromodomain inhibitors. J. Med. Chem. 2014, 57, 9019–9027. [Google Scholar] [CrossRef] [PubMed]
  6. Jiang, W.; Wei, B.; Tang, Y.; Zhou, D. Ordered visibility graph average aggregation operator: An application in produced water management. Chaos 2017, 27. [Google Scholar] [CrossRef] [PubMed]
  7. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
  8. Zhang, L.; Wu, X.; Qin, Y.; Skibniewski, M.J.; Liu, W. Towards a Fuzzy Bayesian Network Based Approach for Safety Risk Analysis of Tunnel-Induced Pipeline Damage. Risk Anal. 2016, 36, 278–301. [Google Scholar] [CrossRef] [PubMed]
  9. Jiang, W.; Xie, C.; Luo, Y.; Tang, Y. Ranking Z-numbers with an improved ranking method for generalized fuzzy numbers. J. Intell. Fuzzy Syst. 2017, 32, 1931–1943. [Google Scholar] [CrossRef]
  10. Quost, B.; Denoeux, T. Clustering and classification of fuzzy data using the fuzzy EM algorithm. Fuzzy Sets Syst. 2016, 286, 134–156. [Google Scholar] [CrossRef]
  11. Dempster, A.P. Upper and Lower Probabilities Induced by a Multi-valued Mapping. Ann. Math. Stat. 1967, 38, 325–339. [Google Scholar] [CrossRef]
  12. Shafer, G. A Mathematical Theory of Evidence; Princeton University Press: Princeton, NJ, USA, 1976. [Google Scholar]
  13. Zadeh, L.A. A Note on Z-numbers. Inf. Sci. 2011, 181, 2923–2932. [Google Scholar] [CrossRef]
  14. Jiang, W.; Xie, C.; Zhuang, M.; Shou, Y.; Tang, Y. Sensor Data Fusion with Z-Numbers and Its Application in Fault Diagnosis. Sensors 2016, 16, 1509. [Google Scholar] [CrossRef] [PubMed]
  15. Mo, H.; Deng, Y. A new aggregating operator for linguistic information based on D numbers. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2016, 24, 831–846. [Google Scholar] [CrossRef]
  16. Zhou, X.; Deng, X.; Deng, Y.; Mahadevan, S. Dependence assessment in human reliability analysis based on D numbers and AHP. Nucl. Eng. Des. 2017, 313, 243–252. [Google Scholar] [CrossRef]
  17. Deng, X.; Lu, X.; Chan, F.T.; Sadiq, R.; Mahadevan, S.; Deng, Y. D-CFPR: D numbers extended consistent fuzzy preference relations. Knowl.-Based Syst. 2015, 73, 61–68. [Google Scholar] [CrossRef]
  18. Alexandridis, A.; Chondrodima, E.; Sarimveis, H. Radial basis function network training using a nonsymmetric partition of the input space and particle swarm optimization. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 219–230. [Google Scholar] [CrossRef] [PubMed]
  19. Alexandridis, A. Evolving RBF neural networks for adaptive soft-sensor design. Int. J. Neural Syst. 2013, 23. [Google Scholar] [CrossRef] [PubMed]
  20. Du, W.B.; Zhou, X.L.; Lordan, O.; Wang, Z.; Zhao, C.; Zhu, Y.B. Analysis of the Chinese Airline Network as multi-layer networks. Transp. Res. Part E Logist. Transp. Rev. 2016, 89, 108–116. [Google Scholar] [CrossRef]
  21. Wang, Z.; Bauch, C.T.; Bhattacharyya, S.; d’Onofrio, A.; Manfredi, P.; Perc, M.; Perra, N.; Salathé, M.; Zhao, D. Statistical physics of vaccination. Phys. Rep. 2016, 664, 1–113. [Google Scholar] [CrossRef]
  22. Zhang, R.; Ran, X.; Wang, C.; Deng, Y. Fuzzy Evaluation of Network Vulnerability. Qual. Reliab. Eng. Int. 2016, 32, 1715–1730. [Google Scholar] [CrossRef]
  23. Jiang, W.; Wei, B.; Zhan, J.; Xie, C.; Zhou, D. A visibility graph power averaging aggregation operator: A methodology based on network analysis. Comput. Ind. Eng. 2016, 101, 260–268. [Google Scholar] [CrossRef]
  24. Yu, Y.; Xiao, G.; Zhou, J.; Wang, Y.; Wang, Z.; Kurths, J.; Schellnhuber, H.J. System crash as dynamics of complex networks. Proc. Natl. Acad. Sci. USA 2016, 113, 11726–11731. [Google Scholar] [CrossRef] [PubMed]
  25. Zhao, X.; Wang, R.; Gu, H.; Song, G.; Mo, Y. Innovative data fusion enabled structural health monitoring approach. Math. Probl. Eng. 2014, 2014. [Google Scholar] [CrossRef]
  26. He, Y.; Hu, L.; Guan, X.; Han, D.; Deng, Y. New conflict representation model in generalized power space. J. Syst. Eng. Electron. 2012, 23, 1–9. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Zhang, W.; Chao, H.C.; Lai, C.F. Toward belief function-based cooperative sensing for interference resistant industrial wireless sensor networks. IEEE Trans. Ind. Inf. 2016, 12, 2115–2126. [Google Scholar] [CrossRef]
  28. Yager, R. A class of fuzzy measures generated from a Dempster-Shafer belief structure. Int. J. Intell. Syst. 1999, 14, 1239–1247. [Google Scholar] [CrossRef]
  29. Zhu, Y.M.; Bentabet, L.; Dupuis, O.; Kaftandjian, V.; Babot, D.; Rombaut, M. Automatic determination of mass functions in Dempster-Shafer theory using fuzzy C-means and spatial neighborhood information for image segmentation. Opt. Eng. 2002, 41, 760–770. [Google Scholar] [CrossRef]
  30. Dubois, D.; Prade, H.; Sandri, S. On possibility/probability transformations. Theory and Decision Library of Fuzzy Logic: State of the Art. 1993, 12, 103–112. [Google Scholar]
  31. Baudrit, C.; Dubois, D. Practical representations of incomplete probabilistic knowledge. Comput. Stat. Data Anal. 2006, 51, 86–108. [Google Scholar] [CrossRef]
  32. Wang, H.; Mcclean, S. Deriving evidence theoretical functions in multivariate data spaces: a systematic approach. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2008, 38, 455–465. [Google Scholar] [CrossRef] [PubMed]
  33. Masson, M.H.; Denoeux, T. Inferring a possibility distribution from empirical data. Fuzzy Sets Syst. 2006, 157, 319–340. [Google Scholar] [CrossRef]
  34. Antoine, V.; Quost, B.; Masson, M.H.; Denoeux, T. CEVCLUS: evidential clustering with instance-level constraints for relational data. Soft Comput. 2013, 18, 1321–1335. [Google Scholar] [CrossRef]
  35. De Campos, L.; Huete, J. Measurement of possibility distributions. Int. J. Gen. Syst. 2001, 30, 309–346. [Google Scholar] [CrossRef]
  36. Song, M.; Jiang, W. Engine fault diagnosis based on sensor data fusion using evidence theory. Adv. Mech. Eng. 2016, 8, 1–9. [Google Scholar] [CrossRef]
  37. Song, M.; Jiang, W.; Xie, C.; Zhou, D. A new interval numbers power average operator in multiple attribute decision making. Int. J. Intell. Syst. 2017. [Google Scholar] [CrossRef]
  38. Smets, P.; Kennes, R. The transferable belief model. Artif. Intell. 1994, 66, 191–234. [Google Scholar] [CrossRef]
  39. Ayoun, A.; Smets, P. Data association in multi-target detection using the transferable belief model. Int. J. Intell. Syst. 2001, 16, 1167–1182. [Google Scholar] [CrossRef]
  40. Quost, B.; Masson, M.H.; Denoeux, T. Classifier fusion in the Dempster-Shafer framework using optimized t-norm based combination rules. Int. J. Approx. Reason. 2011, 52, 353–374. [Google Scholar] [CrossRef]
  41. Daniel, J.; Lauffenburger, J.P. Fusing navigation and vision information with the Transferable Belief Model: Application to an intelligent speed limit assistant. Inf. Fusion 2014, 18, 62–77. [Google Scholar] [CrossRef]
  42. Deng, Y. Generalized evidence theory. Appl. Intell. 2015, 43, 530–543. [Google Scholar] [CrossRef]
  43. Jiang, W.; Zhan, J. A modified combination rule in generalized evidence theory. Appl. Intell. 2017, 46, 630–640. [Google Scholar] [CrossRef]
  44. Deng, X.; Han, D.; Dezert, J.; Deng, Y.; Shyr, Y. Evidence combination from an evolutionary game theory perspective. IEEE Trans. Cybern. 2016, 46, 2070–2082. [Google Scholar] [CrossRef] [PubMed]
  45. Deng, X.; Xiao, F.; Deng, Y. An improved distance-based total uncertainty measure in belief function theory. Appl. Intell. 2017. [Google Scholar] [CrossRef]
  46. Wang, J.; Xiao, F.; Deng, X.; Fei, L.; Deng, Y. Weighted evidence combination based on distance of evidence and entropy function. Int. J. Distrib. Sens. Netw. 2016, 12, 3218784. [Google Scholar] [CrossRef]
  47. Fu, C.; Wang, Y. An interval difference based evidential reasoning approach with unknown attribute weights and utilities of assessment grades. Comput. Ind. Eng. 2015, 81, 109–117. [Google Scholar] [CrossRef]
  48. Zhang, W.; Zhang, Z. Belief function based decision fusion for decentralized target classification in wireless sensor networks. Sensors 2015, 15, 20524–20540. [Google Scholar] [CrossRef] [PubMed]
  49. Liu, Y.Z.; Jiang, Y.C.; Liu, X.; Yang, S.L. A combination strategy for multiclass classification based on multiple association rules. Knowl.-Based Syst. 2008, 21, 786–793. [Google Scholar] [CrossRef]
  50. Fisher, R. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  51. Iris Data Set. Available online: http://archive.ics.uci.edu/ml/datasets/Iris (accessed on 29 March 2017).
  52. UCI Machine Learning Repository. Available online: http://archive.ics.uci.edu/ml/datasets (accessed on 29 March 2017).
  53. Wen, C.; Xu, X. Theories and Applications in Multi-Source Uncertain Information Fusion–Fault Diagnosis and Reliability Evaluation; Beijing Science Press: Beijing, China, 2012. [Google Scholar]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Sensors 17 00721 g001
Figure 2. The modeling of the singleton subsect and compound subsect.
Figure 2. The modeling of the singleton subsect and compound subsect.
Sensors 17 00721 g002
Figure 3. The measurement of similarity where the test model is a discrete value. (a) Only one intersection; (b) multiple intersections.
Figure 3. The measurement of similarity where the test model is a discrete value. (a) Only one intersection; (b) multiple intersections.
Sensors 17 00721 g003
Figure 4. The measurement of similarity where the test model is a continuous value. (a) Only one intersection; (b) multiple intersections.
Figure 4. The measurement of similarity where the test model is a continuous value. (a) Only one intersection; (b) multiple intersections.
Sensors 17 00721 g004
Figure 5. The models and similarity of each attribute in the closed world. (a) SL; (b) SW; (c) PL; (d) PW.
Figure 5. The models and similarity of each attribute in the closed world. (a) SL; (b) SW; (c) PL; (d) PW.
Sensors 17 00721 g005
Figure 6. The models and similarity of each attribute in the open world.(a) SL; (b) SW; (c) PL; (d) PW.
Figure 6. The models and similarity of each attribute in the open world.(a) SL; (b) SW; (c) PL; (d) PW.
Sensors 17 00721 g006
Table 1. Similarity of Iris in the closed world.
Table 1. Similarity of Iris in the closed world.
AttributesSimilarity
sim(S)sim(E)sim(V)sim(SE)sim(SV)sim(EV)sim(SEV)
SL0.5005000.0285000
SW000.403500.320100
PL0.9712000000
PW0.9098000000
Table 2. BBA of Iris in the closed world.
Table 2. BBA of Iris in the closed world.
AttributesBBA
m(S)m(E)m(V)m(SE)m(SV)m(EV)m(SEV)
SL0.3046000.0515000.6439
SW0.16650000.028300.8052
PL0.7442000000.2558
PW0.8067000000.1933
Table 3. Similarity of Iris in the open world.
Table 3. Similarity of Iris in the open world.
AttributesSimilarity
sim(S)sim(E)sim(SE)
SL00.72610.4027
SW0.118100
PL0.949300
PW0.933300
Table 4. BBA of Iris in the open world.
Table 4. BBA of Iris in the open world.
AttributesBBA
m(S)m(E)m(SE)m(∅)
SL00.59370.36780.0385
SW0.097300.45130.4513
PL0.949300.02530.0253
PW0.933300.03330.0333
Table 5. General information about the real datasets.
Table 5. General information about the real datasets.
Dataset#Instance#Class#Attribute
Iris15034
Seed21037
Wine178313
Table 6. The comparison results of Iris.
Table 6. The comparison results of Iris.
CasesFrameMethodsClassesOverall Average
Setosa (S)Versicolour (E)Virginica (V)
Closed world{S,E,V}SVM-RBF100.00%94.30%96.31%96.87%
REPTree100.00%92.38%92.65%95.01%
NB100%91.62%94.53%95.38%
Our method99.00%93.00%95.00%95.67%
Open world{S,E}SVM-RBF100.00%100.00%0 (V=∅)66.67%
REPTree100.00%100.00%0 (V=∅)66.67%
NB100.00%100.00%0 (V=∅)66.67%
Our method92.00%88.00%84.00% (V=∅)88%
{S,V}SVM-RBF100.00%0 (E=∅)100.00%66.67%
REPTree99.41%0 (E=∅)99.67%66.36%
NB100%0 (E=∅)100.00%66.67%
Our method84.00%80.00% (E=∅)90.00%84.67%
{E,V}SVM-RBF0 (S=∅)95.53%97.89%64.48%
REPTree0 (S=∅)93.11%90.61%61.24%
NB0 (S=∅)92.51%94.89%62.47%
Our method100.00% (S=∅)82.00%84.00%88.67%
Table 7. The comparison results of Seeds.
Table 7. The comparison results of Seeds.
CasesFrameMethodsClassesOverall Average
Kama (K)Rosa (R)Canadian (C)
Closed world{K,R,C}SVM-RBF82.14%94.41%94.08%90.21%
REPTree84.32%92.33%91.82%89.49%
NB76.03%69.81%88.42%78.09%
Our method87.71%93.43%90.57%90.57%
Open world{K,R}SVM-RBF93.07%93.58%0 (C=∅)62.21%
REPTree93.57%95.43%0 (C=∅)63.00%
NB78.04%77.22%0 (C=∅)51.75%
Our method85.71%82.86%88.57% (C=∅)85.71%
{K,C}SVM-RBF89.84%0 (R=∅)93.61%61.15%
REPTree89.44%0 (R=∅)91.04%60.16%
NB83.67%0 (R=∅)87.66%57.11%
Our method80.00%88.57% (R=∅)95.71%88.09%
{R,C}SVM-RBF0 (K=∅)100%100%66.67%
REPTree0 (K=∅)98.81%99.53%66.11%
NB0 (K=∅)91.91%89.12%60.34%
Our method84.29% (K=∅)91.43%84.29%86.67%
Table 8. The comparison results of Wine.
Table 8. The comparison results of Wine.
CasesFrameMethodsClassesOverall Average
ABC
Closed world{A,B,C}SVM-RBF6.86%99.93%4.75%37.18%
REPTree91.43%88.06%91.58%90.36%
NB88.36%82.88%84.58%85.27%
Our method89.85%95.90%95.78%93.84%
Open world{A,B}SVM-RBF7.36%99.36%0 (C=∅)35.68%
REPTree95.35%95.87%0 (C=∅)63.74%
NB88.07%93.19%0 (C=∅)60.42%
Our method84.85%88.95%89.33% (C=∅)87.71%
{A,C}SVM-RBF100.00%0 (B=∅)6.91%35.64%
REPTree99.29%0 (B=∅)99.17%66.15%
NB86.93%0 (B=∅)95.92%60.95%
Our method95.00%93.78% (B=∅)81.90%90.23%
{B,C}SVM-RBF0 (A=∅)100%5.25%35.08%
REPTree0 (A=∅)94.31%91.92%62.08%
NB0 (A=∅)84.88%88.00%57.63%
Our method91.62% (A=∅)96.00%79.85%89.16%
Table 9. Five groups of observations of 1X of the rotor unbalance [53].
Table 9. Five groups of observations of 1X of the rotor unbalance [53].
GroupsObservations
X110.16630.15900.15680.14850.17230.20060.19030.19080.19860.1843
0.17850.16100.15790.15110.15320.16470.16280.16460.16340.1642
0.16480.16400.16740.06610.16590.16500.16330.16320.16040.1542
0.15550.15620.15400.15640.15570.15420.15460.15710.15370.1536
X120.1540.15180.15370.15480.15420.15380.15450.15370.15710.1560
0.15840.15520.15860.15740.15690.15650.15510.15850.15850.1593
0.15480.15580.15470.15930.15320.16320.15750.1590.15940.1541
0.1650.16740.16510.16040.17870.18180.18200.16560.16580.1644
X130.16470.16470.16540.16510.16560.16530.16520.16520.16480.1649
0.16530.16500.16500.16520.16530.16520.16480.16470.16460.1645
0.16510.16520.16520.16490.16500.16430.16400.16390.16410.1633
0.16320.16290.16300.16300.16340.16310.16340.16290.16320.1629
X140.16300.16290.16270.16260.16220.16240.16270.16180.16140.1617
0.16210.16150.16180.16110.16140.16100.16120.16110.16160.1612
0.16120.16130.16230.16160.16210.16130.16110.16100.16100.1613
0.16150.16160.16180.16160.16140.16120.16060.16140.16190.1614
X150.16090.16100.16120.16150.16090.16060.16040.16060.16050.1601
0.16040.16080.16100.16030.15990.16010.16020.15990.15980.1598
0.15980.15960.15950.15930.15940.15980.15960.15970.15950.1593
0.15980.15960.15970.15950.15930.15770.15800.15760.15770.1579
Table 10. The comparison results of the fault diagnosis.
Table 10. The comparison results of the fault diagnosis.
CasesFrameMethodsClassesOverall Average
F 1 F 2 F 3
Closed world{ F 1 , F 2 , F 3 }SVM-RBF94.15%92.86%100%95.67%
REPTree99.05%98.68%99.78%99.17%
NB98.05%96.94%100%98.33%
Our method99.50%98.50%100%99.33%
Open world{ F 1 , F 2 }SVM-RBF99.14%98.49%0( F 3 =∅)65.88%
REPTree99.66%100%0( F 3 =∅)66.55%
NB98.86%99.48%0( F 3 =∅)66.11%
Our method96.00%90.00%100%( F 3 =∅)95.33%
{ F 1 , F 3 }SVM-RBF99.86%0( F 2 =∅)100%66.62%
REPTree99.76%0( F 2 =∅)100%66.59%
NB99.99%0( F 2 =∅)100%66.66%
Our method96.50%100%( F 2 =∅)94%96.83%
{ F 2 , F 3 }SVM-RBF0( F 1 =∅)99.38%100%66.46%
REPTree0( F 1 =∅)99.70%99.78%66.49%
NB0( F 1 =∅)99.41%100%66.47%
Our method100%( F 1 =∅)91.00%93.00%94.67%

Share and Cite

MDPI and ACS Style

Jiang, W.; Zhuang, M.; Xie, C.; Wu, J. Sensing Attribute Weights: A Novel Basic Belief Assignment Method. Sensors 2017, 17, 721. https://doi.org/10.3390/s17040721

AMA Style

Jiang W, Zhuang M, Xie C, Wu J. Sensing Attribute Weights: A Novel Basic Belief Assignment Method. Sensors. 2017; 17(4):721. https://doi.org/10.3390/s17040721

Chicago/Turabian Style

Jiang, Wen, Miaoyan Zhuang, Chunhe Xie, and Jun Wu. 2017. "Sensing Attribute Weights: A Novel Basic Belief Assignment Method" Sensors 17, no. 4: 721. https://doi.org/10.3390/s17040721

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop