Next Article in Journal
Symmetry in Organic/Inorganic Hybrid Materials
Previous Article in Journal
The kth Local Exponent of Doubly Symmetric Primitive Digraphs with d Loops
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolutionary Optimization for the Belief-Rule-Based System: Method and Applications

1
State Key Laboratory of Complex Eletromagnetic Environment Effects on Electronics and Information System, Luoyang 471003, China
2
Department of Automation, Hangzhou Dianzi University, Hangzhou 310005, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(8), 1622; https://doi.org/10.3390/sym14081622
Submission received: 14 July 2022 / Revised: 31 July 2022 / Accepted: 4 August 2022 / Published: 7 August 2022
(This article belongs to the Section Computer)

Abstract

:
Evolutionary optimization (EO) has been proven to be highly effective computation means in solving asymmetry problems in engineering practices. In this study, a novel evolutionary optimization approach for the belief rule base (BRB) system is proposed, namely EO-BRB, by constructing an optimization model and employing the Differential Evolutionary (DE) algorithm as its optimization engine due to its ability to locate an optimal solution for problems with nonlinear complexity. In the EO-BRB approach, the most representative referenced values of the attributes which are pre-determined in traditional learning approaches are to be optimized. In the optimization model, the mean squared error (MSE) between the actual and observed data is taken as the objective, while the initial weights of all the rules, the beliefs of the scales in the conclusion part, and the referenced values of the attributes are taken as the restraints. Compared with the traditional learning approaches for the BRB system, the EO-BRB approach (1) does not require transforming the numerical referenced values of the attributes into linguistic terms; (2) does not require identifying any initial solution; (3) does not require any mathematical deduction and/or case-specific information which verifies it as a general approach; and (4) can help downsize the BRB system while producing superior performances. Thus, the proposed EO-BRB approach can make the best use of the nonlinear modeling ability of BRB and the optimization superiority of the EO algorithms. Three asymmetry numerical and practical cases are studied to validate the efficiency of the proposed EO-BRB approach.

1. Introduction

The belief rule base (BRB) system is comprised of multiple rules in the same belief structure [1,2,3]. The BRB system is used to represent system dynamics, predict complex system behavior, and integrate different types of information under uncertainty. BRB has the advantages of generalizing the conventional probability distribution by allowing inexact reasoning [4] and dealing with uncertain information which is different from ignorance and equal likelihood [5]. In recent years, the BRB system has been successfully applied in solving problems and meeting challenges in multiple fields, including the multiple attribute decision analysis (MADA) problem [6], group decision making [7], clinical risk analysis [8], trade-off analysis [9], system readiness assessment [10], military capability evaluation [11], etc.
To apply the BRB system in modeling complex practical systems, each attribute (especially continuous attribute) is required to be discretized into a certain number of referenced values [12]. The number of the referenced values should be enough to represent system dynamics and complex system behavior. However, too many referenced values of the attributes may lead to an oversized BRB system which may increase the computational complexity and eventually make the BRB system impractical.
So far, there have been multiple studies regarding the learning frameworks and approaches for the BRB system. The first generic BRB learning framework and the corresponding optimization model were proposed in [13]. Then, a BRB inferring and training approach was applied in the pipeline leak detection, which had since become a benchmark for validating BRB and is also used in this study as Case III [14]. More approaches on BRB training and learning were proposed and applied in more theoretical and practical case studies [15,16,17,18,19,20,21], including the offline and online BRB updating approach, adaptive BRB learning approach, dynamic rule adjustment approach for BRB, etc.
This study summarizes the four challenges of current BRB training and learning endeavors as follows.
First, numerical values of the attributes need to be transformed into linguistic terms when a BRB system is constructed. Among present studies, the referenced values were all pre-determined and were never regarded as the parameters to be optimized, which essentially differentiates the work of this study from previous studies. The transformation process depends merely on human judgment which may be biased due to personal preference and prejudice and eventually affect the validity of the learning result. For example, in the pipeline leak detection case [14,15,16,17,18,19,20,21] (Case III in this study), the two attributes, PressureDiff and Flowdiff, are transformed into eight and seven linguistic terms, respectively.
Second, the traditional learning approaches are sensitive to the initial solution. The traditional learning approaches require first identifying the initial solution [13,14,15,16,17,18,19,20,21,22]. Furthermore, the optimization algorithm used in traditional learning approaches is based on the initial solution. Thus, the selection of the initial solution becomes very important because a poorly selected initial solution could invalidate the efficiency of the learning approach. For example, in [22], on the multiple-extremal function (Case I in this study), the local minimum/maximum points must be known as prior knowledge. If the selected referenced values of the attribute, “x”, are trapped in a local optimality, then it could severely affect the learning efficiency. To meet this challenge, the training dataset must be specifically chosen in a way that is consistent with the facts in most traditional studies. However, it may lose the universal versatility when the learning approach is introduced to other practical fields.
Third, the traditional learning approaches require complex mathematical deduction and strict assumptions. For example, the studies [14,15,16,17,18,19,20,21] on the pipeline leak detection case required calculating the ladder direction and the derivative function for the requirement of the optimization algorithm. In the residual life probability prediction case (Case II in this study), strict assumptions were made and multiple mathematical steps were taken in order to obtain the probability distribution function in previous studies [23,24].
Finally, the BRB system is not necessarily downsized after certain learning process (partially because it is never the intention of traditional learning approaches to downsize the BRB system). In most of the pipeline leak detection case-related studies [14,15,16,17,18], there were still 56 rules in the BRB system both before and after the learning process. Only [19,20,21] used the “statistical utility” to rate the rules and downsized the BRB system to 5 rules where the full-sized BRB system with 56 rules was still required to be constructed.
Therefore, a new EO-BRB approach is proposed in this study to meet these challenges by adopting evolutionary optimization (EO) [25,26,27,28,29,30,31] algorithms as the optimization engine. The EO-BRB approach includes the referenced values of the attribute as parameters to be optimized and therefore does not need to transform the numerical terms into linguistic terms. Moreover, the EO-BRB approach is no longer limited by the initial solution and can employ the evolution mechanism within an EO algorithm instead of resorting to a complex mathematical deduction process. Finally, the case studies results show that the EO-BRB approach can help downsize the BRB system as well as maintain high accuracy. Note that it is not the intention of this study to identify a specific EO algorithm as the most fitting technique for the EO-BRB approach. The metric to validate the efficiency and optimization objective of EO-BRB is the mean squared error (MSE) to maintain consistency with previous studies as well as for the purpose of fair comparison.
The remainder of this study is organized as follows. The background and the inference mechanism are introduced in Section 2. The EO-BRB approach is proposed in Section 3. In Section 4, three cases are studied to validate the efficiency of the EO-BRB approach. This study is concluded in Section 5.

2. Belief-Rule-Based System and Its Inferencing Mechanism

The BRB system [1,2,3] is composed of a series of rules in the same belief structure. The kth rule is described as:
R k : if   ( x 1 = A 1 k ) ( x m = A m k ) ( x m = A M k ) , then   { ( D 1 , β 1 , k ) , , ( D n , β n , k ) , , ( D N , β N , k ) } with   an   initial   rule   weight   θ k
where A m k (m = 1, …, M; k = 1, …, K) represents the referenced values of the mth attribute, M represents the number of the precedent attributes, N represents the number of the scales in the conclusion part. K represents the number of the rules, βn,k represents the belief for the nth scale, and Dn(n = 1, …, N). Additionally, the initial rule weight of the kth rule is θk.
The analytical ER algorithm [32] is used to integrate the activated rules. First, the activated rules are transformed into the basic probability mass (BMP) as in Equations (2)–(5).
m n , k = w k β ¯ n , k
m D , k = ( 1 w k n = 1 N β ¯ n , k )
m ¯ D , k = 1 w k
m ˜ D , k = w k ( 1 n = 1 N β ¯ n , k )
where wk denotes the weight of the activated kth rule, which is calculated based on θk.
Then, K activated rules are integrated using the following Equations (6)–(11).
m n = μ [ k = 1 K ( m n , k + m ¯ R , k + m ˜ R , k ) k = 1 K ( m ¯ R , k + m ˜ R , k ) ]
m ˜ D = μ [ k = 1 K ( m ¯ D , k + m ˜ D , k ) k = 1 K ( m ¯ D , k ) ]
m ¯ H = μ [ k = 1 K ( m ¯ D , k ) ]
μ = [ n = 1 N k = 1 K ( m n , k + m ¯ D , k + m ˜ D , k ) ( N 1 ) k = 1 K ( m ¯ D , k + m ˜ D , k ) ] 1
β n = m n 1 m ¯ D
β D = m ˜ D 1 m ¯ D
where βn represents the belief for the nth scale ( n = 1 , 2 , , N ) , and βD represents the belief for the incomplete information.
Specifically, the calculation of βn can be translated into the following Equations (12) and (13) by integrating Equations (2)–(11),
β n = μ [ k = 1 K ( w k β n , k + 1 w k n = 1 N β n , k ) k = 1 K ( 1 w k n = 1 N β n , k ) ] 1 μ [ k = 1 K ( 1 w k ) ]
μ = [ n = 1 N k = 1 K ( w k β n , k + 1 w k n = 1 N β n , k ) ( N 1 ) k = 1 K ( 1 w k n = 1 N β n , k ) ] 1
where wk and βn,k share the same meaning as in Equations (2)–(11).

3. Method

3.1. Framework

By adopting the DE algorithm as the optimization engine, the EO-BRB approach (see Figure 1) with six steps is proposed as follows:
  • Step 1: Divide the initial dataset into the training dataset and the testing dataset;
  • Step 2: Determine the parameters for further optimization, and modify the DE algorithm to fit the needs of specific requirements;
  • Step 3: Calculate the MSE of the training dataset;
  • Step 4: Identify the best trained model with the smallest MSE;
  • Step 5: Construct the new BRB system using the identified parameters including the key referenced values of the attributes;
  • Step 6: Calculate the MSE of the testing dataset for efficiency validation.

3.2. The Optimization Model

The output of the ER algorithm as in Equation (14) is in the same belief structure as that of the input rules.
S = { ( D n , β n ) , n = 1 , , N }
Consider that the utility of the nth scale u(Dn) [22], and then the numerical inference output can be calculated by Equation (15)
f ^ ( ) = n = 1 N u ( D n ) β n
Since MSE is used to measure the difference between the actual output, f ( ) , and the inference output, f ^ ( ) , the MSE can be modeled as in Equation (16),
M S E = 1 T t = 1 T ( f ( t ) f ^ ( t ) ) 2
where T denotes the size of the dataset. Note that the MSE function is also the optimization objective function in the DE algorithm.
With the consideration of the parameters to be optimized and the MSE model as in Equation (16), the objective function of the optimization model is given as in Equation (17) and the restraints are given in Equations (18)–(22):
min M S E ( A m k , θ k , β j , k )
s. t.
l b m A m k u b m , k = 1 , , K , m = 1 , , M
A m 1 = l b m
A m K = u b m
0 θ k 1
0 β n , k 1 , n = 1 , , N
where Equation (18) describes that the referenced values should be bounded by the lower and upper bounds of the mth attribute, lbm and ubm. Equations (19) and (20) describe that, among all K rules, the minimum/maximum of the referenced rules of the ith attribute should be the lower/upper bound of the ith attribute. Equation (21) describes that the initial weight of any rule, θk, should be within the range of 0–1. Equation (22) describes that the belief of any scale in any rule, βn,k, should be within the range of 0–1.

3.3. The Optimization Algorithm

As an effective EA algorithm, the differential evolutionary (DE) algorithm is used as the optimization engine for solving the constructed optimization model in Section 3.2. The DE algorithm was first proposed to solve the Chebyshev problem. Then, it was widely used because of the simple mechanism and the high quality of the optimization results [28]. In this study, the classical DE (still referred to as DE) is applied with the following six steps.
  • Step 1: Initialization.
In the initialization step, the initial population is generated with each individual representing a BRB. Moreover, the parameters are also determined, including the number of the individuals, the number of the generations, etc.
  • Step 2: Mutation/Crossover.
The mutation step aims to generate new individuals. The mutation for an individual xi in the population {xi, I = 1, …, NP} is shown in Equation (23)
v i = x r 1 + F ( x r 2 x r 3 )
where, x r 1 , x r 2 and x r 3 are three different individuals, F = 0.5 is an appropriate choice.
In the crossover step, two individuals exchange several genes and a new individual is generated, as shown in Equation (24).
u i ( j ) = { v i ( j ) if r C R x i ( j ) , otherwise
where r is a random number between 0 and 1. CR ∈ [0, 1] is the crossover probability which is used to control where the mutated value should be applied in the decision variables. Normally, CR = 0.9 is an appropriate choice.
  • Step 3: Fitness calculation.
The fitness function of DE produces the objective the optimization objective. Thus, the calculation procedures of the fitness function are Equations (14)–(16) in Section 3.2.
  • Step 4: Selection.
For an individual ui generated in the mutation and the crossover steps, the selected individual is determined by Equation (25):
x i = { u i if f ( u i ) f ( x i ) x i otherwise
where f ( ) is the objective function and x i is the selected individual.
  • Step 5: Stop check.
The stop criterion is the preset number of generations. If it has met the stop criterion, i.e., the maximum number of generations, then, stop and go to Step 6. If it has not met the stop criterion, then go to Step 3 for a new round of optimization.
  • Step 6: Output
Produce the final output.
By repeating this process, the population is updated and the solution of the problem is optimized. When the stop criterion is met, the loop could be stopped.

4. Case Study

Three cases are studied to validate the efficiency of the EO-BRB approach. Case I is based a multi-extremal function [22]. Case II is to predict the residual life probability of the metalized film capacitor for its reliability analysis [23,24]. Case III studies the pipeline leak detection which has been studied under multiple conditions [14,15,16,17,18,19,20,21]. Since DE would inevitably bring randomness, all experiments are conducted in 30 runs. Moreover, the mean square error (MSE) is adopted as the metric for the performance evaluation for maintaining consistency with previous studies.

4.1. Case I: Multi-Extremal Function

A multi-extremal function case in Reference [22] is given as in Equation (26),
g ( x ) = e ( x 2 ) 2 + 0.5 e ( x + 2 ) 2 , 5 x 5
In [22], a BRB system with five rules is constructed. We apply similar settings as that of [22] by supposing that there are five scales in the consequences with the utilities as {D1, D2, D3, D4, D5} = {−0.5, 0, 0.5, 1, 1.5}. Since the DE algorithm is applied, there is no need to locate the initial BRB. The settings are that (1) BRB is still assumed with 5 rules, (2) 20 individuals, and (3) 2000 generations. The objective is to minimize the MSE between the actual and the predicted outputs. The training/testing dataset is the same as that in [22] for maintaining consistency and conducting a fair comparison of 101 sets of data which are uniformed sampled from the interval [−5, 5].
Figure 2 shows the comparison of two approaches. Table 1 shows the derived BRB. The authors in [22] calculated the MSE between the actual output and the optimized output for the 101 sets of data and it reached 6.32 × 10−5. Comparatively, the MSE between the actual output and the predicted output by the BRB in Table 1 is 2.55 × 10−5, which is 59.65% less than [22].

4.2. Case II: Residual Life Probability Prediction for the Metalized Film Capacitor

The metallized film capacitors (MFCs) are critical parts of the inertial confinement fusion device which uses the high-power laser for the thermonuclear fusion under laboratory conditions [33]. It is necessary to conduct the reliability evaluation of MFC to make maintenance policies for the energy systems as well as for the high-power laser devices themselves, which in essence is to predict the MFC’s residual life or the residual life probability. For each 100 s, the residual life probability is predicted using a deduction process with the consideration of all influential factors at that time. Ref. [24] proposed a generic process to model this deduction process and predict the residual life probability of MFC, which derived the residual life probability inverse-Gauss function as in Equation (27).
f ( t | μ , σ , y ) = l 2 π σ 2 t 3 exp ( ( l μ t ) 2 2 σ 2 t )
where l =245, μ = 0.0001195, and σ = 0.0057. The MSE between the real-time probability and the predicted probability is 1.31 × 10−2.
To apply the EO-BRB approach, the settings are that (1) the BRB is with 5 rules, (2) 20 individuals, and (3) 2000 generations. A total of 250 sets of data (which is half of the total dataset) are used as the training dataset and the total dataset is used as the testing dataset, which is the same as the prior study [24] for maintaining consistency and conducting a fair comparison. Additionally, suppose that there are five scales in the consequences with the utilities as {D1, D2, D3, D4, D5} = {0, 2, 4, 6, 8}.
After using the EO-BRB approach, a BRB with five rules are constructed, as presented in Table 2. Furthermore, upon observing the BRB with five rules in Table 2, it is found that the last rule could be combined with the first rule and the BRB with four rules is derived as in Table 2. It is obvious that the BRBs in Table 2 are and Table 3 equivalent. Consequently, the BRB in Table 3 reaches the MSE of 9.09 × 10−3. Compared with the MSE = 1.31 × 10−2 by [24], the MSE by the BRB with four rules is 30.42% less. Figure 3 shows the comparison of two approaches.

4.3. Case III: Pipeline Leak Detection

4.3.1. Background

The pipeline leak detection case has been extensively studied [14,15,16,17,18,19,20,21] where two factors, namely PressureDiff and Flowdiff, were used to predict pipeline leaksize. Of all the previous studies regarding on the pipeline leak detection case, there are two basic scenarios. For Scenario I, 500 sets of the total dataset is taken as the training dataset while the total dataset is taken as the testing dataset. This training/testing datasets division is the same as in [14,15,16,17,18]. For Scenario II, it is specifically based on [19,20,21] where a partial of 322 sets of data were selected from the total dataset. Of the selected 322 sets of data, the first 305 sets of data were used as the training dataset and the remaining 17 sets of data as the testing dataset. However, we do not have the same dataset of 322 sets of data as [19,20,21]. In this scenario, we select six datasets with the starting points ranging from 650th to 900th of the total dataset, and the next 17 sets of data are used as the testing dataset. Be reminded again that the training/testing datasets division in Scenarios I/II is the same as in prior studies for maintaining consistency and conducting a fair comparison.

4.3.2. Scenario I

In previous studies, 500 sets of data were randomly selected from the total dataset as the training dataset (compared with 500 sets of data in [14,15] and 800/900 sets of data in [16,17,18] from a specific time interval). The total dataset is used as the testing dataset. To maintain consistency, the settings for EO-BRB are that: (1) BRB is assumed with 3 rules, (2) 20 individuals, and (3) 500 generations, whilst (4) 5 scales are assumed for leaksize as the output of BRB with the utilities as {D1, D2, D3, D4, D5} = {−3, 0, 3, 6, 9}.
The derived BRB is shown in Table 3, and Figure 4 shows the actual/predicted output/error. The MSE of the BRB as in Table 4 reaches 5.49 × 10−2 which is far smaller than 7.78 × 10−1 in [16] as it is the best result to date. Additionally, the MAE of the BRB system as in Table 4 reaches 1.35 × 10−1 which is also small than 2.06 × 10−1 in [15].

4.3.3. Scenario II

Scenario II is based on [19,20,21], where 322 sets of data were selected from the total dataset among which the first 305 sets of data become the training dataset and the rest 17 sets of data become the testing dataset. To execute a fair comparison with [19,20,21], we selected six independent datasets with similar settings. These datasets are of 322 sets of data which are selected with the start points of the 650th, 700th, 750th, 800th, 850th, and 900th in the total dataset, respectively. The settings are that (1) the BRB is with 3 rules, (2) 20 individuals, and (3) 500 generations. Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10 shows the corresponding BRBs. Figure 5 shows the actual/predicted outputs and the errors of the EO-BRB approach. The errors in the six settings are 7.42 × 10−4, 1.45 × 10−2, 1.34 × 10−2, 3.54 × 10−2, 2.16 × 10−2, and 2.79 × 10−3, respectively. In other words, there is an improvement of 38.84% since the average MSE of six settings is 1.47 × 10−2 by EO-BRB compared with 2.41 × 10−2 in [19].

4.4. Summary of Results and Discussion

Table 11 summarizes the statistics of traditional studies and this study. For most cases and scenarios, the number of rules and the number of the parameters to be optimized are reduced (especially in Case 3). More importantly, MSEs in these cases/scenarios are reduced as well, which validates the efficiency of the EO-BRB approach.
Table 12 further explores the efficiency of DE as the optimization engine. It shows that the variances of the MSEs in 30 runs of all cases/scenarios are far smaller compared with the average/median of MSEs. It shows that there are very stable results using DE, which further proves the validity of the application of the DE algorithm.
According to the results in Table 11 and Table 12, there are many advantages to the proposed EO-BRB approach. (1) First, EO-BRB can effectively improve the accuracy of BRB in multiple numerical and practical conditions. (2) Moreover, only 3–5 rules are assumed in BRBs in Cases I–III, which is very cost-effective compared with other approaches that may require many more parameters. Admittedly, there are also challenges that still need to be addressed. (1) The primary concern is overfitting. Although only 3–5 rules are assumed in BRBs in this study, more rules can be used for improving the accuracy of BRB which, however, could result in overfitting. (2) Performance of EO-BRB may be affected by the selected evolutionary algorithm. Therefore, it is a challenging task to determine the optimization engine as well as its related parameters.

5. Conclusions

The EO-BRB approach is proposed by identifying the most representative referenced values of the attributes. Comparatively, the traditional learning approaches focus more on better approximating the practical systems, which contributes as the main difference between the proposed EO-BRB approach and the traditional learning approaches. Besides, the EO-BRB approach uses the DE algorithm as the optimization engine. Case study results validate the efficiency of the EO-BRB approach:
(1)
No need to transform numerical values into linguistic terms. The inclusion of the referenced values of the attributes as the parameters to be optimized makes the transformation process unnecessary. In addition, the DE algorithm is very effective in dealing with continuous parameters (and can also be adapted to deal with discretized parameters). Therefore, there is no need to transform numerical values into linguistic terms, which is a common step in traditional approaches. Without this transformation, the EO-BRB approach is less prejudicial from human intervention.
(2)
No need to pre-determine the initial solution. The DE algorithm has been proven to be very effective in finding a global optimal solution with randomly generated initial solutions. Comparatively, the traditional approaches need to pre-determine the initial solution based on certain information from the initial dataset.
(3)
No need for a complicated mathematical deduction process. The evolutionary mechanism in finding the optimal solution is inherited within DE, which means that this approach is universal and general. There is no need to calculate the ladder direction and/or the derivative function which are required by a traditional learning approach.
(4)
Furthermore, EO-BRB can help downsize the BRB system while still preserving excellent approximation with practical system. The results of three case studies show that not only the BRB systems are downsized or at least maintained as the same size, but also even smaller MSEs than the results in existed literatures are achieved.
For future work, two directions are worthy of paying attention to. Other variants of the DE algorithm or other evolutionary algorithms (GA, PSO, ACO, etc.) may be a better choice for other application cases. However, this is not the main topic of this study and is worthy of more effort. Another direction is to test on more benchmarks. The cases in this paper, especially the pipeline leak detection case, are chosen because they have been studied in the literature regarding on the learning frameworks and approaches for the BRB system. Then, more benchmark cases from other backgrounds should be further studied because only in this way can the EO-BRB approach be improved.

Author Contributions

Writing—original draft preparation, L.C. and F.S.; writing—review and editing, Z.Q. and L.C.; supervision, L.C.; programming, Z.Q., X.X. and F.S.; validation, Z.Q., F.S., X.X. and J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Funding of State Key Laboratory of Complex Eletromagnetic Environment Effects on Electronics and Information System (CEMEE2022Z0301B), the National Natural Science Foundation of China (Grant No. 61903108, 52171352), Zhejiang Province Key R&D projects (2021C03142), National Natural Science Foundation of Zhejiang Province (Grant No. LY21F030011), and Zhejiang Province Public Welfare Technology Application Research Projects (LGF21F020013).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon on request.

Acknowledgments

The authors gratefully thank Xiaobin Xu from the School of Automation, Hangzhou Dianzi University, for providing constructive suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, J.B.; Liu, J.; Wang, J.; Sii, H.S.; Wang, H.W. Belief rule-base inference methodology using the evidential reasoning approach-RIMER. IEEE Trans. Syst. Man Cybern. Syst. Hum. 2006, 36, 266–285. [Google Scholar] [CrossRef]
  2. Zhou, Z.J.; Hu, G.Y.; Hu, C.H.; Wen, C.L.; Chang, L.L. A Survey of Belief Rule-Base Expert System. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 4944–4958. [Google Scholar] [CrossRef]
  3. Chang, L.; Zhang, L.; Fu, C.; Chen, Y.W. Transparent digital twin for output control using belief rule base. IEEE Trans. Cybern. 2021, 2021, 1–15. [Google Scholar] [CrossRef] [PubMed]
  4. Sen, M.K.; Dutta, S.; Kabir, G. Development of flood resilience framework for housing infrastructure system: Integration of best-worst method with evidence theory. J. Clean. Prod. 2021, 290, 125197. [Google Scholar] [CrossRef]
  5. Yang, J.B.; Xu, D.L. Evidential Reasoning rule for evidence combination. Artif. Intell. 2013, 205, 1–29. [Google Scholar] [CrossRef]
  6. Chang, L.; Zhang, L.; Xu, X. Correlation-oriented complex system structural risk assessment using Copula and belief rule base. Inf. Sci. 2021, 564, 220–236. [Google Scholar] [CrossRef]
  7. Fu, C.; Hou, B.; Xue, M.; Chang, L.; Liu, W. Extended Belief Rule-Based System with Accurate Rule Weights and Efficient Rule Activation for Diagnosis of Thyroid Nodules. IEEE Trans. Syst. Man Cybern. Syst. 2022, 2022, 1–13. [Google Scholar] [CrossRef]
  8. Dymova, L.; Kaczmarek, K.; Sevastjanov, P. An extension of rule base evidential reasoning in the interval-valued intuitionistic fuzzy setting applied to the type 2 diabetes diagnostic. Expert Syst. Appl. 2022, 201, 117100. [Google Scholar] [CrossRef]
  9. Cao, Y.; Zhou, Z.J.; Hu, C.H.; Tang, S.W.; Wang, J. A new approximate belief rule base expert system for complex system modelling. Decis. Support Syst. 2021, 150, 113558. [Google Scholar] [CrossRef]
  10. Zhao, D.; Zhou, Z.; Tang, S.; Cao, Y.; Wang, J.; Zhang, P.; Zhang, Y. Online estimation of satellite lithium-ion battery capacity based on approximate belief rule base and hidden Markov model. Energy 2022, 256, 124632. [Google Scholar] [CrossRef]
  11. Jia, Q.; Hu, J.; Zhang, W. A fault detection method for FADS system based on interval-valued neutrosophic sets, belief rule base, and DS evidence reasoning. Aerosp. Sci. Technol. 2021, 114, 106758. [Google Scholar] [CrossRef]
  12. Chang, L.L.; Zhou, Y.; Jiang, J.; Li, M.J.; Zhang, X.H. Structure learning for belief rule base expert system: A comparative study. Knowl. Based Syst. 2013, 39, 159–172. [Google Scholar] [CrossRef]
  13. Yang, J.B.; Liu, J.; Xu, D.L.; Wang, J.; Wang, Y.M. Optimization models for training belief-rule-based systems. IEEE Trans. Syst. Man Cybern. A Syst. Hum. 2007, 37, 569–585. [Google Scholar] [CrossRef]
  14. Zhu, H.; Xiao, M.; Yang, L.; Tang, X.; Liang, Y.; Li, J. A minimum centre distance rule activation method for extended belief rule-based classification systems. Appl. Soft Comput. 2020, 91, 106214. [Google Scholar] [CrossRef]
  15. Chen, Y.W.; Yang, J.B.; Xu, D.L.; Zhou, Z.J.; Tang, D.W. Inference analysis and adaptive training for belief rule based systems. Expert Syst. Appl. 2011, 38, 12845–12860. [Google Scholar] [CrossRef]
  16. Diao, H.; Lu, Y.; Deng, A.; Zou, L.; Li, X.; Pedrycz, W. Convolutional rule inference network based on belief rule-based system using an evidential reasoning approach. Knowl. Based Syst. 2022, 237, 107713. [Google Scholar] [CrossRef]
  17. Yang, L.H.; Wang, S.; Ye, F.F.; Liu, J.; Wang, Y.M.; Hu, H. Environmental investment prediction using extended belief rule-based system and evidential reasoning rule. J. Clean. Prod. 2021, 289, 125661. [Google Scholar] [CrossRef]
  18. Zhenjie, Z.; Xiaobin, X.; Peng, C.; Xudong, W.; Xiaojian, X.; Guodong, W. A novel nonlinear causal inference approach using vector-based belief rule base. Int. J. Intell. Syst. 2021, 36, 5005–5027. [Google Scholar] [CrossRef]
  19. Zhou, Y.; Chang, L.; Qian, B. A belief-rule-based model for information fusion with insufficient multi-sensor data and domain knowledge using evolutionary algorithms with operator recommendations. Soft Comput. 2019, 23, 5129–5142. [Google Scholar] [CrossRef]
  20. Kabir, S.; Islam, R.U.; Hossain, M.S.; Andersson, K. An integrated approach of belief rule base and deep learning to predict air pollution. Sensors 2020, 20, 1956. [Google Scholar] [CrossRef]
  21. Xu, X.; Zhao, Z.; Xu, X.; Yang, J.; Chang, L.; Yan, X.; Wang, G. Machine learning-based wear fault diagnosis for marine diesel engine by fusing multiple data-driven models. Knowl. Based Syst. 2020, 190, 105324. [Google Scholar] [CrossRef]
  22. Chen, Y.W.; Yang, J.B.; Xu, D.L.; Yang, S.L. On the inference and approximation properties of belief rule based systems. Inf. Sci. 2013, 234, 121–135. [Google Scholar] [CrossRef]
  23. Wang, X.L.; Guo, B.; Cheng, Z.J. Residual life estimation based on bivariate Wiener degradation process with time-scale transformations. J. Stat. Comput. Sim. 2014, 84, 545–563. [Google Scholar] [CrossRef]
  24. Liu, S.; Zhang, J.; Zhang, Z. Lifetime Prediction of Mica Paper Capacitors Based on an Improved Iterative Grey-Markov Chain Model. IEEE J. Emerg. Sel. Top. Power Electron. 2022, 2022, 1. [Google Scholar] [CrossRef]
  25. Srorn, R.; Price, K. Differential Evolution—A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar]
  26. Yang, J.Q.; Chen, C.H.; Li, J.Y.; Liu, D.; Li, T.; Zhan, Z.H. Compressed-Encoding Particle Swarm Optimization with Fuzzy Learning for Large-Scale Feature Selection. Symmetry 2022, 14, 1142. [Google Scholar] [CrossRef]
  27. Fu, B.J.; Zhang, J.P.; Li, W.J.; Zhang, M.J.; He, Y.; Mao, Q.J. A Differential Evolutionary Influence Maximization Algorithm Based on Network Discreteness. Symmetry 2022, 14, 1397. [Google Scholar] [CrossRef]
  28. Das, S.; Suganthan, P.N. Differential Evolution: A survey of the state-of-the-art. IEEE Trans. Evol. Comput. 2011, 15, 4–31. [Google Scholar] [CrossRef]
  29. Lv, C.; Liu, J.; Zhang, Y.; Lei, W.; Cao, R.; Lv, G. Reliability modeling for metallized film capacitors based on time-varying stress mission profile and aging of ESR. IEEE J. Emerg. Sel. Top. Power Electron. 2020, 9, 4311–4319. [Google Scholar] [CrossRef]
  30. Mirjalili, S. Evolutionary Algorithms and Neural Networks. Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2019; p. 780. [Google Scholar]
  31. Bui, D.T.; Panahi, M.; Shahabi, H.; Singh, V.P.; Shirzadi, A.; Chapi, K.; Ahmad, B.B. Novel hybrid evolutionary algorithms for spatial prediction of floods. Sci. Rep. 2018, 8, 15364. [Google Scholar] [CrossRef]
  32. Fu, C.; Xu, C.; Xue, M.; Liu, W.; Yang, S. Data-driven decision making based on evidential reasoning approach and machine learning algorithms. Appl. Soft Comput. 2021, 110, 107622. [Google Scholar] [CrossRef]
  33. Belko, V.O.; Emelyanov, O.A.; Ivanov, I.O.; Plotnikov, A.P. Self-Healing Processes of Metallized Film Capacitors in Overload Modes—Part 1: Experimental Observations. IEEE Trans. Plasma Sci. 2021, 49, 1580–1587. [Google Scholar] [CrossRef]
Figure 1. Framework of the EO-BRB approach.
Figure 1. Framework of the EO-BRB approach.
Symmetry 14 01622 g001
Figure 2. Comparative results for Case I: (a) compares the errors of [22] and EO-BRB while (b) and (c) are the outputs of [22] and EO-BRB, respectively.
Figure 2. Comparative results for Case I: (a) compares the errors of [22] and EO-BRB while (b) and (c) are the outputs of [22] and EO-BRB, respectively.
Symmetry 14 01622 g002
Figure 3. Comparative results for Case II: (a,b) give outputs of the previous study and EO-BRB while (c,d) give the errors.
Figure 3. Comparative results for Case II: (a,b) give outputs of the previous study and EO-BRB while (c,d) give the errors.
Symmetry 14 01622 g003
Figure 4. Comparative results for Case III in Scenario I: (a) and (b) give the outputs and errors of EO-BRB, respectively.
Figure 4. Comparative results for Case III in Scenario I: (a) and (b) give the outputs and errors of EO-BRB, respectively.
Symmetry 14 01622 g004
Figure 5. Comparative results for Case III/Scenario II with different starting points.
Figure 5. Comparative results for Case III/Scenario II with different starting points.
Symmetry 14 01622 g005
Table 1. BRB with five rules for Case I.
Table 1. BRB with five rules for Case I.
No.Rule WeightxBelief Degrees on {D1, D2, D3, D4, D5}
10.9311−5(0.3425, 0.3998, 0.1912, 0.0574, 0.0091)
20.3679−2.0710(0.2160, 0.2919, 0.0002, 0.2482, 0.2436)
30.4216−0.0821(0.5043, 0.1990, 0.0359, 0.2587, 0.0021)
40.28582.0415(0.1153, 0.0121, 0.0004, 0.5008, 0.3714)
50.74925(0.5016, 0.0125, 0.4854, 0.0001, 0.0005)
Table 2. BRB with five rules for Case II.
Table 2. BRB with five rules for Case II.
No.Rule WeighttBelief Degrees on {D1, D2, D3, D4, D5}
11 t < 7100 (1, 0, 0, 0, 0)
20.2769 t = 7100 (0.9639, 0, 0, 0, 0.0361)
30.3303 t = 17,323 (0.1942, 0, 0, 0, 0.8058)
40.8781 t = 49,899 (0.9893, 0, 0, 0, 0.0107)
50.0137 t = 50,000 (0.0035, 0.0018, 0.0012, 0.4726, 0.5210)
Table 3. BRB with four rules for Case II.
Table 3. BRB with four rules for Case II.
No.Rule WeighttBelief Degrees on {D1, D5}
11 ( t < 7100 ) ( t > 49,899 ) (1, 0)
20.2769 t = 7100 (0.9639, 0.0361)
30.3303 t = 17,323 (0.1942, 0.8058)
40.8781 t = 49,899 (0.9893, 0.0107)
Table 4. BRB with three rules for Scenario I of Case III.
Table 4. BRB with three rules for Scenario I of Case III.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.7829−10.0000−0.0200(0.1549, 0.0051, 0.0246, 0.0112, 0.8041)
20.0820−0.37500.0157(0.9984, 0.0010, 0.004, 0.0001, 0.0001)
30.00172.00000.0400(0.4503, 0.3094, 0.1050, 0.0484, 0.0868)
Table 5. BRB with three rules for Scenario II of Case III: start point 650.
Table 5. BRB with three rules for Scenario II of Case III: start point 650.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.9882−10.0000−0.0200(0.0113, 0.0949, 0.1061, 0.3867, 0.4009)
20.0756−1.3128−0.0094(0.9960, 0.0027, 0.0003, 0.0004, 0.0006)
30.00142.00000.0400(0.7592, 0.0347, 0.1559, 0.0264, 0.0239)
Table 6. BRB with three rules for Scenario II of Case III: start point 700.
Table 6. BRB with three rules for Scenario II of Case III: start point 700.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.9157−10.0000−0.0200(0.0069, 0.1781, 0.0120, 0.2432, 0.5598)
20.0801−1.13630.3318(0.9991, 0.0007, 0.0002, 0, 0)
30.00032.00000.0400(0.4749, 0.1902, 0.1932, 0.1159, 0.0257)
Table 7. BRB with three rules for Scenario II of Case III: start point 750.
Table 7. BRB with three rules for Scenario II of Case III: start point 750.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.8921−10.0000−0.0200(0.0297, 0.0802, 0.1340, 0.3228, 0.4334)
20.1050−1.1886−0.0149(0.9997, 0.0002, 0, 0, 0.0001)
30.000012.00000.0400(0.7490, 0.0043, 0.0541, 0.0937, 0.1015)
Table 8. BRB with three rules for Scenario II of Case III: start point 800.
Table 8. BRB with three rules for Scenario II of Case III: start point 800.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.9040−10.0000−0.0200(0.0888, 0.0640, 0.0104, 0.2698, 0.5669)
20.0006−3.4693−0.0109(0.0015, 0.0040, 0.0050, 0.0034, 0.9861)
30.05702.00000.0400(0.9983, 0.0002, 0.0012, 0.0001, 0.0002)
Table 9. BRB with three rules for Scenario II of Case III: start point 850.
Table 9. BRB with three rules for Scenario II of Case III: start point 850.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.0955−10.0000−0.0200(0.0194, 0.0560, 0.1336, 0.4007, 0.3904)
20.0147−4.82610.0349(0.1648, 0.1095, 0.0593, 0.0416, 0.6249)
30.93122.00000.0400(0.9881, 0.0012, 0.0007, 0, 0)
Table 10. BRB with three rules for Scenario II of Case III: start point 900.
Table 10. BRB with three rules for Scenario II of Case III: start point 900.
No.Rule WeightFlowDiffPressureDiffBelief Degrees on {D1, D2, D3, D4, D5}
10.3893−10.0000−0.0200(0.0115, 0.0395, 0.2883, 0.0946, 0.5662)
20.0180−4.26860.0228(0.1920, 0.1318, 0.3311, 0.3073, 0.0377)
30.89282.00000.0400(0.9982, 0.0002, 0.0010, 0.0005, 0.0001)
Table 11. Statistics on the efficiency of the EO-BRB approach.
Table 11. Statistics on the efficiency of the EO-BRB approach.
Case 1Case 2Case 3: Scenario I
Previous studiesNo. rules543
No. parameters332620
MSE2.55E-059.07E-035.49E-02
EO-BRBNo. rules5 56
No. parameters33 336
MSE6.32E-051.31E-027.78E-01
Improvement59.65%30.76%92.94%
Case 3: Scenario II (Start points)
650700750800850900
Previous studiesNo. rules3
No. parameters20
MSE7.42E-041.45E-021.34E-023.54E-022.16E-022.79E-03
EO-BRBNo. rules5
No. parameters30
MSE2.41E-02
Improvement38.84%
Table 12. Statistics on the efficiency of the application of the DE algorithm.
Table 12. Statistics on the efficiency of the application of the DE algorithm.
Case 1Case 2Case 3: Scenario ICase 3: Scenario II (Start Points)
650700750800850900
MSE_average3.84E-059.07E-035.64E-026.44E-031.68E-021.42E-023.86E-022.29E-023.47E-03
MSE_median3.88E-059.07E-035.63E-026.60E-041.68E-021.40E-033.88E-022.32E-023.45E-03
MSE_variance5.56E-111.06E-108.54E-073.71E-066.34E-072.63E-071.10E-063.24E-061.93E-07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qi, Z.; Chang, L.; Shi, F.; Xu, X.; Feng, J. Evolutionary Optimization for the Belief-Rule-Based System: Method and Applications. Symmetry 2022, 14, 1622. https://doi.org/10.3390/sym14081622

AMA Style

Qi Z, Chang L, Shi F, Xu X, Feng J. Evolutionary Optimization for the Belief-Rule-Based System: Method and Applications. Symmetry. 2022; 14(8):1622. https://doi.org/10.3390/sym14081622

Chicago/Turabian Style

Qi, Zongfeng, Leilei Chang, Fan Shi, Xiaojian Xu, and Jing Feng. 2022. "Evolutionary Optimization for the Belief-Rule-Based System: Method and Applications" Symmetry 14, no. 8: 1622. https://doi.org/10.3390/sym14081622

APA Style

Qi, Z., Chang, L., Shi, F., Xu, X., & Feng, J. (2022). Evolutionary Optimization for the Belief-Rule-Based System: Method and Applications. Symmetry, 14(8), 1622. https://doi.org/10.3390/sym14081622

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop