Next Article in Journal
Collective Action and Social Innovation in the Energy Sector: A Mobilization Model Perspective
Previous Article in Journal
Optimization of Cold Pressing Process Parameters of Chopped Corn Straws for Fuel
Previous Article in Special Issue
A Test of Using Markov-Switching GARCH Models in Oil and Natural Gas Trading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble of Regression-Type and Interpolation-Type Metamodels

1
School of Aerospace Engineering, Xiamen University, Xiamen 361005, China
2
School of Energy and Power Engineering, Beihang University, Beijing 100191, China
3
Army Aviation Institute, Beijing 100000, China
4
AECC Hunan Aviation Powerplant Research Institute, Zhuzhou 412002, China
*
Authors to whom correspondence should be addressed.
Energies 2020, 13(3), 654; https://doi.org/10.3390/en13030654
Submission received: 21 December 2019 / Revised: 20 January 2020 / Accepted: 21 January 2020 / Published: 4 February 2020
(This article belongs to the Special Issue Intelligent Optimization Modelling in Energy Forecasting)

Abstract

:
Metamodels have become increasingly popular in the field of energy sources because of their significant advantages in reducing the computational cost of time-consuming tasks. Lacking the prior knowledge of actual physical systems, it may be difficult to find an appropriate metamodel in advance for a new task. A favorite way of overcoming this difficulty is to construct an ensemble metamodel by assembling two or more individual metamodels. Motivated by the existing works, a novel metamodeling approach for building the ensemble metamodels is proposed in this paper. By thoroughly exploring the characteristics of regression-type and interpolation-type metamodels, some useful information is extracted from the feedback of the regression-type metamodels to further improve the functional fitting capability of the ensemble metamodels. Four types of ensemble metamodels were constructed by choosing four individual metamodels. Common benchmark problems are chosen to compare the performance of the individual and ensemble metamodels. The results show that the proposed metamodeling approach reduces the risk of selecting the worst individual metamodel and improves the accuracy of the used individual metamodels.

Graphical Abstract

1. Introduction

Metamodels, which are also referred to as surrogate models, are essentially approximate mathematical models of real physical systems. In the past decade, metamodels have become increasingly popular in the field of energy sources because of their significant advantages in reducing the computational cost of time-consuming tasks [1,2]. Melo et al. [3] pointed out that researchers in many countries are developing metamodels to estimate the energy performance of the building stock. Bornatico et al. [4] used a kind of metamodel to optimize energy systems, and found that the metamodel converged to the same solution at 150 times the speed of the fine model. Westermann and Evins [5] summarized and discussed recent studies on the application of metamodels in sustainable building design. Ferrero Bermejo et al. [6] reviewed and compared two typical metamodels, namely the artificial neural networks and the support vector machine, for energy forecasting and condition-based maintenance in PV plants.
Actually, a good metamodel mainly depends on its accuracy and generality for different design tasks. To enhance the performance of metamodels, researchers have carried out a lot of studies over the past few decades [7,8,9,10,11]. As a result, a large number of metamodels have been proposed, of which several types have gained wide acceptance in various applications. They are polynomial response surface (PRS) [12,13,14], support vector regression (SVR) [15,16,17], radial basis functions (RBF) [18,19], extended radial basis functions (E-RBF) [20], moving least squares (MLS) [21], artificial neural networks (ANN) [22,23], multivariate adaptive regressive splines (MARS) [24] and Kriging (KRG) [25,26]. These different metamodels give us more options for different tasks. However, lacking the prior knowledge of the actual physical systems, it is challenging to find a suitable metamodel in advance for a new task. In particular, the worst metamodel may be chosen for the task.
A simple way to overcome the difficulty is to build a series of metamodels based on a given training dataset at first, and then select the best one on the basis of some statistical techniques like the cross-validation method. Another favorite way is to construct an ensemble metamodel, which assembles two or more individual metamodels by introducing weight factors. The basic idea of such an ensemble metamodel can be traced back to 1990s [27,28], and currently it has become a research hotspot [8,29]. According to the characteristics of the weight factors, the techniques for building the ensemble metamodels can be mainly categorized into methods based on local errors, methods based on global errors, and methods based on regression.
In the first category, the weight factors ω i = ω i ( x ) are functions of design space, which are determined by the local errors of individual metamodels at the point of interest. Zerpa et al. [30] introduced a local weighted average model for the optimization of alkaline-surfactant-polymer flooding processes by using the prediction variances of three individual metamodels (PRS, KRG, and RBF). Sanchez, Pintos, and Queipo [31] proposed a general approach toward the ensemble of kernel-based models based on the local prediction variances. Acar [32] investigated the efficiency of methods based on the local errors, and developed a new approach to determine the weight factors by using the pointwise cross-validation errors instead of the prediction variances. Zhang, Chowdhury, and Messac [33] proposed a new metamodeling technique called adaptively hybrid functions, whose weight factors are determined based on the local measure of accuracy in the pertinent trust region. Lee and Choi [34] presented a new pointwise ensemble of metamodels, of which the weight factors are calculated by using the v nearest points cross-validation errors.
In the second category, the weight factors ( ω i = C i , x ) are constant values in the entire design space, which are determined by the global errors of individual metamodels. Goel et al. [35] studied a global weight factor selection approach based on the generalized mean square cross-validation errors (GMSE). Acar and Rais-Rohani [36] developed an accurate ensemble of metamodels by solving an optimization problem that minimizes GMSE or root mean square errors (RMSE). Viana, Haftka, and Steffen [37] obtained the optimal weight factors of the optimization problem by using the Lagrange multipliers. This method was also employed by Toal and Keane [38] to construct an ensemble of ordinary, universal, non-stationary and limit KRG models. Additionally, Acar [39] performed the simultaneous optimization of the weight factors and the shape parameters in the ensemble of RBFs.
It should be noted that in the first two categories the weight factors of individual metamodels are restricted to a positive range ω i > 0 and the sum of these factors is equal to 1 i = 1 M ω i = 1 . Since they are different from the first two categories, the techniques in the third category mainly use the regression methods (like least squares) to determine the weight factors. Accordingly, there is no longer any restriction on the weight factors, which may even have negative values. Polynkin and Toropov [40] introduced a novel mid-range metamodel assembly for the large-scale optimization problems, which is constructed based on the linear regression method. Ferreira and Serpa [41] developed an augmented least-square approach for creating the ensemble of metamodels, which can be extended to the efficient global optimization. Zhou and Jiang [42] constructed an ensemble of four individual metamodels (PRS, KRG, SVR, and RBF) from the view of the polynomial regression, and proposed a metamodel selection method on the basis of the stepwise regression to eliminate the redundant ones from the set of the candidate metamodels.
Motivated by these existing works, this paper proposes a different method for constructing the ensemble metamodels, which combines the advantages of regression-type and interpolation-type metamodels. The regression-type metamodels have better global trend fitting capacity than the interpolation-type metamodels, while the interpolation-type metamodels perform better than the regression-type metamodels in the vicinity of the sampling locations. By thoroughly exploring the characteristics of regression-type and interpolation-type metamodels, the proposed method could extract some useful information from the feedback of the regression-type metamodels to further improve the functional fitting capability of the ensemble metamodels.

2. Proposed Ensemble of Metamodels

2.1. Motivation and Basic Characteristics

The existing individual metamodels can be classified into regression-type and interpolation-type metamodels. The regression-type metamodels aim to fit the global trend of the underlying functions of the real physical systems in the entire design space, while the interpolation-type metamodels aim to achieve the local accuracy in the vicinity of the sampling locations. Accordingly, the regression-type metamodels can build smooth surfaces that pass across all the training points, while the interpolation-type metamodels can construct models that go through each training point. That is to say, for the regression-type metamodels there may be obvious deviations between the actual responses and the approximate responses at the sampling locations, while for the interpolation-type metamodels there is no deviation. These different characteristics make the two types of metamodels possess different advantages and limitations. For example: (i) the regression-type metamodels have better global trend fitting capacity than the interpolation-type metamodels, while (ii) the interpolation-type metamodels perform better than the regression-type metamodels in the vicinity of the sampling locations.
It should be noted that obtaining the training dataset required for constructing the metamodels may be time-consuming. Therefore, as much information as possible should be extracted from these data. However, for the regression-type metamodels, there are apparent deviations between the actual responses and the approximate responses at the sampling locations, from where some useful information may be still extracted to further improve the performance of these metamodels. Exploring the underlying knowledge of the training dataset and combining the characteristics of regression-type and interpolation-type metamodels, this paper proposes a novel metamodeling approach for the ensemble metamodels. The flowchart of the proposed metamodeling technique is shown in Figure 1, which involves four main steps as follows.
Step 1:
An appropriate design of experiment (DOE) should be first chosen to generate n sampling locations ( x 1 , x 2 , , x n ) , at where the actual responses ( y 1 , y 2 , , y n ) are obtained by conducting experiments or simulations. By using the initial training dataset ( x i , y i ) ( i = 1 , , n ) , a regression-type metamodel y ^ 1 ( x ) in Equation (1) is subsequently constructed to approximate the actual model y ( x ) .
y ^ 1 ( x ) y ( x ) , x = ( x 1 , x 2 , , x k ) T
where x denotes any point of interest.
Step 2:
We suppose that there is a deviation function y d ( x ) . It is obtained by subtracting the approximate model y ^ 1 ( x ) from the actual model y ( x ) .
y d ( x ) = y ( x ) y ^ 1 ( x )
Some useful information may be still extracted from the deviation function y d ( x ) . To approximate the deviation function, the training dataset should be updated. In detail, this paper first uses the established regression-type metamodel in Equation (1) to predict the approximate responses ( y ^ 1 1 , y ^ 1 2 , , y ^ 1 n ) at the initial sampling locations. Subsequently, the deviations ( y d 1 , y d 2 , , y d n ) between the actual responses and approximate responses at these locations are calculated as the updated training dataset.
( x 1 , y d 1 ) , ( x 2 , y d 2 ) , , ( x n , y d n ) = ( x 1 , y 1 y ^ 1 1 ) , ( x 2 , y 2 y ^ 1 2 ) , , ( x n , y n y ^ 1 n )
Step 3:
By using the updated training dataset in Equation (3), an interpolation-type metamodel y ^ 2 ( x ) in Equation (4) is constructed to approximate the deviation function y d ( x ) .
y ^ 2 ( x ) y d ( x )
Step 4:
Finally, the ensemble metamodel y ^ e n s ( x ) in Equation (5) is constructed by adding the established regression-type metamodel y ^ 1 ( x ) and interpolation-type metamodel y ^ 2 ( x ) together. By using Equations (1), (4) and (5), the established ensemble metamodel y ^ e n s ( x ) can be used to predict the response at any point of interest in the entire design space.
y ^ e n s ( x ) = y ^ 1 ( x ) + y ^ 2 ( x ) y ^ 1 ( x ) + y d ( x ) y ( x )

2.2. Detailed Modeling Process

To clearly illustrate the proposed metamodeling technique, this paper selects two common regression-type metamodels (PRS and SVR) and two popular interpolation-type metamodels, namely RBFM (RBF with multiquadric-form basis function) and RBFI (RBF with inverse multiquadric-form basis function). Accordingly, four types of ensemble metamodels can be obtained, which are PrsRbfm (Ensemble Scheme 1, ensemble of PRS and RBFM), PrsRbfi (Ensemble Scheme 2, ensemble of PRS and RBFI), SvrRbfm (Ensemble Scheme 3, ensemble of SVR and RBFM) and SvrRbfi (Ensemble Scheme 4, ensemble of SVR and RBFI). The detailed modeling processes of these involved metamodels are introduced as follows.

2.2.1. Step 1: Construction of Regression-Type Metamodels

PRS is a general designation of a series of polynomial regression functions, of which the most popular one is the second-order polynomial model. This paper adopts the second-order polynomial model y ^ 1 , p r s ( x ) , which can be written as
y ^ 1 , p r s ( x ) = z T β = β 0 + i = 1 k β i x i + i = 1 k j = i k β 2 j + i ( 2 k i + 1 ) 2 x i x j
where β = ( β 0 , β 1 , , β k 2 + 3 k 2 ) T denotes a coefficient vector, z = ( 1 , x 1 , x 2 , , x k 1 x k , x k x k ) T denotes a polynomial basis-function vector.
To estimate β , the regression problem in Equation (6) can be transformed as follows by using the initial training dataset.
y 1 y 2 y n = y d , p r s 1 y d , p r s 2 y d , p r s n + 1 x 1 1 x k 1 x 1 1 x 1 1 x k 1 1 x k 1 x k 1 x k 1 1 x 1 2 x k 2 x 1 2 x 1 2 x k 1 2 x k 2 x k 2 x k 2 1 x 1 n x k n x 1 n x 1 n x k 1 n x k n x k n x k n β 0 β 1 β k 2 + 3 k 2
where y d , p r s = ( y d , p r s 1 , y d , p r s 2 , , y d , p r s n ) T denotes the deviation vector.
Equation (7) can be also expressed as
y = X β + y d , p r s
According to the least squares method, β can be calculated as follows.
β = ( X T X ) 1 X T y
SVR is a regression function y ^ 1 , s v r ( x ) in the high-dimensional space, as shown in Equation (10).
y ^ 1 , s v r ( x ) = ω T ψ ( x ) + b
where ω denotes the weight vector, ψ ( x ) denotes the mapping function, and b denotes the bias.
To estimate ω and b, the regression problem in Equation (10) can be transformed as an optimization problem in Equation (11) by introducing ϵ -insensitive loss function.
min 1 2 | | ω | | 2 subject to ω T ψ ( x i ) + b y i ϵ y i ω T ψ ( x i ) b ϵ i = 1 , , n
To solve Equation (11), the regularization parameter, C > 0 , and the slack variables, ξ + ( i ) and ξ ( i ) , are introduced. In addition, Equation (12) can be obtained
min 1 2 | | ω | | 2 + C i = 1 n ( ξ + ( i ) + ξ ( i ) ) subject to ω T ψ ( x i ) + b y i ϵ + ξ + ( i ) y i ω T ψ ( x i ) b ϵ + ξ ( i ) ξ + ( i ) , ξ ( i ) 0 i = 1 , , n
The Lagrange dual model of Equation (12) can be expressed as
max 1 2 i , j = 1 n ( α + ( i ) α ( i ) ) ( α + ( j ) α ( j ) ) k x i , x j + i = 1 n ( α + ( i ) α ( i ) ) y i i = 1 n ( α + ( i ) + α ( i ) ) ϵ subject to i = 1 n ( α + ( i ) α ( i ) ) = 0 0 α + ( i ) , α ( i ) C i = 1 , , n
where α + ( i ) and α ( i ) denote the Lagrange multipliers, k x i , x j = ψ ( x i ) T ψ ( x j ) denotes a kernel function, which has several different forms. This paper chooses the Gaussian kernel function, which can be expressed as
k x , x i = exp ( γ | | x x i | | 2 )
According to Equation (13), α + ( i ) and α ( i ) can be first obtained. According to KKT conditions [43], ω and b can be then calculated.

2.2.2. Step 2: Update of Training Dataset

First, β calculated by Equation (9) can be used to substitute the one in Equation (6). Second, the approximate responses of established PRS ( y ^ 1 , p r s 1 , y ^ 1 , p r s 2 , , y ^ 1 , p r s n ) at the initial sampling locations ( x 1 , x 2 , , x n ) can be calculated according to Equation (6). Then, the updated training dataset of PRS can be expressed as
( x 1 , y d , p r s 1 ) , ( x 2 , y d , p r s 2 ) , , ( x n , y d , p r s n ) = ( x 1 , y 1 y ^ 1 , p r s 1 ) , ( x 2 , y 2 y ^ 1 , p r s 2 ) , , ( x n , y n y ^ 1 , p r s n )
Similarly, according to Equation (10), the updated training dataset of SVR can be obtained and expressed as
( x 1 , y d , s v r 1 ) , ( x 2 , y d , s v r 2 ) , , ( x n , y d , s v r n ) = ( x 1 , y 1 y ^ 1 , s v r 1 ) , ( x 2 , y 2 y ^ 1 , s v r 2 ) , , ( x n , y n y ^ 1 , s v r n )

2.2.3. Step 3: Construction of Interpolation-Type Metamodels

The general form of RBF can be expressed as
y ^ r b f ( x ) = i = 1 n λ i ϕ | | x x i | |
where λ i denotes an interpolation coefficient, r = | | x x i | | = ( x x i ) T ( x x i ) denotes the distance between points x and x i . ϕ ( r ) denotes a radially symmetric basis function, which has several different forms, such as:
  • Gaussian ϕ ( r ) = e ( r 2 / c 2 )
  • Multiquadric ϕ ( r ) = ( r 2 + c 2 ) 1 2
  • Inverse multiquadric ϕ ( r ) = ( r 2 + c 2 ) 1 2
  • Thin plate spline ϕ ( r ) = ( r 2 ) log ( r )
The interpolation coefficient λ i can be calculated by using the given training dataset ( x i , y i ) ( i = 1 , , n ) .
λ = A 1 y
where
λ = λ 1 , λ 2 , , λ n T A = ϕ | | x 1 x 1 | | , ϕ | | x 1 x 2 | | ϕ | | x 1 x n | | ϕ | | x 2 x 1 | | , ϕ | | x 2 x 2 | | ϕ | | x 2 x n | | ϕ | | x n x 1 | | , ϕ | | x n x 2 | | ϕ | | x n x n | |
After choosing the multiquadric-form basis function, RBFM ( y ^ r b f m ( x ) ) can be constructed to approximate the actual model y ( x ) by replacing y ^ r b f ( x ) and λ i in Equation (17) with y ^ r b f m ( x ) and λ i , r b f m . The coefficient λ i , r b f m can be calculated based on Equation (18). Similarly, after choosing the inverse multiquadric-form basis function, RBFI ( y ^ r b f i ( x ) ) can be constructed to approximate the actual model y ( x ) . The coefficient λ i , r b f i of y ^ r b f i ( x ) can be calculated based on Equation (18).
Additionally, by choosing the multiquadric-form basis function, a model y ^ 2 , r b f m 1 ( x ) can be constructed to approximate the deviation function of PRS y d , p r s . By replacing the initial training dataset ( x i , y i ) ( i = 1 , , n ) with the updated training dataset of PRS ( x i , y d , p r s i ) ( i = 1 , , n ) , the coefficient λ i , 2 r b f m 1 of y ^ 2 , r b f m 1 ( x ) can be calculated on the basis of Equation (18). Similarly, by choosing the inverse multiquadric-form basis function, a model y ^ 2 , r b f i 1 ( x ) can be constructed to approximate the deviation function of PRS y d , p r s .
Finally, by choosing the multiquadric-form basis function, a model y ^ 2 , r b f m 2 ( x ) can be constructed to approximate the deviation function of SVR y d , s v r . By choosing the interpolation-type metamodel, a model y ^ 2 , r b f i 2 ( x ) can be constructed to approximate the deviation function of SVR y d , s v r .

2.2.4. Step 4: Construction of Ensemble Metamodels

By adding the established y ^ 1 , p r s ( x ) and y ^ 2 , r b f m 1 ( x ) together, PrsRbfm ( y ^ p r s r b f m ( x ) ) can be subsequently constructed as follows.
y ^ p r s r b f m ( x ) = y ^ 1 , p r s ( x ) + y ^ 2 , r b f m 1 ( x )
Being similar to PrsRbfm, PrsRbfi ( y ^ p r s r b f i ( x ) ) can be constructed as follows.
y ^ p r s r b f i ( x ) = y ^ 1 , p r s ( x ) + y ^ 2 , r b f i 1 ( x )
SvrRbfm ( y ^ s v r r b f m ( x ) ) can be constructed as follows.
y ^ s v r r b f m ( x ) = y ^ 1 , s v r ( x ) + y ^ 2 , r b f m 2 ( x )
SvrRbfi ( y ^ s v r r b f i ( x ) ) can be constructed as follows.
y ^ s v r r b f i ( x ) = y ^ 1 , s v r ( x ) + y ^ 2 , r b f i 2 ( x )
The established ensemble metamodels, namely PrsRbfm, PrsRbfi, SvrRbfm, and SvrRbfi, can be used to predict the response at any point of interest in the entire design space by using Equations (19)–(22).

3. Numerical Experiments

3.1. Benchmark Problems

Referred to the website (http://www.sfu.ca/~ssurjano/index.html) and relevant literature [32], six common benchmark problems (BPs) are selected to compare the performance of the individual metamodels (PRS, SVR, RBFM, and RBFI) and the ensemble metamodels (PrsRbfm, PrsRbfi, SvrRbfm, and SvrRbfi).
BP1: 
Goldstein Price Function
f ( x ) = 1 + ( x 1 + x 2 + 1 ) 2 × ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) × 30 + ( 2 x 1 3 x 2 ) 2 × ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 )
where x i [ 2 , 2 ] , for i = 1 , 2 .
BP2: 
Friedman Function
f ( x ) = 10 sin π x 1 x 2 + 20 x 3 0.5 2 + 10 x 4 + 5 x 5
where x i [ 0 , 1 ] , for all i = 1 , , 5 .
BP3: 
Power Sum Function
f ( x ) = j = 1 6 i = 1 6 x i j 36 2
where x i [ 0 , 6 ] , for all i = 1 , , 6 .
BP4: 
Rosenbrock Function
f ( x ) = i = 1 6 100 x i + 1 x i 2 2 + x i 1 2
where x i [ 5 , 10 ] , for all i = 1 , , 7 .
BP5: 
Zakharov Function
f ( x ) = i = 1 9 x i 2 + i = 1 9 0.5 i x i 2 + i = 1 9 0.5 i x i 4
where x i [ 5 , 10 ] , for all i = 1 , , 9 .
BP6: 
Powell Function
f ( x ) = i = 1 2 x 4 i 3 + 10 x 4 i 2 2 + 5 x 4 i 1 x 4 i 2 + x 4 i 2 2 x 4 i 1 4 + 10 x 4 i 3 x 4 i 4
where x i [ 4 , 5 ] , for all i = 1 , , 10 .

3.2. Numerical Setting

For all the benchmark problems, the MATLAB routine “lhsdesign” is used to generate training points and test points. Referred to Jin, Chen, and Simpson [44], n = 3 ( k + 1 ) ( k + 2 ) 2 training points are selected for a k-dimension problem. Moreover, as many test points as possible should be used in practice, since insufficient test points may increase the uncertainty of the results. This paper selects n t s t = 20,000 test points for each benchmark problem. Since the DOE sampling scheme may have an obvious influence on the performance of the metamodels, 100 different training and test sets are selected for each problem. The detailed numerical settings for all the benchmark problems are listed in Table 1. The shape parameters (c) of RBFM and RBFI are both selected as 1 by referring to relevant literature [34,45,46]. The parameters ( ϵ , C, and γ ) of SVR are selected by using the cross-validation method, which was introduced in detail in the published paper of the authors [47].

3.3. Performance Criteria

The root mean square error (RMSE) and the max absolute error (MAE) are selected as the performance criteria.
RMSE can be expressed as
R M S E = i = 1 n t s t ( y i y ^ i ) 2 n t s t
where n t s t denotes the number of test points.
MAE can be expressed as
M A E = max | y i y i ^ | , i = 1 , 2 , , n t s t

4. Results and Discussion

4.1. RMSE

Figure 2 shows the boxplots of RMSE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points. It can be seen that: (1) for all the benchmark problems, the most accurate ensemble metamodels outperform the most accurate individual metamodels; (2) without exception, the least accurate individual metamodels perform worse than the least accurate ensemble metamodels; (3) for each benchmark problem, the performance differences among the four individual metamodels are greater than that among the four ensemble metamodels.
To provide a better comparison for these metamodels, the error values are normalized with respect to the most accurate individual metamodel for each benchmark problem. Table 2 shows the normalized means of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points. The bold values in Table 2 are the most accurate individual/ensemble metamodels, the italic values are the least accurate individual/ensemble metamodels, the underlined values are the ensemble metamodels that perform better than all the individual metamodels, the “Best & Best” values denote the differences between the most accurate ensemble metamodels and individual metamodels, and the “Worst & Worst” values denote the differences between the least accurate ensemble metamodels and individual metamodels. From Table 2, it can be seen that: (1) compared with the most accurate individual metamodels, the means of RMSE of the most accurate ensemble metamodels are reduced, ranging from 1.1% to 22.2%; (2) compared with the least accurate individual metamodels, the means of RMSE of the least accurate ensemble metamodels are reduced, ranging from 21.1% to 52.5%; (3) except for BP3, more than two ensemble metamodels perform better than the most accurate individual metamodels; (4) for BP5, all the four ensemble metamodels perform better than the most accurate individual metamodel.
Table 3 shows the frequency of the accuracy ranking (using RMSE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points. It can be seen that: (1) the frequency of the ensemble metamodels that rank 1st or 2nd is 11, yet the frequency of the individual metamodels is only one; (2) the frequency of the individual metamodels that rank 7th or 8th is 12, yet the frequency of the ensemble metamodels is zero; (3) considered the frequency of the metamodels that rank the top/bottom two, all the ensemble metamodels have better performance than the individual metamodels; (4) PrsRbfm performs best among the four ensemble metamodels, followed by SvrRbfm, PrsRbfi, and SvrRbfi.
To clearly compare the accuracy of each ensemble metamodel with their corresponding individual metamodels, Figure 3 shows the normalized means of RMSE of each ensemble scheme for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points. It can be seen that: (1) in Scheme 1, PrsRbfm ranks 1st among PRS, RBFM, and PrsRbfm for all the benchmark problems; (2) in Scheme 2, PrsRbfi ranks 1st for all the benchmark problems; (3) in Scheme 3, SvrRbfm ranks 1st for four benchmark problems and 2nd for two benchmark problems; although RBFM ranks 1st for two benchmark problems, it is the worst performer for three benchmark problems; (4) in Scheme 4, without exception, the accuracy of SvrRbfi outperforms that of SVR and RBFI.
Table 4 shows the normalized standard deviations of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points. It can be seen that: (1) compared with the most accurate individual metamodels, the standard deviations of RMSE of the most accurate ensemble metamodels are reduced for BP5 and BP6, yet the standard deviations are increased for the other four benchmark problems; (2) compared with the least accurate individual metamodels, the standard deviations of RMSE of the least accurate ensemble metamodels are reduced, ranging from 8.4% to 35.5%.
According to the above experimental results, we think the proposed metamodeling approach could reduce the risk of selecting the worst individual metamodel, and the constructed ensemble metamodels perform better than the used individual metamodels in terms of accuracy. In particular, PrsRbfm performs best among the four ensemble metamodels, followed by SvrRbfm, PrsRbfi, and SvrRbfi.
To provide an explicit explanation for the better performance of the proposed approach, a low-dimensional problem (BP1) and an ensemble scheme (ensemble of SVR and RBFM) are selected as examples. Figure 4 shows the contour plot of the actual function and the approximate functions of SVR, RBFM, and SvrRbfm. It can be seen that: (1) SVR has better global trend fitting capacity than RBFM, such as in the red box area; (2) RBFM performs better in the vicinity of the sampling locations, such as in the red ellipse region; (3) SvrRbfm combines the global trend of SVR and the local accuracy of RBFM, such as in the red box area and the red ellipse region.
Therefore, the reason for the better performance of the ensemble metamodels may be that the proposed metamodeling approach combines the advantages of the regression-type and interpolation-type metamodels. The actual model is regarded as the sum of a regression-type model and a deviation function. Some useful information is first extracted by the regression-type metamodel to capture the global trend of the actual model in the entire design space. Then, some other information is extracted from the deviations at the sampling locations by using the interpolation-type metamodel to achieve the local accuracy in the vicinity of sampling locations.

4.2. Effect of Performance Criteria

The choice of different performance criteria may influence the results of the metamodels. To reduce the source of uncertainty in the results as much as possible, the max absolute error (MAE) is selected as another performance criterion.
Figure 5 shows the boxplots of MAE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points. Table 5 shows the normalized means of MAE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points. From Figure 5 and Table 5, it can be seen that: (1) for each benchmark problem, the performance differences among the four ensemble metamodels are less than that among the four individual metamodels; (2) except for BP6, more than two ensemble metamodels perform better than the most accurate individual metamodels; (3) compared with the most accurate individual metamodels, the means of MAE of the most accurate ensemble metamodels are reduced for five benchmark problems; (4) compared with the least accurate individual metamodels, the means of MAE of the least accurate ensemble metamodels are reduced, ranging from 14.2% to 48.9%.
Table 6 shows the frequency of the accuracy ranking (using MAE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points. It can be seen that: (1) considered the frequency of the metamodels that rank the top/bottom two, PrsRbfm, PrsRbfi, and SvrRbfm outperform all the individual metamodels; (2) although SvrRbfi is a little worse than PRS, it still performs better than its corresponding individual metamodels (SVR and RBFI); (3) PrsRbfm is the best performer of the four ensemble metamodels, followed by SvrRbfm, PrsRbfi, and SvrRbfi.
In summary, the choice of the performance criteria influence the results slightly, but the conclusions obtained by the two criteria remain unchanged.

4.3. Effect of Sampling Densities

The choice of different sampling densities may also influence the results of the metamodels. To investigate the effect of the sampling densities, this paper selects another two schemes with different sampling densities, which are n = 5 ( k + 1 ) ( k + 2 ) 4 and n = 7 ( k + 1 ) ( k + 2 ) 4 .
Table 7 shows the normalized means of RMSE of the metamodels for each benchmark problem with 7 ( k + 1 ) ( k + 2 ) 4 training points. It can be seen that: (1) compared with the most accurate individual metamodels, the means of RMSE of the most accurate ensemble metamodels are reduced, ranging from 0.9% to 8.1%; (2) compared with the least accurate individual metamodels, the means of RMSE of the least accurate ensemble metamodels are reduced, ranging from 23.4% to 53.8%; (3) except for BP3, more than two ensemble metamodels perform better than the most accurate individual metamodels; (4) all the ensemble metamodels perform better than the four individual metamodels; (5) PrsRbfm is the best performer among the four metamodels, while SvrRbfi is the worst performer.
Table 8 shows the normalized means of RMSE of the metamodels for each benchmark problem with 5 ( k + 1 ) ( k + 2 ) 4 training points. It can be seen that: (1) compared with the most accurate individual metamodels, the means of RMSE of the most accurate ensemble metamodels are reduced for five benchmark problems, ranging from 0.9% to 16.9%; (2) compared with the least accurate individual metamodels, the means of RMSE of the least accurate ensemble metamodels are reduced, ranging from 20.9% to 51.3%; (3) all the ensemble metamodels have better performance than the four individual metamodels.
In summary, the choice of different sampling densities influences the results slightly, but the conclusions obtained by the three schemes with different sampling densities remain unchanged.

4.4. Significance of Results

The results above have proven the effectiveness of the proposed method to some extent. To further demonstrate the advantages, the proposed method is compared with some other popular ensemble metamodels, which are BPS (Best PRESS surrogate), PWS (PRESS weighted average surrogate), and OWSD (Optimal weighted surrogate using the diagonal elements). The detailed descriptions of these ensemble metamodels can be found in relevant literature [35,37]. Additionally, Kriging with first order polynomial regression function (KRG1) and Kriging with second-order polynomial regression function (KRG2) are also included in the performance comparison. To be noted, the principle and modeling process of Kriging are different from that of the proposed metamodeling approach in this paper.
Figure 6 compares the performance of PrsRbfm, SvrRbfm, KRG1, KRG2, BPS, PWS, and OWSD. It can be seen that: (1) for BP1, PrsRbfm and SvrRbfm perform better than the other five metamodels; (2) for BP2, SvrRbfm and BPS are the best two performers; (3) for BP3, the accuracy of PrsRbfm and BPS are better than that of the other metamodels; (4) for BP4, PrsRbfm and KRG2 are the best two performers; (5) for BP5, SvrRbfm and BPS are more accurate than other metamodels; (6) for BP6, PrsRbfm and KRG2 perform better the other metamodels.
In summary, the proposed metamodeling approach possesses some advantages when compared with KRG1, KRG2, BPS, PWS, and OWSD.

5. Conclusions

This paper proposed a novel metamodeling approach for building ensemble metamodels. Four types of ensemble metamodels, namely PrsRbfm, PrsRbfi, SvrRbfm, and SvrRbfi, were constructed by choosing four individual metamodels, namely PRS, SVR, RBFM, and RBFI. The performance of these metamodels was investigated through six popular benchmark problems. The effects of the performance criteria and sampling densities on the performance of the metamodels were studied. Additionally, the significance of the results was discussed by comparing the proposed method with some other popular ensemble metamodels. According to the results, some findings of this work could be concluded as follows:
(1)
According to the experimental results, the proposed metamodeling approach could reduce the risk of choosing the worst individual metamodel, and the constructed ensemble metamodels perform better than the selected individual metamodels in terms of accuracy.
(2)
The reason for the better performance of the ensemble metamodels may be that the proposed metamodeling approach combines the advantages of the regression-type and interpolation-type metamodels. The ensemble metamodels not only capture the global trend of the actual model in the entire design space, but also achieve the local accuracy in the vicinity of sampling locations.
(3)
The choices of different performance criteria and sampling densities influence the results slightly, but the obtained conclusions remain unchanged.
(4)
The proposed metamodeling approach possesses some advantages when compared with some other popular ensemble metamodels.

Author Contributions

Conceptualization, C.Y. and J.Z.; Formal analysis, C.Y. and J.F.; Methodology, C.Y. and J.Z.; Software, Z.Q. and J.F.; Validation, C.Y. and Z.Q.; Investigation, C.Y. and Z.Q.; Resources, X.S. and D.M.; Data curation, J.Z., Z.Q. and J.F.; Writing—original draft, C.Y.; Writing—review and editing, C.Y., J.F. and J.Z.; Visualization, J.Z. and D.M.; Supervision, X.S. and D.M.; Project administration, X.S. and D.M.; Funding acquisition, X.S. and D.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. An, X.; Song, B.; Mao, Z.; Ma, C. Layout Optimization Design of Two Vortex Induced Piezoelectric Energy Converters (VIPECs) Using the Combined Kriging Surrogate Model and Particle Swarm Optimization Method. Energies 2018, 11, 2069. [Google Scholar] [CrossRef] [Green Version]
  2. Wang, D.; Hu, Q.; Tang, J.; Jia, H.; Li, Y.; Gao, S.; Fan, M. A kriging model based optimization of active distribution networks considering loss reduction and voltage profile improvement. Energies 2017, 10, 2162. [Google Scholar] [CrossRef] [Green Version]
  3. Melo, A.; Versage, R.; Sawaya, G.; Lamberts, R. A novel surrogate model to support building energy labelling system: A new approach to assess cooling energy demand in commercial buildings. Energy Build. 2016, 131, 233–247. [Google Scholar] [CrossRef]
  4. Bornatico, R.; Hüssy, J.; Witzig, A.; Guzzella, L. Surrogate modeling for the fast optimization of energy systems. Energy 2013, 57, 653–662. [Google Scholar] [CrossRef]
  5. Westermann, P.; Evins, R. Surrogate modelling for sustainable building design—A review. Energy Build. 2019, 198, 170–186. [Google Scholar] [CrossRef]
  6. Ferrero Bermejo, J.; Gómez Fernández, J.F.; Pino, R.; Crespo Márquez, A.; Guillén López, A.J. Review and Comparison of Intelligent Optimization Modelling Techniques for Energy Forecasting and Condition-Based Maintenance in PV Plants. Energies 2019, 12, 4163. [Google Scholar] [CrossRef] [Green Version]
  7. Asher, M.J.; Croke, B.F.W.; Jakeman, A.J.; Peeters, L.J.M. A review of surrogate models and their application to groundwater modeling. Water Resour. Res. 2015, 51, 5957–5973. [Google Scholar] [CrossRef]
  8. Viana, F.A.C.; Simpson, T.W.; Balabanov, V.; Toropov, V. Special section on multidisciplinary design optimization: metamodeling in multidisciplinary design optimization: How far have we really come? AIAA J. 2014, 52, 670–690. [Google Scholar] [CrossRef] [Green Version]
  9. Razavi, S.; Tolson, B.A.; Burn, D.H. Review of surrogate modeling in water resources. Water Resour. Res. 2012, 48, 54–62. [Google Scholar] [CrossRef]
  10. Forrester, A.I.J.; Keane, A.J. Recent advances in surrogate-based optimization. Prog. Aerosp. Sci. 2009, 45, 50–79. [Google Scholar] [CrossRef]
  11. Wang, G.G.; Shan, S. Review of metamodeling techniques in support of engineering design optimization. J. Mech. Des. 2007, 129, 370–380. [Google Scholar] [CrossRef]
  12. González-Fernández, C.; Molinuevo-Salces, B.; García-González, M.C. Evaluation of anaerobic codigestion of microalgal biomass and swine manure via response surface methodology. Appl. Energy 2011, 88, 3448–3453. [Google Scholar] [CrossRef]
  13. Yan, C.; Shen, X.; Guo, F. Novel two-stage method for low-order polynomial model. Math. Probl. Eng. 2018, 2018, 8156390. [Google Scholar] [CrossRef]
  14. Yan, C.; Yin, Z.; Shen, X.; Guo, F.; Wu, Y. Axisymmetric hub-endwall profile optimization for a transonic fan to improve aerodynamic performance based on an integrated design optimization method. Struct. Multidiscip. Optim. 2019, 60, 1267–1282. [Google Scholar] [CrossRef]
  15. Yan, C.; Shen, X.; Guo, F.; Zhao, S.; Zhang, L. A novel model modification method for support vector regression based on radial basis functions. Struct. Multidiscip. Optim. 2019, 60, 983–997. [Google Scholar] [CrossRef]
  16. Lee, C.W.; Lin, B.Y. Applications of the chaotic quantum genetic algorithm with support vector regression in load forecasting. Energies 2017, 10, 1832. [Google Scholar] [CrossRef] [Green Version]
  17. Hong, W.C.; Fan, G.F. Hybrid Empirical Mode Decomposition with Support Vector Regression Model for Short Term Load Forecasting. Energies 2019, 12, 1093. [Google Scholar] [CrossRef] [Green Version]
  18. Fang, H.; Horstemeyer, M.F. Global response approximation with radial basis functions. Eng. Optim. 2006, 38, 407–424. [Google Scholar] [CrossRef]
  19. Zhou, Q.; Cao, L.; Zhou, H.; Huang, X. Prediction of angular distortion in the fiber laser keyhole welding process based on a variable-fidelity approximation modeling approach. J. Intell. Manuf. 2018, 29, 719–736. [Google Scholar] [CrossRef]
  20. Mullur, A.; Messac, A. Metamodeling using extended radial basis functions: A comparative approach. Eng. Comput. 2006, 21, 203–217. [Google Scholar] [CrossRef]
  21. Kim, C.; Wang, S.; Choi, K.K. Efficient response surface modeling by using moving least-squares method and sensitivity. AIAA J. 2005, 43, 2404–2411. [Google Scholar] [CrossRef]
  22. Runge, J.; Zmeureanu, R. Forecasting Energy Use in Buildings Using Artificial Neural Networks: A Review. Energies 2019, 12, 3254. [Google Scholar] [CrossRef] [Green Version]
  23. Silitonga, A.S.; Mahlia, T.M.I.; Shamsuddin, A.H.; Ong, H.C.; Milano, J.; Kusumo, F.; Sebayang, A.H.; Dharma, S.; Ibrahim, H.; Husin, H.; et al. Optimization of Cerbera manghas Biodiesel Production Using Artificial Neural Networks Integrated with Ant Colony Optimization. Energies 2019, 12, 3811. [Google Scholar] [CrossRef] [Green Version]
  24. Crino, S.; Brown, D.E. Global optimization with multivariate adaptive regression splines. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2007, 37, 333–340. [Google Scholar] [CrossRef]
  25. Nam, S.; Hur, J. Probabilistic Forecasting Model of Solar Power Outputs Based on the Naïve Bayes Classifier and Kriging Models. Energies 2018, 11, 2982. [Google Scholar] [CrossRef] [Green Version]
  26. Venturelli, G.; Benini, E.; L, L.W. A Kriging-assisted multiobjective evolutionary algorithm. Appl. Soft Comput. 2017, 58, 155–175. [Google Scholar] [CrossRef]
  27. Perrone, M.P.; Cooper, L.N. When networks disagree: Ensemble methods for hybrid neural networks. In Artificial Neural Networks for Speech and Vision; Mammone, R.J., Ed.; Chapman and Hall: London, UK, 1993; pp. 126–142. [Google Scholar]
  28. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: New York, NY, USA, 1995; pp. 364–371. [Google Scholar]
  29. Zhou, Q.; Rong, Y.; Shao, X.; Jiang, P. Optimization of laser brazing onto galvanized steel based on ensemble of metamodels. J. Intell. Manuf. 2018, 29, 1417–1431. [Google Scholar] [CrossRef]
  30. Zerpa, L.E.; Queipo, N.V.; Pintos, S.; Salager, J.L. An optimization methodology of alkaline-surfactant-polymer flooding processes using field scale numerical simulation and multiple surrogates. J. Pet. Sci. Eng. 2005, 47, 197–208. [Google Scholar] [CrossRef]
  31. Sanchez, E.; Pintos, S.; Queipo, N.V. Toward an optimal ensemble of kernel-based approximations with engineering applications. Struct. Multidiscip. Optim. 2008, 36, 247–261. [Google Scholar] [CrossRef]
  32. Acar, E. Various approaches for constructing an ensemble of metamodels using local measures. Struct. Multidiscip. Optim. 2010, 42, 879–896. [Google Scholar] [CrossRef]
  33. Zhang, J.; Chowdhury, S.; Messac, A. An adaptive hybrid surrogate model. Struct. Multidiscip. Optim. 2012, 46, 223–238. [Google Scholar] [CrossRef]
  34. Lee, Y.; Choi, D.H. Pointwise ensemble of meta-models using v nearest points cross-validation. Struct. Multidiscip. Optim. 2014, 50, 383–394. [Google Scholar] [CrossRef]
  35. Goel, T.; Haftka, R.T.; Shyy, W.; Queipo, N.V. Ensemble of surrogates. Struct. Multidiscip. Optim. 2007, 33, 199–216. [Google Scholar] [CrossRef]
  36. Acar, E.; Rais-Rohani, M. Ensemble of metamodels with optimized weight factors. Struct. Multidiscip. Optim. 2009, 37, 279–294. [Google Scholar] [CrossRef]
  37. Viana, F.A.C.; Haftka, R.T.; Steffen, V. Multiple surrogates: How cross-validation errors can help us to obtain the best predictor. Struct. Multidiscip. Optim. 2009, 39, 439–457. [Google Scholar] [CrossRef]
  38. Toal, D.J.; Keane, A.J. Performance of an ensemble of ordinary, universal, non-stationary and limit Kriging predictors. Struct. Multidiscip. Optim. 2013, 47, 893–903. [Google Scholar] [CrossRef]
  39. Acar, E. Simultaneous optimization of shape parameters and weight factors in ensemble of radial basis functions. Struct. Multidiscip. Optim. 2014, 49, 969–978. [Google Scholar] [CrossRef]
  40. Polynkin, A.; Toropov, V.V. Mid-range metamodel assembly building based on linear regression for large scale optimization problems. Struct. Multidiscip. Optim. 2012, 45, 515–527. [Google Scholar] [CrossRef]
  41. Ferreira, W.G.; Serpa, A.L. Ensemble of metamodels: The augmented least squares approach. Struct. Multidiscip. Optim. 2016, 53, 1019–1046. [Google Scholar] [CrossRef]
  42. Zhou, X.; Jiang, T. Metamodel selection based on stepwise regression. Struct. Multidiscip. Optim. 2016. [Google Scholar] [CrossRef]
  43. Fletcher, R. Practical Methods of Optimization; Wiley: New York, NY, USA, 2013. [Google Scholar]
  44. Jin, R.; Chen, W.; Simpson, T.W. Comparative Studies Of Metamodeling Techniques Under Multiple Modeling Criteria. Struct. Multidiscip. Optim. 2001, 23, 1–13. [Google Scholar] [CrossRef]
  45. Forrester, A.; Sobester, A.; Keane, A. Engineering Design via Surrogate Modelling: A Practical Guide; Wiley: New York, NY, USA, 2008. [Google Scholar] [CrossRef]
  46. Chen, R.; Liang, C.Y.; Hong, W.C.; Gu, D.X. Forecasting holiday daily tourist flow based on seasonal support vector regression with adaptive genetic algorithm. Appl. Soft Comput. 2015, 26, 435–443. [Google Scholar] [CrossRef]
  47. Yan, C.; Shen, X.; Guo, F. An improved support vector regression using least squares method. Struct. Multidiscip. Optim. 2017, 57, 2431–2445. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed approach for building ensembles of regression-type and interpolation-type metamodels.
Figure 1. Flowchart of the proposed approach for building ensembles of regression-type and interpolation-type metamodels.
Energies 13 00654 g001
Figure 2. Boxplots of RMSE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Figure 2. Boxplots of RMSE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Energies 13 00654 g002
Figure 3. Normalized means of RMSE of each ensemble scheme for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Figure 3. Normalized means of RMSE of each ensemble scheme for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Energies 13 00654 g003
Figure 4. Contour plot of the actual function and the approximate functions of SVR, RBFM, and SvrRbfm.
Figure 4. Contour plot of the actual function and the approximate functions of SVR, RBFM, and SvrRbfm.
Energies 13 00654 g004
Figure 5. Boxplots of MAE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Figure 5. Boxplots of MAE of the metamodels over 100 test sets for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Energies 13 00654 g005
Figure 6. Boxplots of RMSE of PrsRbfm, SvrRbfm, KRG1, KRG2, BPS, PWS, and OWSD for the benchmark problems.
Figure 6. Boxplots of RMSE of PrsRbfm, SvrRbfm, KRG1, KRG2, BPS, PWS, and OWSD for the benchmark problems.
Energies 13 00654 g006
Table 1. Detailed numerical settings for the benchmark problems.
Table 1. Detailed numerical settings for the benchmark problems.
Benchmark ProblemNO. of VariablesNO. of Training PointsNO. of Test PointsNO. of Training and Test Sets
BP121820,000100
BP256320,000100
BP368420,000100
BP4710820,000100
BP5916520,000100
BP61019820,000100
Table 2. Normalized means of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Table 2. Normalized means of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
BPsBP1BP2BP3BP4BP5BP6
PRS1.2801.8661.1131.0001.0001.000
SVR1.2241.0001.2621.1491.0061.133
RBFM1.0001.1081.0001.0011.1231.385
RBFI1.1331.2611.5362.1752.0732.166
PrsRbfm0.9291.0430.9810.9570.8890.989
PrsRbfi0.9771.1731.0620.9900.9850.994
SvrRbfm0.9680.9221.0391.0060.7781.080
SvrRbfi1.0100.9371.1761.1020.9101.109
Best & Best−7.1%−7.8%−1.9%−4.3%−22.2%−1.1%
Worst & Worst−21.1%−37.1%−23.5%−49.3%−52.5%−48.8%
Table 3. Frequency of the accuracy ranking (using RMSE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Table 3. Frequency of the accuracy ranking (using RMSE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Ranking1st2nd3rd4th5th6th7th8th
PRS00202002
SVR00100230
RBFM01021020
RBFI00000114
Total01323366
PrsRbfm41010000
PrsRbfi02120100
SvrRbfm21111000
SvrRbfi01102200
Total65343300
Table 4. Normalized standard deviations of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Table 4. Normalized standard deviations of RMSE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
BPsBP1BP2BP3BP4BP5BP6
PRS1.0001.0001.6461.5661.0011.072
SVR2.0671.5421.12110.6636.1886.844
RBFM1.4621.1671.0001.0001.2041.488
RBFI1.6601.1001.2351.1171.0001.000
PrsRbfm1.3971.1851.3041.5920.7440.995
PrsRbfi1.4231.1931.4931.5770.9531.055
SvrRbfm1.5811.1431.5092.0211.4273.714
SvrRbfi1.6961.1991.1237.0413.9915.542
Best & Best39.7%14.3%12.3%57.7%−25.6%−0.5%
Worst & Worst−17.9%−22.2%−8.4%−34.0%−35.5%−19.0%
Table 5. Normalized means of MAE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Table 5. Normalized means of MAE of the metamodels for each benchmark problem with 3 ( k + 1 ) ( k + 2 ) 2 training points.
BPsBP1BP2BP3BP4BP5BP6
PRS1.0551.5831.0001.0001.0001.000
SVR1.1491.0001.4051.3121.0801.325
RBFM1.0001.1481.1651.1741.3851.708
RBFI1.1891.2001.6262.5151.9392.353
PrsRbfm0.9101.1110.9520.9800.9441.015
PrsRbfi0.9651.2780.9990.9990.9961.002
SvrRbfm0.9500.9331.1641.1680.9571.244
SvrRbfi1.0210.9561.3381.2851.0521.302
Best & Best−9.0%−6.7%−4.8%−2.0%−5.6%0.2%
Worst & Worst−14.2%−19.3%−17.7%−48.9%−45.7%−44.7%
Table 6. Frequency of the accuracy ranking (using MAE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Table 6. Frequency of the accuracy ranking (using MAE) of the metamodels for the six benchmark problems with 3 ( k + 1 ) ( k + 2 ) 2 training points.
Ranking1st2nd3rd4th5th6th7th8th
PRS10210101
SVR00100230
RBFM00013020
RBFI00000105
Total10323456
PrsRbfm40110000
PrsRbfi03200010
SvrRbfm12030000
SvrRbfi01003200
Total56343210
Table 7. Normalized means of RMSE of the metamodels for each benchmark problem with 7 ( k + 1 ) ( k + 2 ) 4 training points.
Table 7. Normalized means of RMSE of the metamodels for each benchmark problem with 7 ( k + 1 ) ( k + 2 ) 4 training points.
BPsBP1BP2BP3BP4BP5BP6
PRS1.4002.2821.1041.0001.3421.000
SVR1.2091.0001.2831.1541.0001.130
RBFM1.0001.1861.0001.0121.4961.371
RBFI1.1741.3831.5552.1882.8472.195
PrsRbfm0.9211.1020.9520.9461.1580.991
PrsRbfi0.9851.2711.0440.9871.3160.992
SvrRbfm0.9580.9371.0401.0000.9191.076
SvrRbfi1.0140.9491.1911.1020.9721.105
Best & Best−7.9%−6.3%−4.8%−5.4%−8.1%−0.9%
Worst & Worst−27.6%−44.3%−23.4%−49.6%−53.8%−49.7%
Table 8. Normalized means of RMSE of the metamodels for each benchmark problem with 5 ( k + 1 ) ( k + 2 ) 4 training points.
Table 8. Normalized means of RMSE of the metamodels for each benchmark problem with 5 ( k + 1 ) ( k + 2 ) 4 training points.
BPsBP1BP2BP3BP4BP5BP6
PRS1.2681.5771.1151.0221.0001.000
SVR1.2481.0001.2421.2061.0271.174
RBFM1.0001.0201.0001.0001.1341.379
RBFI1.1001.1261.5172.1512.0302.093
PrsRbfm0.9371.0081.0150.9910.9180.991
PrsRbfi0.9801.0981.0791.0140.9890.995
SvrRbfm0.9650.9241.0531.0350.8311.073
SvrRbfi1.0020.9381.1671.1520.9511.139
Best & Best−6.3%−7.6%1.5%−0.9%−16.9%−0.9%
Worst & Worst−20.9%−30.4%−23.1%−46.4%−51.3%−45.6%

Share and Cite

MDPI and ACS Style

Yan, C.; Zhu, J.; Shen, X.; Fan, J.; Mi, D.; Qian, Z. Ensemble of Regression-Type and Interpolation-Type Metamodels. Energies 2020, 13, 654. https://doi.org/10.3390/en13030654

AMA Style

Yan C, Zhu J, Shen X, Fan J, Mi D, Qian Z. Ensemble of Regression-Type and Interpolation-Type Metamodels. Energies. 2020; 13(3):654. https://doi.org/10.3390/en13030654

Chicago/Turabian Style

Yan, Cheng, Jianfeng Zhu, Xiuli Shen, Jun Fan, Dong Mi, and Zhengming Qian. 2020. "Ensemble of Regression-Type and Interpolation-Type Metamodels" Energies 13, no. 3: 654. https://doi.org/10.3390/en13030654

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop