Next Article in Journal
The Consistent Couple Stress Theory-Based Vibration and Post-Buckling Analysis of Bi-directional Functionally Graded Microbeam
Previous Article in Journal
Formation of Fuzzy Patterns in Logical Analysis of Data Using a Multi-Criteria Genetic Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Penalty Function Optimization in Dual Response Surfaces Based on Decision Maker’s Preference and Its Application to Real Data

1
Faculty of Computer & Mathematical Statistics, UiTM Kelantan, Kota Bharu Campus, Kota Bharu 150150, Kelantan, Malaysia
2
Department of Mathematics and Statistics, Faculty of Science, Universiti Putra Malaysia, Serdang 43400, Selangor, Malaysia
3
Institute for Mathematical Research, Universiti Putra Malaysia, Serdang 43400, Selangor, Malaysia
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(3), 601; https://doi.org/10.3390/sym14030601
Submission received: 15 February 2022 / Revised: 11 March 2022 / Accepted: 15 March 2022 / Published: 17 March 2022

Abstract

:
The dual response surface methodology is a widely used technique in industrial engineering for simultaneously optimizing both the process mean and process standard deviation functions of the response variables. Many optimization techniques have been proposed to optimize the two fitted response surface functions that include the penalty function method (PM). The PM method has been shown to be more efficient than some existing methods. However, the drawback of the PM method is that it does not have a specific rule for determining the penalty constant; thus, in practice, practitioners will find this method difficult since it depends on subjective judgments. Moreover, in most dual response optimization methods, the sample mean and sample standard deviation of the response often use non-outlier-resistant estimators. The ordinary least squares (OLS) method is also usually used to estimate the parameters of the process mean and process standard deviation functions. Nevertheless, not many statistics practitioners are aware that the OLS procedure and the classical sample mean and sample standard deviation are easily influenced by the presence of outliers. Alternatively, instead of using those classical methods, we propose using a high breakdown and highly efficient robust MM-mean, robust MM-standard deviation, and robust MM regression estimators to overcome these shortcomings. We also propose a new optimization technique that incorporates a systematic method to determine the penalty constant. We call this method the penalty function method based on the decision maker’s (DM) preference structure in obtaining the penalty constant, denoted as PMDM. The performance of our proposed method is investigated by a Monte Carlo simulation study and real examples that employ symmetrical factorial design of experiments (DOE). The results signify that our proposed PMDM method is the most efficient method compared to the other commonly used methods in this study.

1. Introduction

Response surface methodology (RSM) was first developed by Box and Draper [1]. It is an important tool to find the optimal factor settings of the design point, which can either maximize or minimize the given response function. RSM is widely used in many disciplines, such as in manufacturing industries, engineering, and agricultural sciences. The food industry, in particular, has been a prime user of RSM since the early 1970s [2]. For example, Dey and Dora [3] studied the effects of temperature, pH, enzyme concentration/substrate concentration (E/S) ratio on the response, i.e., degree of hydrolysis (DH) for marine shrimp. They noted that RSM was successfully applied to determine the optimum operating conditions on the control variables for maximum DH value. The traditional RSM emphasizes locating the optimal process parameters to achieve the target mean value where it assumes a homogeneous variance. However, in a real situation, this assumption may not be achieved. In this situation, both the mean and the standard deviation of the responses should be considered when determining the optimum conditions for the input variables.
The dual response surface methodology is the most efficient method for simultaneously optimizing the mean and the standard deviation functions of the responses to achieve the desired target while keeping the standard deviation small. It is common to employ both symmetrical and asymmetrical designs along with the regression method to model the mean and the standard deviations of the responses. Asymmetric design refers to a situation where all the factors have the same levels of variation while asymmetrical design is when not all the factors have the same levels of variation. However, symmetrical 2 k , 3 k and 4 k factorial designs of experiments are the commonly used designs in industrial engineering because a large amount of information can be acquired from a relatively small number of experiments. In dual RSM, the mean and the standard deviation of the responses at each design point are first calculated. Then, statistical models for the mean and standard deviation of the responses are established by fitting the response surface models (usually using a second-degree polynomial model) to the experimental data. The parameters of the models are then estimated prior to simultaneously optimizing both the estimated mean and the estimated standard deviation of the responses to obtain an optimal setting for the input/control variables.
Various optimization methods of dual response surface optimization (DRSO), which attempt to simultaneously optimize both the mean and the standard deviation of the responses, were proposed in the literature [4,5,6,7]. Vining and Myers [8] were the first to introduce dual response surface optimization (DRSO) using response surface methodology (RSM). We denote this method as the VM method. However, their method received criticism from Del Castillo and Montgomery [4] due to using Lagrange multipliers to simultaneously optimize the process mean and process standard deviation. They noted that this approach may be impractical and imprecise for a global best solution due to the restriction of the process mean which was made to be equal to a specific target value. Thus, they proposed the generalized reduced gradient (GRD) technique with inequality constraints to rectify the drawback of the VM optimization method. Similarly, Lin and Tu [5] also pointed out the weakness of the VM method in which it did not guarantee a global optimal due to its constraint to a specific value. Hence, they proposed an optimization scheme which we refer to as the LT optimization scheme. It is based on the mean squared error (MSE) objective function that allows for a small bias. Kim and Lin [9] pointed out that the objective function of LT does not consider the measure of the violation of constraint, and it does not clearly specify how far the estimated mean response is from the desired target value. In an attempt to reduce the effect of bias and variance in obtaining the optimal setting conditions for the input variables, Ding et al. [7] modified the LT method by imposing relative weights, w on the bias and the variance terms of the MSE optimization scheme. Nonetheless, this method, which is referred to as WMSE, received criticism from Jeong et al. [10] due to the lack of consideration of the decision maker’s interest in determining the weight functions. Therefore, [9,10,11] then introduced a proper procedure to determine the value of w by considering the decision maker’s preference. Nevertheless, most of the existing methods do not show much improvement in terms of acquiring an estimated mean response that is closer to the target value with small variation.
Despite the shortcoming of the LT method, this optimization scheme is a widely used method to solve dual response and multiple response problems [12,13,14,15,16]. Baba et al. [17] developed a new optimization scheme by employing the penalty function (PM) approach which was found to be more efficient compared to other approaches discussed earlier. However, the weakness of this method is that it has neither a clear rule nor a systematic approach to determine the value of the penalty constant, ξ .
It is important to highlight that in most optimization methods of dual RSM including the PM, the ordinary least squares (OLS) is the commonly used method for estimating the models’ parameters. The classical sample mean and the classical sample standard deviation are also popular used measures of location/central tendency and measures of spread, respectively. Unfortunately, many statistics practitioners are not aware that the classical mean, classical standard deviation, and OLS method are not resistant to outliers [18]. Outlier(s) is a single or group of observations that are markedly different among the bulk of data or pattern set by most observations. It is now evident that outliers may have an unduly effect on the sample mean, sample standard deviation, and OLS estimates [19,20,21,22]. As a result, using those classical approaches will give inaccurate estimates of the mean response and standard deviation of the response.
Therefore, to address the problem of outliers on the parameter estimates, robust methods which are not easily affected by outliers are put forward. Robustness refers to the ability of a method to remain unaffected or less affected by outliers [23]. Many robust estimation methods such as the M, MM, LMS, and LTS can be found in the literature [24,25,26]. Rousseeuw and Leroy [25] proved that the robust MM estimator is highly efficient and has a high breakdown point.
The work of [17] has inspired us to propose a new method called the penalty function method based on the decision maker’s (DM) preference structure in obtaining the penalty constant, denoted as PMDM. The PMDM is the extension work of the PM optimization method, which is based on our newly developed method of determining penalty constant, ξ . To reduce the effect of outliers, robust MM estimators are incorporated in the algorithm of PMDM to estimate the mean and the standard deviation of the responses as well as to obtain the fitted models of the mean and the standard deviations.
The objectives of this study are: (1) to establish a new optimization method for simultaneously optimizing the fitted models of the mean and the standard deviation of the responses by incorporating a penalty function based on our newly proposed penalty constant, ξ ; (2) to evaluate the performance of the proposed method compared to VM, LT, WMSE, and PM by using simulation study; (3) to apply the proposed method on two real datasets, namely the catapult study data and printing process study data. The significance of this study is that it can contribute to the development of a very powerful tool for simultaneously optimizing the fitted models of the mean and the standard deviation to determine optimum conditions on the models’ control variables which could give an optimum response with the least bias and least variability. Researchers will find our proposed PMDM method very useful, especially to those who wish to determine the optimum combination of factor settings that result in the estimated mean response closer to the target value with the smallest bias and smallest variability, which is desirable. However, it is worth mentioning that the accuracy of the estimators relies on the selected data (i.e., experimental design). Hence, it is crucial to choose a good design for an experiment to be able to capture maximum information about its behavior. A poor design cannot produce valuable knowledge about the experiment, even if effective estimation methods are used [27,28].
The rest of the paper is structured as follows. Section 2 presents a brief review of the dual response surface optimization. The proposed Penalty Function Optimization Method is discussed in Section 3. Section 4 presents the results of a simulation study and two real datasets. Finally, concluding remarks are described in Section 5.

2. The Dual Response Surface Optimization

The dual response surface method aims to find the optimum setting condition of the controllable factors to diminish the variability and deviation from the desired target of the decision-maker. Three basic strategies are considered to achieve such aim: experimental design, regression fitting, and optimization aspect. Let us say an experiment was conducted to analyze the effect of some control factors for a given experimental design. Suppose the characteristics of interest, i.e., the response variable depends on a set of control variables (coded), x = ( x 1 ,   x 2 ,   , x p ) . Let y i j represent the response variable at the ith design point ( i = 1 , 2 , , n ) and the jth replicate ( j = 1 , 2 , , m ) . Table 1 presents the summary of the observed data where x ,   y ¯ i ,   s i 2 represent a set of controlled variables or factors, sample mean, and sample variance of the response, respectively.
At each of the design points, the sample mean and sample standard deviation of the responses are usually calculated as in Equation (1).
y ¯ i = 1 m i = 1 m y i j   and   s i = 1 m 1 ( j = 1 m y i j y ¯ i ) 2
Following Myers et al. [29], the control variables are firstly translated from the natural units (original) into coded as in Equation (2).
x i j = δ i j [ max ( δ i j ) + min ( δ i j ) ] / 2 [ max ( δ i j ) min ( δ i j ) ] / 2 i = 1 , 2 , , n   ;   j = 1 , 2 , , p ,
where δ is the original value of x . Max (.) and min (.) refer to the maximum and minimum values of the original values of x, respectively.
Subsequently, the second-order polynomial model response functions for the mean and standard deviation of the responses are formulated. The ordinary least squares (OLS) method is usually used to estimate the parameters of the models. The fitted dual response functions for the sample mean and sample standard deviation at each design point, denoted as ω ^ μ and ω ^ σ , respectively, are written as follows:
ω ^ μ = β ^ 0 + i = 1 p β ^ i x i + i = 1 p β ^ i i x i 2 +   i < j β ^ i j x i x j
and
ω ^ σ = γ ^ 0 + i = 1 p γ ^ i x i + i = 1 p γ ^ i i x i 2 +   i < j γ ^ i j x i x j
where β ^ and γ ^ are the estimates of the model parameters.
The control variables, x i in Equations (3) and (4) are firstly translated from the natural units into coded units. Then, an optimization method is used to simultaneously optimize both the mean and standard deviation functions of Equations (3) and (4) to obtain the estimated optimal factor settings. Then, the estimated mean response and estimated standard deviation of the responses can be determined.
In this section, we briefly summarized some of the commonly used optimization methods. Prior to formulating the optimization scheme, the processes mean and process standard deviations’ models need to be established at the outset. The VM method simultaneously optimizes the objective function of the process standard deviation in which it restricts the process mean towards the desired target. Using the second-order polynomial model in Equations (3) and (4), the VM method of optimization scheme is to
minimize   ω ^ σ subject   to   ω ^ μ = τ
where τ is the target value.
Lin and Tu [5] proposed an optimization scheme based on the squared-loss model which is known as the mean square error (MSE). This model deals with square bias and minimizes the variance component of the factor settings. The Lin and Tu (LT) method is to
minimize   MSE = ( ω ^ μ τ ) 2 + ( ω ^ σ ) 2
Ding et al. [7] proposed a natural extension of the LT method by imposing relative weights on the bias and the variance terms. Their method which is called weighted MSE (WMSE) is to
minimize   WMSE = w ( ω ^ μ τ ) 2 + ( 1 w ) ω ^ σ
where w = relative weight factor, ( 0 w 1 ) . The data-driven approach is used to determine the value of w .
Baba et al. [17] proposed a penalty function-based approach as another alternative optimization scheme. The penalty method (PM) is to
minimize = ( ξ 2 )   ( ω ^ μ τ ) 2 + ( ω ^ σ ) 2
where ξ is the penalty constant, 0 ξ . Nonetheless, this method concentrates more on the bias and forces the estimated mean response to be close to the target value, which is not adequately efficient. Moreover, no specific penalty constant value, ξ was given to Equation (8). Hence, practitioners will find these approaches difficult since they must do trial and error subjective judgment.

3. The Proposed Penalty Function Optimization in Dual Response Based on Decision Maker’s Preference

In the preceding sections, we have reviewed some of the optimization schemes of the dual response surface methodology, and their weaknesses are noted. The penalty function-based approach (PM) seems to be better than the other methods but still has shortcomings, whereby the determination of the penalty constant ξ is not clearly specified. Hence, we propose a new method which is called the penalty method based on the decision maker’s (PMDM) preference, where the penalty constant ξ is determined by considering the decision maker’s judgments. The proposed method is less complicated and can be implemented by practitioners. Prior to the development of the method, the sample mean, sample standard deviations, and the fitted response models of Equations (3) and (4) are estimated using the MM estimator which is highly robust and highly efficient with a high breakdown point. The resultant MM estimates are highly robust and less affected by outliers.

3.1. Robust Measures of Location and Spread

Mean and standard deviation are popularly used measures of location/central tendency and measure spread of data, respectively. However, there is evidence that they may perform poorly when outliers occur in data [23]. As already mentioned, the robust method which is not easily affected by outliers is recommended. This section discusses the MM robust estimator as an alternative to the classical methods presented in Equation (1) for estimating the mean and standard deviation of the response values at each design point. This estimator, known as the MM estimator, was proposed by Yohai [26]. It is highly efficient with high breakdown points and is not easily affected by outliers. Consider the following location-scale model
x i = μ + σ ε i
where x 1 ,   x 2 , , x n are n observations and ε i ,   i = 1 , 2 , , n , are independent and identically distributed (i.i.d) observations with variance equals to 1. The following algorithm (see Maronna et al. [23]) summarized the MM algorithm for estimating μ , and the scale σ .
Step 1: The initial consistent estimator of the location μ 0 and scale σ 0 were computed using high breakdown point S-estimator [26].
Step 2: From Step 1, compute the residuals e i and subsequently compute the M estimate of standard deviation parameter (scale), σ ^ where σ ^ is the solution to
1 n   ρ 0 ( e i σ ^ ) = 0 . 5
where t = e i / σ ^   and   ψ = ρ 0 ( t / c 0 ) must be a redescending ρ function such as Hampel, Tukey’s biweight, and Tanh functions. Tukey’s biweight function is employed in this paper.
Step 3: Compute the M estimate of μ ^ using ρ 1 . Yohai [26] noted that for Tukey’s bisquare weight function, employing, c 0 = 0.4685 and c 1 = 4.68 , results in high breakdown and high-efficiency estimator, respectively. μ ^ is a solution to
  ψ ( e i σ ^ ) x i = 0
where ψ = ρ 1 ( t / c 1 ) . Upon convergence, μ ^ and σ ^ are the estimates for the mean (MM-mean) and standard deviation (MM-standard deviation) based on the MM estimator. In the subsequent sections, they are referred to as MM l and MM s , respectively. The same procedures were applied to estimate the parameters of Equations (3) and (4) using the MM estimator with a slight change, where the polynomial regression model was considered instead of the location model.

3.2. The New Optimization Approach

The new optimization approach (PMDM) consists of two stages whereby in the first stage, penalty constant ξ is determined. Subsequently, in the second stage, Equation (13) is optimized to obtain the estimated optimal factor settings.
Stage 1:
Determining the value ξ
Determining the value ξ is critical and challenging because of its large interval [11]. The optimal penalty constant denoted as ξ * is defined as the value that can provide an estimated mean response with reasonably close to zero bias and at the same time minimizes the estimated standard deviation of the response. Here, we illustrate how the value of penalty value ξ is determined, by using, the roman catapult problem data. This data set was also used by Kim and Lin [9] to assess their proposed method. The aim of the experiment is to identify a combination of factor settings for which a variety of projectile types can be delivered to a target with minimal variability. The response variable is the distance of the projectile from the base of the catapult. The distance was measured in inches. The input variables are x 1 the arm length (0.32–3.68 inches), x 2 stop angle (30–90°) x 3 and pivot height (2.5–5.5 inches). The target distance of the projectile from the base is set to 80. Figure 1 depicts the corresponding estimated mean and estimated variance of responses for Catapult study data, for some values of penalty constant,   ξ , in the interval of 1–200, i.e., [1, 200]. From the figure, it can be seen that the estimated variance of the response increases as the ξ value increases. On the contrary, the estimated mean response decreases when the ξ value increases. However, when the ξ value reaches a certain point, the performance of both estimated responses becomes stable. In such circumstances, it is essential to diminish the unnecessary value of ξ .
Jeong et al. [10] stated that in such a case, it is necessary to narrow down the interval to make a substantial choice of ξ . Following the idea of [10], we firstly defined a vector z i = ( z 1 i , z 2 i ) T where   z 1 = ( y ¯ i τ ) 2   ( bias )   and   z 2 = s i 2 (variance). The penalty function (PF) denoted as PF ξ for each ξ , 0 ξ can be expressed as follows:
PF ξ i = ( ξ 2 )   z 1 i + ( z 2 i ) 2
The idea behind this basic premise is that the ξ value should be congruent with the DM’s preference structure whereby the ranking of the z i given by the DM should be consistent with the ordering of the corresponding PF ξ values. For example, if the DM prefers z i to z j (denoted as z i z j ), PF ξ z i should be less than PF ξ z j . We employed the pairwise ranking scheme that compares only one pair of vectors at each time. The maximum number of pairwise ranks is C 2 n . If the expression PF ξ z i < PF ξ z j is true, then the possible interval of ξ is obtained, otherwise, it will not be considered as the interval of ξ . This procedure is repeated for all pairwise vectors. Next, the set of all interval ξ i , j n that satisfies inequalities PF ξ z i < PF ξ z j is obtained by intersecting the ξ i , j n for all possible pairs of vectors.
Owing to this, we propose a systematic algorithm to find the optimal penalty constant ξ * . Hence, the proposed algorithm of determining the penalty constant ξ method can be summarized as follows
Step 1: Firstly compute y ¯ i and s i as in Equation (1).
Step 2: Compute a vector z i = ( z 1 i , z 2 i ) T where   z 1 = ( y ¯ i τ ) 2   ( bias )   and   z 2 = s i 2   ( variance ) .
Step 3: Calculate criteria value, i.e., CV i = | y ¯ i τ | + s i . We suggest a simple algorithm of reducing the number of pairwise comparisons of z i by computing where △ is an outlier-resistant estimator, i.e., median of CV i . If CV i value exceeds the value, then vectors of z i = ( z 1 i , z 2 i ) T will not be considered in the ranking of z i .
To clarify the idea, the process mean, y ¯ i , process standard deviation, s i , criteria values, CV i along with the value z i , are presented in Table 2 for Catapult data. In this example, the design point i = 1 , 3 , 6 , 8 , 9 , 10 , 11 , 12 , 13 , 14 were removed from the analysis since their corresponding CV i exceed the median   CV i = 27.72 , Consequently, only design point i = 2 , 4 , 5 , 7 , 15 , 16 , 17 , 18 , 19 , 20 are remained. After that, the remaining design points were ranked in ascending order, i.e., z 1 , , z 10 based on the value of CV i . The number of pairwise comparisons to be made is reduced to C 2 k = 45 , i.e., { ( z 1 , z 2 ) , ( z 1 , z 3 ) , ( z 1 , z 4 ) , , ( z 9 , z 10 ) } , where k is the remaining number of design points.
Step 4: Conduct C 2 k pairwise comparisons. For each pair of ( z i , z j ) , where z i z j ), obtain the values for PF ξ z i and PF ξ z j for some values of ξ in an interval, i . e . ,   0 ξ a that satisfy the condition PF ξ z i < PF ξ z j . Denote the interval of the corresponding ξ for each pair in term of sets, i.e., S 1   S 2     S C 2 k . Then, the intersection of these sets, i.e., S 1   S 2     S C 2 k will be considered as the final set of the optimal penalty constant, ξ .
Stage 2:
Optimization stage
In this stage, optimization models are formulated and solved. This yield estimated optimum values of process mean and process standard deviation of the responses. As mentioned earlier, the new procedure introduced in this paper is called a penalty method based on the decision maker’s (PMDM) preference which simultaneously optimizes the process mean and process standard deviation functions of the response variables. The PMDM is defined as:
minimize = ( ξ * 2 )   ( ω ^ μ τ ) 2 + ( ω ^ σ ) 2
where ξ * is the optimal penalty constant determined in Step 4 of Stage 1. The optimal setting ( x * ) with the corresponding ξ value within the interval that gives the smallest RMSE and bias will be considered as the final optimal setting. The ω ^ μ and ω ^ σ from Equations (3) and (4) are the fitted values based on the MM regression estimator. It is also noted that the MM estimator is used to estimate the mean and the standard deviation of the response variables. The PMDM can also be used if it is based on classical estimators. However, we anticipate that the results based on a robust estimator are more efficient than on a classical estimator.
The PMDM in Equation (13) is optimized by using any unconstrained optimization method such as Newton’s method, BFGS method, conjugate gradient method, and steepest ascent (descent) method. In addition, nonlinear optimization software can be used to find the optimal factor settings for dual response surface problems. In this study, the genetic algorithm (GA) optimization was implemented to obtain the estimated optimal factor settings. The R software is utilized to analyze this problem. The resultant estimated optimal factor settings are then substituted in Equations (3) and (4) to obtain the estimated optimum mean response and standard deviation of the responses. We expect that the PMDM could give an optimum response with the least bias and least variability, compared to other methods.

4. Results

Monte Carlo simulation study and two real datasets are presented in this section to assess the performance of the proposed PMDM compared to VM, LT, WMSE, and PM methods. All analyses were computed using the R program. The robust package in R, i.e., lmrob, rrcov, and Rsolnp are used to obtain the robust MM estimates of model parameters, process mean, and process standard deviation, respectively.

4.1. Monte Carlo Simulation Study

Monte Carlo simulation study is designed to assess the performance of our new proposed PMDM method and to compare its performance with other methods namely VM, LT, WMSE and PM. Following Park and Cho [12] and Goethals and Cho [13], a symmetric 33 factorial design of experiments is considered, whereby at each control factor settings, x i = ( x i 1 , x i 2 , x i 3 ) , i = 1 , 2 , , 27 , five responses ( y i 1 , y i 2 , , y i 5 ) are randomly generated from normal distribution with mean μ ( x i ) and standard deviation σ ( x i ) , as follows:
μ ( x i ) = 500 + ( x 1 + x 2 + x 3 ) 2 + x 1 + x 2 + x 3 σ ( x i ) = 100 + ( x 1 + x 2 + x 3 ) 2 + x 1 + x 2 + x 3
where the desired target is assumed to be 500, i.e., τ = 500 . We refer to this generated data as ‘clean’ data. The performance of the proposed method when compared to other methods is evaluated under three different frameworks as shown in Table 3.
In order to see the effect of outliers on the different methods, three observations of the original responses are randomly replaced with contaminated values which are drawn from a normal distribution, i.e., N ( 150 , 10 2 ) . The five methods are then applied to each set of the generated data. In each simulation run, 500 and 1000 iterations are considered. The performance of the proposed method is evaluated based on bias, standard error (SE) and root mean squares errors (RMSE) of the estimated optimal mean response, μ ^ . The estimated mean and variance of the estimated optimal mean response, computed over k iterations are given by μ ^ ¯ = i = 1 k ( μ ^ / k ) and i = 1 k ( μ ^ μ ^ ¯ ) / k , respectively. The bias is computed as bias = μ ^ ¯ 500 and the root mean square error ( RMSE ) = ( bias ) 2 + var ( μ ^ ) . The MSE consists of two components; one measures the variability (precision) and the other measures its bias (accuracy). A good method is one that has the least values of bias, standard error (SE), and RMSE. The bias, standard error (SE), and RMSE of the estimated optimal mean response for models A, B, and C are exhibited in Table 4, Table 5 and Table 6, respectively.
First, the focus is on the results of Table 4 for clean data where the process mean estimates, process standard deviation estimates, and regression estimates were calculated based on the classical methods. It can be observed that our proposed PMDM-optimization method outperforms the other four methods evident by having the smallest bias and RMSE for both the simulation iterations.
Similar results are obtained in Table 5 for the data having three outliers with the classical framework as described in Table 3 (model B). It can be observed from Table 5 that the biases and RMSE of all methods have markedly increased. This is due to the use of the classical methods to compute the estimates of mean and standard deviation and also the use of the OLS method in obtaining the estimates of the predicted models in Equations (3) and (4). However, our proposed PMDM method is still the most efficient method among the five optimization schemes because it has the smallest value of RMSE, followed by the PM, LT, WMSE, and VM methods.
The bias, SE, and RMSE of the estimated optimal mean responses for contaminated data based on robust estimators (model C) are presented in Table 6. It is interesting to observe that the bias, SE, and RMSE for the five estimates have significantly reduced. These attractive results are due to the usage of the robust MM estimator, which is resistant to outliers. The results of our proposed PMDM are very appealing. Similar to models A and B, the proposed PMDM emerges to be conspicuously the most efficient method. It has the smallest bias, SE, and RMSE among the five methods. Again, the performance of PM comes second and is followed by the LT, WMSE, and VM methods. Our proposed PMDM optimization method based on the robust MM-fitted response surface function has substantially improved the estimated optimal mean responses in terms of having the smallest bias, SE, and RMSE. Smaller RMSE indicates that the combination of optimum factor settings results in the estimated mean response being closer to the target value with less bias and less variability, which is desirable.

4.2. The Catapult Study Data

The roman style catapult problem data is the first real example to show the merit of our proposed PMDM method when compared to the other four methods. This data set was also used by Kim and Lin [9] and Baba et al. [17] to assess their proposed methods. The central composite design (CCD) with three replicates was evaluated at each design point. Following Kim and Lin [9], the target of interest for distance point is assumed to be 80, i.e., τ = 80 .
In order to see the effect of an outlier on the estimated optimal mean response, we deliberately changed one data point, that is the y 3 , 20 observation corresponding to y value equals 79 (in bold) with higher values of 790. Figure 2a,b shows the boxplot of the original and modified data sets with one outlier. It can be observed from the boxplot that the distribution of the response variable of the original data set is fairly close to a normal distribution. However, the response variables for the modified data are skewed to the right with one outlier detected. The data set is shown in Table 7 (the control variables, x i are coded and the original values are in parenthesis) while in Table 8, the mean, standard deviation, MM-mean, MM-standard deviation, and criteria value (CV) for model A, model B, and model C are presented. What is immediately clear from Table 8 is that the process mean and process standard deviation based on the sample mean and sample standard deviation are very sensitive to the contaminated data (shown in bold) when compared with the proposed estimation approach using MM-mean and MM-standard deviation.
If all of the 20 alternative vectors were to be compared, the total number of pairwise comparisons would be 190. In addition, the number of pairwise comparisons can be reduced by screening out the vectors in which their CVs exceed the median of CV. Similar to the simulation experiment, we consider three model scenarios as depicted in Table 3; model A (the classical method is used to estimate the mean and the standard deviation of the responses for clean data), model B (the classical method is used to estimate the mean and the standard deviation of the responses for contaminated data), and model C (the robust MM method is used to estimate the mean and the standard deviation of the responses for contaminated data). From Table 8, it can be observed that only 10 remaining vectors for data with model A and model C, and 9 vectors for data with model B were considered out of 20 vectors. Pairwise comparisons among the vectors are conducted. For models A and B, they consist of C 2 10 while for model C, C 2 9   pairwise comparisons.
Figure 3 shows that the proposed method has established a set value of ξ as a result of the intersection of the I ( i , j ) . First, in Figure 3a, the resultant intersection plot is determined by I ( 3 , 5 ) and I ( 6 , 7 ) which corresponds to the lower and upper limits of ξ , i.e., ξ = 5 and ξ = 6 . Then, the final set of optimal penalty constant ξ * to be used for model A is ξ A * ϵ [ 5 ,   6 ] . Secondly, for model B, the lower and upper limits are ξ = 2 and ξ = 3 which are determined by I ( 2 , 4 ) and I ( 4 , 5 ) . Hence, the optimal penalty constant is ξ B * ϵ [ 2 ,   3 ] as shown in Figure 3b. Finally, Figure 3c shows the final penalty constant for model C, i.e., ξ C * ϵ [ 1 ,   5 ] . Subsequently, the optimal setting ( x * ) for each model is obtained by minimizing the PMDM in Equation (13) using the resultant ξ * values. The optimal setting ( x * ) with the corresponding ξ value within the interval that gives the smallest RMSE and bias will be considered as the final optimal setting.
The VM, LT, WMSE, and PM methods are then applied to the data and their performances are compared with the PMDM. The estimated optimal factor setting, estimated optimal mean response, estimated standard deviation, and RMSE for models A, B, and C are exhibited in Table 9, Table 10 and Table 11, respectively. It is interesting to observe the results of our proposed PMDM in Table 9, Table 10 and Table 11 for clean and contaminated data. Let us first focus on Table 9 for clean data. The PMDM method shows a slightly better result when compared to other methods for clean data. It can be observed from Table 9 that the estimated mean optimal response of our proposed PMDM is the closest to the target value and has the least standard deviation and RMSE compared to other methods. Similar results are obtained for model B (using classical methods to obtain various estimates for contaminated data) in which our PMDM method has the smallest standard deviations and RMSE values among the five methods. As displayed in Table 11 for contaminated data (model C) that use robust methods to obtain various estimates, our proposed PMDM method has also outperformed the other methods evident by having the smallest value of standard deviation and RMSE. Moreover, the estimated mean optimal response of the PMDM is very close to the target value, i.e., 80 with the least standard deviation, which is very desirable. The results of this data set agree reasonably well with the results of the simulation study. Here, we present the interpretation of the results, in particular for the PMDM method, model C as in Table 11 after the coded units of the estimated factor settings are translated back to natural units using Equation (2). The results indicate that the optimum combination of factors where arm length at 1.7628 inches, stop angle at 65.0340, and pivot height at 3.603 inches would give a maximum value of 79.9856 inches of the response, i.e., distance point where a projectile lands from the base of the catapult.

4.3. Printing Process Study Data

The second real example in this paper is the printing process study data [1]. The experiment introduced by [1] was conducted to determine the effect of the three control variables: x 1 (speed), x 2 (pressure), and x 3 (distance) on the characteristics of the printing process, which include the machine’s index to apply colored inks to package labels. A symmetric 33 factorial design of experiments with three replicates at each of the 27 design points is considered, assuming that the target of interest for distance point is τ = 500 .
In order to see the effect of an outlier on the estimated mean optimal response, we deliberately changed one data point, that was the y 1 , 8 observation corresponding to y value equaling 259 (in bold) with higher values of 9259. Figure 4a,b displays the boxplot of the original and the modified data sets with an outlier. From the boxplot, it suggests that the response variable of the original data is reasonably close to normal distribution. On the other hand, the modified data of the response variables are skewed to the right with one outlier detected. The data set is shown in Table 12 (the control variable, x i are coded) while in Table 13, the mean, standard deviation, MM-mean, MM-standard deviation, and criteria value (CV) for model A, model B, and model C are presented.
If all 27 alternative vectors were to be compared, the total number of pairwise comparisons would be 351. In addition, the number of pairwise comparisons can be reduced by screening out the vectors in which their CVs exceed the median of CV. Similar to the simulation experiment and Catapult data example, we consider three model scenarios, i.e., models A, B, and C as depicted in Table 2. From Table 13, it can be observed that there are only 13 remaining vectors of data that include all three models. Thus, C 2 13 pairwise comparisons among the vectors are considered.
Figure 5 shows that the proposed method has established a set value of ξ as a result of the intersection of the I ( i , j ) . First, in Figure 5a, the resultant intersection plot is determined by I ( 3 , 4 ) which corresponds to the lower and upper limits of ξ , i.e., ξ = 3 and ξ = 4 . Then the final set of optimal penalty constant ξ * to be used for models A and B is ξ A , B * ϵ [ 3 ,   4 ] . Finally, Figure 5b shows the final penalty constant for model C, i.e., ξ C * ϵ [ 2 ,   4 ] .
All five methods were used to determine the optimal factor setting, optimal mean response, standard deviation, and RMSE, which are exhibited in Table 14, Table 15 and Table 16, respectively. Similar to the Catapult data, it can be observed from Table 14 that the PMDM shows a slightly better result when compared to other methods for clean data. It is also interesting to observe that the PMDM consistently provides the smallest standard deviation and RMSE for models B and C when outliers are present in the data. The results of the real examples were consistent with the results of the simulation study. It is very interesting to note that the results that show the estimated mean optimal response of the PMDM is very close to the target value, i.e., 500 with the least standard deviation, which is very desirable.

5. Conclusions

The main aim of this study is to develop a more reliable alternative scheme (PMDM) that simultaneously optimizes both the mean and standard deviation functions of the response variables in the response surface model. Robust MM-mean, robust MM-standard deviation estimator, and MM regression estimator which are highly efficient and have high breakdown points were employed instead of using the classical non-robust methods. We also establish a systematic method to determine the penalty constant ξ which is based on the decision maker’s (DM) preference structure and integrate this value in the PMDM optimization scheme. The existing optimization scheme is not effective enough as they fail to obtain an estimated mean response that is closer to the target value with a small variation. Moreover, those methods are easily affected by outliers due to the use of non-robust methods to estimate the mean and the standard deviation of the response variables, and the non-robust least squares (OLS) method to estimate the parameters of the mean and standard deviation functions. The results obtained from the real data set and the Monte Carlo simulation study show that our proposed PMDM method is more efficient than the existing VM, LT, WMSE, and PM methods, irrespective of whether or not the outlier exists in the data set. Nonetheless, our proposed PMDM method has its own shortcoming as its computation running time is longer than that of the other methods in this study. A slightly longer running time compared to the other methods is a trade-off one has to consider when using our PMDM method as the method is very efficient in determining the optimum combination of factor settings that result in the estimated mean response closer to the target value with the smallest bias and smallest variability, which is desirable. In future work, we look forward to developing a reliable method of estimating the parameters of the mean and standard deviations functions of the responses for an unbalanced design which is prone to produce heteroscedastic errors. After that, our proposed PMDM method will be employed to simultaneously optimize both functions to obtain more efficient optimal factor settings, hence improving processes.

Author Contributions

Conceptualization, N.A.A. and H.M.; methodology, N.A.A. and H.M.; validation, N.A.A. and H.M.; formal analysis, N.A.A.; writing—original draft preparation, N.A.A.; writing review and editing, H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to its theoretical and mathematical approach.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Catapult study datasets and the Printing process datasets are used to verify the performance of our proposed method. The Catapult study datasets has been used by Kim and Lin [9] and Baba et al. [17]. The Printing Process dataset was introduced by Box and Draper [1].

Acknowledgments

The authors would like to thank the reviewers for their constructive suggestions.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

References

  1. Box, E.; Draper, N.R. Empirical Model-Building and Response Surfaces; John Wiley & Sons: Hoboken, NJ, USA, 1987. [Google Scholar]
  2. Myers, R.H.; Khuri, A.I.; Carter, W.H. Response surface methodology: 1966–l988. Technometrics 1986, 31, 137–157. [Google Scholar] [CrossRef]
  3. Dey, S.S.; Dora, K.C. Optimization of the production of shrimp waste protein hydrolysate using microbial proteases adopting response surface methodology. J. Sci. Technol. 2011, 51, 16–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Del Castilo, E.; Montgomery, D.C. A nonlinear programming solution to the dual response problem. J. Qual. Technol. 1993, 25, 199–204. [Google Scholar] [CrossRef]
  5. Lin, D.K.; Tu, W. Dual response surface optimization. J. Qual. Technol. 1995, 27, 34–39. [Google Scholar] [CrossRef]
  6. Copeland, K.A.; Nelson, P.R. Dual response optimization via direct function minimization. J. Qual. Technol. 1996, 28, 331–336. [Google Scholar] [CrossRef]
  7. Ding, R.; Lin, D.K.; Wei, D. Dual-response surface optimization: A weighted MSE approach. Qual. Eng. 2004, 16, 377–385. [Google Scholar] [CrossRef]
  8. Vining, G.; Myers, R. Combining Taguchi and response surface philosophies: A dual response approach. J. Qual. Technol. 1990, 22, 38–45. [Google Scholar] [CrossRef]
  9. Kim, K.J.; Lin, D.K. Dual response surface optimization: A fuzzy modeling approach. J. Qual. Technol. 1998, 30, 1–10. [Google Scholar] [CrossRef]
  10. Jeong, I.J.; Kim, K.J.; Chang, S.Y. Optimal weighting of bias and variance in dual response surface optimization. J. Qual. Technol. 2005, 37, 236–247. [Google Scholar] [CrossRef]
  11. Lee, D.H.; Kim, K.J. Interactive weighting of bias and variance in dual response surface optimization. Experts Syst. Appl. 2012, 39, 5900–5906. [Google Scholar] [CrossRef]
  12. Park, C.; Cho, B.R. Development of robust design under contaminated and non-normal data. Qual. Eng. 2003, 15, 463–469. [Google Scholar] [CrossRef]
  13. Goethals, P.L.; Cho, B.R. Solving the optimal process target problem using response surface designs in heteroscedastic conditions. Int. J. Prod. Res. 2011, 49, 3455–3478. [Google Scholar] [CrossRef]
  14. Boylan, G.L.; Cho, B.R. Comparative studies on the high-variability embedded robust parameter design from the perspective of estimator. Comput. Ind. Eng. 2013, 64, 442–452. [Google Scholar] [CrossRef]
  15. Park, C.; Leeds, M. A highly efficient robust design under data contamination. Comput. Ind. Eng. 2016, 93, 131–142. [Google Scholar] [CrossRef]
  16. Park, C.; Ouyang, L.; Byun, J.H.; Leeds, M. Robust design under normal model departure. Comput. Ind. Eng. 2017, 113, 206–220. [Google Scholar] [CrossRef]
  17. Baba, I.; Midi, H.; Rana, S.; Ibragimov, G. An alternative approach of dual response surface optimization based on penalty function method. Math. Probl. Eng. 2015, 2015, 450131. [Google Scholar] [CrossRef] [Green Version]
  18. Dhhan, W.; Midi, H. A high breakdown, high efficiency and bounded influence modified GM estimator based on support vector regression. J. Appl. Stat. 2017, 44, 700–714. [Google Scholar] [CrossRef]
  19. Alguraibawi, M.; Midi, H.; Rana, S. Robust Jackknife Ridge Regression to Combat Multicollinearity and High Leverage Points in Multiple Linear Regressions. Econ. Comput. Econ. Cybern. Stud. Res. 2015, 4, 305–322. [Google Scholar]
  20. Rana, S.; Midi, H.; Imon, A.H.M.R. Robust wild bootstrap for stabilizing the variance of parameter estimates in heteroscedastic regression models in the presence of outliers. Math. Probl. Eng. 2012, 2012, 730328. [Google Scholar] [CrossRef] [Green Version]
  21. Uraibi, H.; Midi, H. Robust Variable Selection Method Based on Huberized Lars-Lasso Regression. Econ. Comput. Econ. Cybern. Stud. Res. 2020, 54, 145–160. [Google Scholar]
  22. Rashid, A.M.; Midi, H.; Slwabi, W.D.; Arasan, J. An efficient estimation and classification methods for high dimensional data using robust iteratively reweighted SIMPLS algorithm based on nu-support vector regression. IEEE Access 2021, 9, 45955–45967. [Google Scholar] [CrossRef]
  23. Maronna, R.A.; Martin, R.D.; Yohai, V.J.; Salibian-Barrera, M. Robust Statistics: Theory and Methods; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  24. Huber, P.J. Robust Statistics; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  25. Rousseeuw, P.J.; Leroy, A.M. Robust Regression and Outlier Detection; John Wiley & Sons: Hoboken, NJ, USA, 1987. [Google Scholar]
  26. Yohai, V.J. High breakdown-point and high efficiency robust estimates for regression. Ann. Stat. 1987, 15, 642–656. [Google Scholar] [CrossRef]
  27. Elsawah, A.M. An appealing technique for designing optimal large experiments with three-level factors. J. Comp. Appl. Math. 2021, 384, 113–164. [Google Scholar] [CrossRef]
  28. Elsawah, A.M. Multiple doubling: A simple effective construction technique for optimal tow-level experimental design. Stat. Pap. 2021, 62, 2923–2967. [Google Scholar] [CrossRef]
  29. Myers, R.H.; Montgomery, D.C.; Anderson-Cook, C.M. Response Surface Methodology: Process and Product Optimization Using Designed Experiments; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
Figure 1. The estimated mean and estimated variance versus penalty constant ( ξ ) .
Figure 1. The estimated mean and estimated variance versus penalty constant ( ξ ) .
Symmetry 14 00601 g001
Figure 2. Boxplot for (a) original data (b) modified (with outlier) for Catapult data.
Figure 2. Boxplot for (a) original data (b) modified (with outlier) for Catapult data.
Symmetry 14 00601 g002
Figure 3. The individual set of ξ for each model for Catapult study data set. (a) ξ A * for model A, (b) ξ B * for model B and (c) ξ C * for model C.
Figure 3. The individual set of ξ for each model for Catapult study data set. (a) ξ A * for model A, (b) ξ B * for model B and (c) ξ C * for model C.
Symmetry 14 00601 g003aSymmetry 14 00601 g003b
Figure 4. Boxplot for (a) original (no outlier) data set and (b) modified data set with outlier for Printing process data.
Figure 4. Boxplot for (a) original (no outlier) data set and (b) modified data set with outlier for Printing process data.
Symmetry 14 00601 g004
Figure 5. The individual set of ξ for each model for printing study data set (a) ξ A * and ξ B * for model A and B and (b) ξ C * for model C.
Figure 5. The individual set of ξ for each model for printing study data set (a) ξ A * and ξ B * for model A and B and (b) ξ C * for model C.
Symmetry 14 00601 g005
Table 1. The presentation of the observed data for a given experimental design.
Table 1. The presentation of the observed data for a given experimental design.
Run x Replication y ¯ i s i 2
1 x 11 x 12 x 1 p y 11 y 12 y 1 m y ¯ 1 s 1 2
2 x 21 x 22 x 2 p y 21 y 22 y 2 m y ¯ 2 s 2 2
... ... ...
... ... ...
... ... ...
i x i 1 x i 2 x i p y i 1 y i 2 y i m y ¯ i s i 2
... ... ...
... ... ...
... ... ...
n x n 1 x n 2 x n p y n 1 y n 2 y n m y ¯ n s n 2
Table 2. Summary of ranking process.
Table 2. Summary of ranking process.
i y ¯ i s i CV i z i = ( z 1 i , z 2 i ) T Code
141.01.740.7(1521, 1.7)
280.710.010.7(0.49, 10.0) z 6
347.04.437.4(1089, 4.4)
475.019.524.5(25, 19.5) z 9
560.37.527.2(388.09, 7.5) z 10
6114.711.646.3(1204.09, 11.6)
769.07.818.8(121, 7.8) z 8
894.730.745.4(216.09, 30.7)
956.74.928.2(542.89, 4.9)
10111.38.139.4(979.69, 8.1)
1150.07.037.0(900, 7.0)
1260.024.444.4(400, 24.4)
1354.75.030.3(640.09, 5.0)
14116.76.843.5(1346.89, 6.8)
1584.75.910.6(22.09, 5.9) z 5
1683.33.87.1(10.89, 3.8) z 1
1785.33.89.1(28.09, 3.8) z 3
1886.03.79.7(36, 3.7) z 4
1984.34.79.0(18.49, 4.7) z 2
2085.75.911.6(32.49, 5.9) z 7
Median   CV i 27.72
Table 3. Framework for the purpose of performance comparison.
Table 3. Framework for the purpose of performance comparison.
ModelTypes of DataProcess LocationProcess ScaleRegression Estimator
AWithout outlierSample meanSample standard deviationOLS
BWith outlierSample meanSample standard deviationOLS
CWith outlierMM-meanMM-standard deviationMM estimator
Table 4. Estimated bias, SE, and RMSE of the optimal mean response for model A (classical method were used, data without outliers).
Table 4. Estimated bias, SE, and RMSE of the optimal mean response for model A (classical method were used, data without outliers).
Methods500 Iterations1000 Iterations
BiasSERMSEBiasSERMSE
VM3.5784.6985.9053.5764.6095.834
LT2.1491.6692.7212.1931.7602.811
WMSE0.6232.5312.6072.6262.4773.609
PM0.9340.0320.9350.9300.0310.931
PMDM0.5380.1150.5500.5640.1280.579
Table 5. Estimated bias, SE, and RMSE of the optimal mean response for model B (classical method were used, data with outliers).
Table 5. Estimated bias, SE, and RMSE of the optimal mean response for model B (classical method were used, data with outliers).
Methods500 Iterations1000 Iterations
BiasSERMSEBiasSERMSE
VM0.8404.3834.4621.3024.4124.600
LT1.4633.7344.0101.3043.7473.967
WMSE1.3213.8244.0461.4723.7834.060
PM0.0650.2280.2370.0710.2220.233
PMDM0.0040.1300.1300.0160.1880.189
Table 6. Estimated bias, SE, and RMSE of the optimal mean response for model C (robust method were used, data with outliers).
Table 6. Estimated bias, SE, and RMSE of the optimal mean response for model C (robust method were used, data with outliers).
Methods500 Iterations1000 Iterations
BiasSERMSEBiasSERMSE
VM0.4221.8101.8590.3881.7831.825
LT0.1580.3000.3390.1460.3250.356
WMSE0.2170.5290.5720.1910.5670.598
PM0.0110.0140.0180.0070.0280.029
PMDM0.0040.0100.0110.0050.0170.018
Table 7. The Catapult study data.
Table 7. The Catapult study data.
i x 1 x 2 x 3 y i 1 y i 2 y i 3
1−1 (0.32)−1 (30)−1 (2.5)394242
2−1 (0.32)−1 (30)1 (5.5)809171
3−1 (0.32)1 (90)−1 (2.5)524544
4−1 (0.32)1 (90)1 (5.5)976068
51 (3.68)−1 (30)−1 (2.5)606853
61 (3.68)−1 (30)1 (5.5)113127104
71 (3.68)1 (90)−1 (2.5)786564
81 (3.68)1 (90)1 (5.5)1307579
9−1.682 (−0.823)0 (2)0 (2)596051
101.682 (4.826)0 (2)0 (2)115117102
110 (2)−1.682 (−48.46)0 (2)505743
120 (2)1.682 (52.46)0 (2)884349
130 (2)0 (2)−1.682 (−0.523)546050
140 (2)0 (2)1.682 (4.523)122119109
150 (2)0 (2)0 (2)878978
160 (2)0 (2)0 (2)868579
170 (2)0 (2)0 (2)888781
180 (2)0 (2)0 (2)898782
190 (2)0 (2)0 (2)868879
200 (2)0 (2)0 (2)889079 (790)
Table 8. Process mean and process standard deviation with the criteria value (CV) and ranking code for Catapult study data set.
Table 8. Process mean and process standard deviation with the criteria value (CV) and ranking code for Catapult study data set.
i y ¯ i s i CV i Code
Model A
Code
Model B
MM l MM s CV i Code
Model C
141.01.740.7 41.11.140.0
280.710.010.7 z 6 z 6 80.58.08.5 z 5
347.04.437.4 44.51.136.6
475.019.524.5 z 9 z 8 64.05.921.9 z 10
560.37.527.2 z 10 z 9 60.37.527.2
6114.711.646.3 114.46.340.7
769.07.818.8 z 8 z 7 64.51.116.6 z 9
894.730.745.4 77.04.27.2 z 2
956.74.928.2 59.51.121.6
10111.38.139.4 116.02.138.1
1150.07.037.0 50.04.534.5
1260.024.444.4 46.05.139.1
1354.75.030.3 54.64.329.7
14116.76.843.5 117.83.241.0
1584.75.910.6 z 5 z 5 87.32.19.4 z 7
1683.33.87.1 z 1 z 1 85.51.16.6 z 1
1785.33.89.1 z 3 z 3 87.51.18.6 z 6
1886.03.79.7 z 4 z 4 86.22.18.3 z 4
1984.34.79.0 z 2 z 2 85.22.17.3 z 3
2085.7 (349.7)5.9 (459.3)11.6 (729) z 7 84.5 (84.5)6.5 (6.5)11.0 (11.0) z 8
Median   CV i 27.72 (29.3) Median   CV i 21.6 (21.6)
Table 9. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model A, classical methods were used, data without outlier).
Table 9. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model A, classical methods were used, data without outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(0.0444, −0.3093, −0.2394)79.58183.00923.0381
LT(0.0449, −0.3100, −0.2374)79.62283.01433.0378
WMSE(0.0440, −0.3087, −0.2414)79.54093.00413.0389
PM(0.0471, −0.3138, −0.2262)79.84743.00233.0062
PMDM(0.0166, −0.3743, −0.1778)79.86392.99853.0016
Table 10. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model B, classical methods were used, data with outlier).
Table 10. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model B, classical methods were used, data with outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(0.7948, 0.0526, −1.4118)78.88755.02075.1425
LT(0.7959, 0.0519, −1.4105)78.99455.04205.1413
WMSE(0.7932, 0.0533, −1.4135)78.75494.99425.1471
PM(0.7959, 0.0519, −1.4105)78.99455.04205.1413
PMDM(0.5231, 0.1060, −1.5008)78.98264.85174.9572
Table 11. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model C, robust methods were used, data with outlier).
Table 11. Estimated optimal settings, estimated mean response, estimated standard deviation of response and RMSE for Catapult study data set, (model C, robust methods were used, data with outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(−0.1381, 0.1684, −0.2634)79.93141.54941.5509
LT(−0.1377, 0.1684, −0.2633)79.93831.54961.5508
WMSE(−0.1386, 0.1683, −0.2636)79.92281.54901.5509
PM(−0.1355, 0.1684, −0.2625)79.97731.54821.5484
PMDM(−0.1412, 0.1678, −0.2646)79.98561.54711.5472
Table 12. The printing process study data.
Table 12. The printing process study data.
i x 1 x 2 x 3 y i 1 y i 2 y i 3
1−1−1−1341028
20−1−1115116130
31−1−1192186263
4−10−1828888
500−144178188
610−1322350350
7−11−114111086
801−1259 (9259)251259
911−1290280245
10−1−10818181
110−109012293
121−10319376376
13−100180180154
14000372372372
15100541568396
16−110288192312
17010432336513
18110713725754
19−1−1136499199
200−11232221266
211−11408415443
22−101182233182
23001507515434
24101846535640
25−111236126168
26011660440403
271118789911161
Table 13. Process mean and process standard deviation with the criteria value (CV) and ranking code for printing study data set.
Table 13. Process mean and process standard deviation with the criteria value (CV) and ranking code for printing study data set.
i y ¯ i s i CV i Code
Model A
Code
Model B
MM l MM s CV i Code
Model C
12412.49488.5 26.055.80479.6
2120.38.39388.1 115.51.07385.6
3213.742.8329.1 z 13 1895.13316.1
4863.46417.5 86.591.73415.1
5136.780.41443.7 1836.62323.6
6340.716.17175.5 z 7 z 7 341.512.31170.8 z 7
7112.327.57415.3 111.817.56405.76
8256.3 (3256.33)4.62 (5198.5)248.3 (7954.8) z 9 255 (255)8.55 (8.55)253.6 (253.6) z 12
9271.723.63251.9 z 11 z 10 2856.62221.6 z 10
10810419.0 81.860418.1
11101.717.67416.0 91.53.20411.7
1235732.91175.9 z 8 z 8 358.130.69172.6 z 8
13171.315.01343.7 171.814342.2
143720128.0 z 4 z 4 371.70128.3 z 5
15501.792.594.2 z 2 z 2 554.510.8965.4 z 2
1626463.5299.5 z 13 z 12 30010.26210.3 z 9
1742788.61161.6 z 6 z 6 427.485.88158.5 z 6
18730.721.08251.8 z 10 z 9 729.412.82242.2 z 11
19220.7133.8413.1 149106.84457.8
20239.723.46283.8 z 12 z 11 232.99.44276.5 z 13
2142218.5296.5 z 3 z 3 4136.5593.6 z 3
2219929.45330.5 197.422.42325.0
23485.344.6459.3 z 1 z 1 5115.9216.9 z 1
24673.7158.2331.9 668.3112.18280.5
25176.755.51378.8 175.644.88369.3
26501138.9139.9 z 5 z 5 421.523.94102.4 z 4
271010142.5652.5 934.522.27456.8
Median   CV i 329.1 (330.5) Median   CV i 280.5 (280.5)
Table 14. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model A, classical methods were used, data without outlier).
Table 14. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model A, classical methods were used, data without outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(1.00, 0.06, −0.24)494.6543.4043.73
LT(1.00, 0.07, −0.25)494.6943.4643.78
WMSE(1.00, 0.08, −0.25)496.4443.5243.67
PM(1.00, 0.12, −0.26)500.0042.7542.75
PMDM(1.00, 0.10, −0.25)500.0042.0042.00
Table 15. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model B, classical methods were used, data with outlier).
Table 15. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model B, classical methods were used, data with outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(0.93, 1.00, −1.00)493.2846.7247.20
LT(0.92, 1.00, −1.00)493.9446.8147.20
WMSE(0.92, 1.00, −1.00)493.2846.7247.20
PM(0.92, 1.00, −1.00)493.9346.8047.19
PMDM(0.91, 1.00, −1.00)493.9646.5546.94
Table 16. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model C, robust methods were used, data with outlier).
Table 16. Estimated optimal settings, estimated mean response, estimated standard deviation of response, and RMSE for printing study data set, (model C, robust methods were used, data with outlier).
Method Optimal   Settings   ( x * ) ω ^ μ ( x * ) ω ^ σ ( x * ) RMSE
VM(0.72, 0.22, −0.16)498.9615.5515.58
LT(0.72, 0.22, −0.16)499.0715.5615.59
WMSE(0.72, 0.22, −0.16)498.9615.5515.58
PM(1.00, 0.07, −0.21)499.6215.5915.59
PMDM(0.73, 0.21, −0.15)499.9815.0915.09
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Aziz, N.A.; Midi, H. Penalty Function Optimization in Dual Response Surfaces Based on Decision Maker’s Preference and Its Application to Real Data. Symmetry 2022, 14, 601. https://doi.org/10.3390/sym14030601

AMA Style

Aziz NA, Midi H. Penalty Function Optimization in Dual Response Surfaces Based on Decision Maker’s Preference and Its Application to Real Data. Symmetry. 2022; 14(3):601. https://doi.org/10.3390/sym14030601

Chicago/Turabian Style

Aziz, Nasuhar Ab., and Habshah Midi. 2022. "Penalty Function Optimization in Dual Response Surfaces Based on Decision Maker’s Preference and Its Application to Real Data" Symmetry 14, no. 3: 601. https://doi.org/10.3390/sym14030601

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop