Next Article in Journal
On Discrete Approximation of Analytic Functions by Shifts of the Lerch Zeta Function
Previous Article in Journal
Fuzzy Extension of Crisp Metric by Means of Fuzzy Equivalence Relation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Propose-Specific Information Related to Prediction Level at x and Mean Magnitude of Relative Error: A Case Study of Software Effort Estimation

1
Faculty of Applied Informatics, Tomas Bata University in Zlin, Nad Stranemi 4511, 76001 Zlin, Czech Republic
2
Faculty of Information Technology, School of Engineering and Technology, Van Lang University, Ho Chi Minh City 700000, Vietnam
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(24), 4649; https://doi.org/10.3390/math10244649
Submission received: 2 November 2022 / Revised: 1 December 2022 / Accepted: 5 December 2022 / Published: 8 December 2022
(This article belongs to the Section Computational and Applied Mathematics)

Abstract

:
The prediction level at x ( P R E D ( x ) ) and mean magnitude of relative error ( M M R E ) are measured based on the magnitude of relative error between real and predicted values. They are the standard metrics that evaluate accurate effort estimates. However, these values might not reveal the magnitude of over-/under-estimation. This study aims to define additional information associated with the P R E D ( x ) and M M R E to help practitioners better interpret those values. We propose the formulas associated with the P R E D ( x ) and M M R E to express the level of scatters of predictive values versus actual values on the left ( s i g L e f t ), on the right ( s i g R i g h t ), and on the mean of the scatters ( s i g ). We depict the benefit of the formulas with three use case points datasets. The proposed formulas might contribute to enriching the value of the P R E D ( x ) and M M R E in validating the effort estimation.

1. Introduction

One of the essential aspects of developing software projects is software effort estimation (SEE) [1,2,3,4,5]. In the early stages of project development, resources or budgets must be measured. The estimation might assist software project managers in determining how much money from the budget is being spent on maintenance activities or project completion. The inaccurate estimate might lead to unrealistic resource allocation and project risk, leading to project failure [6]. The accuracy of effort estimation is, therefore, crucial. Scientists have always adopted evaluation criteria when comparing their proposal models to estimate software project efforts with others. The model that best meets the evaluation criteria is likely to be the most appropriate model chosen to be used to estimate subsequent projects [6,7].
The magnitude of relative error ( M R E ), prediction level at x ( P R E D ( x ) ), and mean magnitude of relative error ( M M R E ) that were proposed by [8] are well-known evaluation criteria in SEE (see Equations (1)–(3) below). Although Myrtveit, Stensrud, and Shepperd stated that this criterion might have some disadvantages given the M M R E , it is still widely used in the validation of real effort estimation. As presented in Table 1, many researchers have used them for measuring the accurate predictive effort estimation:
M R E i = | y i y i ^ | y i
M M R E = i = 1 N M R E i N
P R E D ( x ) = 1 N i = 1 N 1 if M R E i x 0 o t h e r w i s e
where y i ^ is the predicted and y i is the i-th observed value.
As can be seen, Equations (2) and (3) only provide evaluation criteria for estimation, and they might not reveal the distribution of predicted values and observed values around a baseline. A baseline is a straight line where predictive and actual values are the same, i.e., a line with an equation y i ^ = y i . Although such information might be determined by analyzing the prediction residuals, additional information associated with the P R E D ( x ) and M M R E might be more interesting. The distribution of these values might be helpful information for researchers because it might bring additional information about the predictive model’s performance.
This article proposes the additional information related to the P R E D ( x ) and M M R E by defining formulas, s i g L e f t , s i g R i g h t , and s i g , to determine how the predicted values in comparison with actual values are symmetric around the baseline. The s i g L e f t , s i g R i g h t , and s i g values may reflect the proposed model’s trend. The closer the values reach 0, the better the proposed model distributes around the baseline. On the other hand, if they approach either –1 or +1, the proposed model is either under- or over-estimated. Hence, their value might contribute to the P R E D ( x ) and M M R E .
The sections of this article are organized as follows: Section 2 presents the related works; Section 3 proposes the three formulas s i g L e f t , s i g R i g h t , and s i g ; the characteristics of the sig formula are presented in Section 4; Section 5 presents the research questions; Section 6 gives the result and discussion; Section 7 conveys the conclusion and future work.

2. Related Works

The P R E D ( x ) and M M R E might be used in SEE situations because they interpret the absolute percentage error. Since the P R E D ( x ) and M M R E do not scale, they might be used to aggregate estimated errors from software development projects of various sizes. According to Jørgensen et al., however, there is no top score constraint for overestimating in the M M R E ; underestimating effort will never result in an M M R E score more significant than one. In their publication [18], they investigated the requirements for the practical usage of the M M R E in effort estimation contexts. As a result, they stated that this criterion was still helpful in effort estimation in most cases.
According to Conte et al., one of the advantages of the M M R E is that its comparison might be adopted for all kinds of predicted models [8]. However, Foss et al. [19] concluded that it might not be correct when they simulated analysis using the model evaluation criteria M M R E . They realized that the M M R E might be an untrustworthy or insufficient criterion when choosing suitable linear predicted models in terms of the M M R E . They stated that the M M R E is likely to select a model that delivers an underestimation of a model. They advised employing a combination of the theoretical justification for the proposed models and additional criteria. A similar conclusion was also shown by Myrtveit et al. [20].
In addition, Kitchenham et al. [21] proposed the variable z (where z = predicted/actual) derived from the M M R E . They argued that the distribution of z was required to evaluate the accuracy of a predictive model. However, they also pointed out that the z variable had a limit regarding the summary statistics. The estimation may prefer prediction models that minimize overestimations over prediction systems rather than underestimates. Furthermore, several researchers used other criteria when evaluating the performance of effort estimation, such as the mean inverse balanced relative error ( M I B R E ), mean balanced relative error ( M B R E ), mean absolute error ( M A E ), and standardized accuracy (SA). They stated that the M M R E might be unbalanced and yield an asymmetry distribution [22,23]. This means that this criterion might cause some problems identifying under-estimating or over-estimating.
Moreover, as mentioned in the reports [24,25], the M M R E is possibly the most widely used measure of estimation error in research and industrial applications. There might be several advantages of the M M R E that can boost its popularity, such as: (1) it is unaffected by the sizing unit used; it makes no difference whether SEE is measured in hours or months of workload [20]; (2) it might be not influenced by scale; this means that the accuracy of the M M R E does not change with the sizing unit selected [26].
This section presents a literature overview of recent papers that used the M M R E and P R E D ( x ) to validate the model’s accuracy in SEE. The search criteria are based on the most recently published articles with high citations. In addition, we will present some review studies in which the authors list the research articles related to SEE and the number of studies that use the M M R E and P R E D ( x ) .
In 2022, Mahmood et al. investigated the effectiveness of machine-learning-based effort estimation based on the M M R E and P R E D ( 0.25 ) [3]. They concluded that ensemble effort estimation techniques might be better than solo techniques. Praynlin 2021 also used the M M R E and other evaluation criteria to validate meta cognitive neuro fuzzy and other methods (particle swarm optimization, genetic algorithm, and backpropagation network) [9]. Furthermore, Fadhil et al. [10] proposed the DolBat model to predict effort estimation; this proposal was compared with the constructive cost model (COCOMO) based on the M M R E and P R E D ( x ) [10]. As a result, they concluded that their model was better than COCOMO II. This criterion was also used by Hamid et al. [27] when they compared the IRDSS model with Delphi and Planning Poker [27].
Bilgaiyan, S. et al. [11] adopted the M M R E , P R E D ( x ) , and mean-squared error ( M S E ) in comparison to the performance of the feedforward back-propagation neural network (NN), cascade correlation NN, and Elman NN in terms of effort estimation [11]. Mustapha et al. also used the P R E D ( 0.25 ) , M M R E , and median of the magnitude of the relative error ( M d M R E ) to identify the accuracy of their approach when they investigated the use of random forest in software effort estimation [12]. The M M R E was also adopted to validate the accurate effort estimation reported in the publication [13,14].
On the other hand, this criterion was employed by Desai and Mohanty [17]; Effendi, Sarno, and Prasetyo [15]; and Khan et al. [16]. Effendi et al. compared the Optimization of COCOMO with COCOMO [15]; Khan et al. used it and the P R E D ( x ) to compare their proposal and Delphi and Planning Poker [16], while Desai et al. applied the M M R E and root-mean-squared error (RMSE) to validate ANN-COA with Other neural network-based techniques [17].
Last but not least, Asad Ali and Carmine Gravino (2019) studied the machine learning approaches employed in software effort prediction from 1991 to 2017 [28]. A total of 75 papers were selected after carefully evaluating the inclusion/exclusion and quality assessment filter. The M M R E , P R E D ( 0.25 ) , and M d M R E were studied in those papers. Out of the 75 papers chosen, 69 used the M M R E as a measure of accuracy, accounting for 92% of the total. The following highest percentages were the P R E D ( 0.25 ) at 69% and 47% for the M d M R E . They claimed that the M M R E and P R E D ( 0.25 ) are frequently employed as accuracy metrics in the papers they chose. In 2018, Gautam et al. examined software effort estimates published between 1981 and 2016 [29]. They provided a list of 56 publications that included the datasets used, validation techniques, performance metrics, statistical tests, and graphical analyses, where the M M R E and P R E D ( x ) take into account 32 out of 56 observations (57%).
As presented above, there are several criteria that researchers might consider for their proposals, such as the S A , M I B R E , M B R E , M A E , and R M S E . However, the M M R E and P R E D ( x ) are frequently used as evaluation criteria. Table 1 summarizes the several articles that adopted those criteria in validating SEE in recent years.

3. Sig Formula

This paper proposes a function s i g n ( y i , y i ^ ) expressed below by Equation (4). “i” represents the i-th item in the survey dataset. If the predicted value of the i-th item, denoted as y i ^ , is greater than the actual value of the i-th item, denoted as y i , the value of s i g n ( y i , y i ^ ) is set to +1. If the predicted value is less than the actual one, the value of s i g n ( y i , y i ^ ) is set to –1. If there is no difference between the predicted and actual values, then s i g n ( y i , y i ^ ) is setto be equal to zero.
Three new s i g formulas ( s i g L e f t ( y , y ^ ) , s i g R i g h t ( y , y ^ ) , and s i g ( y , y ^ ) ) are introduced by Equations (5)–(7). The double sum in each of Equations (5) and (6) is the sum of c u m u l a t i v e s u m of s i g n ( y i , y i ^ ) of the studied pairs. N is the number of all pairs.
s i g n ( y i , y i ^ ) = 1 if y i ^ > y i 0 if y i ^ = y i 1 y i ^ < y i
s i g L e f t ( y , y ^ ) = 2 N ( N + 1 ) k = 1 N i = 1 k s i g n ( y i , y i ^ )
s i g R i g h t ( y , y ^ ) = 2 N ( N + 1 ) k = 1 N i = 1 k s i g n ( y N i + 1 , y ^ N i + 1 )
s i g ( y , y ^ ) = s i g L e f t ( y , y ^ ) + s i g R i g h t ( y , y ^ ) 2
Suppose a coordinate system where actual values y i are on the horizontal axis and predicted values y i ^ are on the vertical axis. In Figure 1, there is such a system with a dashed line of the equation y i ^ = y i (further, this line will be addressed as a “baseline’) and some dots. Each dot ( y i , y i ^ ) represents a studied pair. Its first coordinate is the actual value, and its second coordinate is the predicted value.
Lemma 1.
Assume “ n ” is the number of dots above the baseline and “ m ” is the number of dots below the baseline; we have:
s i g ( y , y ^ ) = > 0 if n > m , < 0 if n < m , = 0 if n = m , o r   a l l d o t s l i e o n t h e b a s e l i n e
Proof. 
Let us denote shortly a i = s i g n ( y i , y i ^ ) ; Equations (5)–(7) will then be:
s i g L e f t ( y , y ^ ) = 2 N ( N + 1 ) k = 1 N i = 1 k s i g n ( y i , y ^ i ) = 2 N ( N + 1 ) k = 1 N i = 1 k a i
s i g L e f t ( y , y ^ ) = 2 N ( N + 1 ) i = 1 N ( N i + 1 ) a i .
s i g R i g h t ( y , y ^ ) = 2 N ( N + 1 ) k = 1 N i = 1 k s i g n ( y N i + 1 , y ^ N i + 1 )
s i g R i g h t ( y , y ^ ) = 2 N ( N + 1 ) k = 1 N i = 1 k a N i + 1 = 2 N ( N + 1 ) i = 1 N ( N i + 1 ) a N i + 1 .
s i g ( y , y ^ ) = s i g L e f t ( y , y ^ ) + s i g R i g h t ( y , y ^ ) 2 ,
s i g ( y , y ^ ) = 1 2 2 N ( N + 1 ) i = 1 N ( N + 1 ) a i = 1 N i = 1 N a i .
The sum in Equation (11) is the sum of all signs. As the sign of each dot is +1 if the dot lies above the baseline, –1 if it lies below the baseline, and 0 if the dot lies on the baseline, it is obvious that the sum in Equation (11) is greater than zero if the number of dots above the baseline is greater than the number of dots below.
Similarly, the sum is less than zero if the number of dots above the baseline is less than the number of dots below the baseline.
Finally, the sum is equal to zero either if the number of dots above the baseline is the same as the number of dots below the baseline or if all dots lie on the baseline. □
Lemma 2.
Let us suppose a chart with N dots such that not all of them lie below the baseline. Changing the y-coordinate of any dot with a position below the baseline (with its x-coordinate unchanged) such that this change causes the dot to move to a position above the baseline will result in an increase of the values of both s i g L e f t ( y , y ^ ) and s i g R i g h t ( y , y ^ ) .
Proof. 
Let us denote shortly again a i = s i g n ( y i , y i ^ ) and denote k ( 1 k N ) as the position of the k t h dot with the changed y-coordinate. The change means than that a k b e f o r e = s i g n ( y k , y ^ k ) = 1 will change to a k a f t e r = s i g n ( y k , y ^ k ) = + 1 . Then, according to Equation (9):
s i g L e f t b e f o r e ( y , y ^ ) = 2 N ( N + 1 ) [ N a 1 + ( N 1 ) a 2 + . . . + ( N k + 1 ) a k b e f o r e + . . . + 1 a N ) ]
s i g L e f t b e f o r e ( y , y ^ ) = 2 N ( N + 1 ) [ N a 1 + ( N 1 ) a 2 + . . . + ( N k + 1 ) × ( 1 ) + . . . + 1 a N ]
and
s i g L e f t a f t e r ( y , y ^ ) = 2 N ( N + 1 ) [ N a 1 + ( N 1 ) a 2 + . . . + ( N k + 1 ) a k a f t e r + . . . + 1 a N ) ]
s i g L e f t a f t e r ( y , y ^ ) = 2 N ( N + 1 ) [ N a 1 + ( N 1 ) a 2 + . . . + ( N k + 1 ) × ( + 1 ) + . . . + 1 a N ]
Thus:
The values of s i g L e f t b e f o r e ( y , y ^ ) and s i g L e f t a f t e r ( y , y ^ ) differ only in a term ( N k + 1 ) × ( 1 ) , respectively ( N k + 1 ) × ( + 1 ) . It is obvious that ( N k + 1 ) × ( 1 ) < ( N k + 1 ) × ( + 1 ) , because:
N + k 1 < N k + 1 ,
2 k < 2 N + 2 ,
k < N + 1 , as the assumption was k N .
Similarly, according to Equation (10),
s i g R i g h t b e f o r e ( y , y ^ ) = 2 N ( N + 1 ) [ N a N + ( N 1 ) a N 1 + . . . + ( k ) a k b e f o r e + . . . + 1 a 1 ) ]
s i g R i g h t b e f o r e ( y , y ^ ) = 2 N ( N + 1 ) [ N a N + ( N 1 ) a N 1 + . . . + ( k ) × ( 1 ) + . . . + 1 a 1 ]
and
s i g R i g h t a f t e r ( y , y ^ ) = 2 N ( N + 1 ) [ N a N + ( N 1 ) a N 1 + . . . + ( k ) a k a f t e r + . . . + 1 a 1 ) ]
s i g R i g h t a f t e r ( y , y ^ ) = 2 N ( N + 1 ) [ N a N + ( N 1 ) a N 1 + . . . + ( k ) × ( + 1 ) + . . . + 1 a 1 ]
The values of s i g R i g h t b e f o r e ( y , y ^ ) and s i g R i g h t a f t e r ( y , y ^ ) differ only in a term ( k ) × ( 1 ) , respectively ( k ) × ( + 1 ) . Again, it is obvious that k × ( 1 ) < k × ( + 1 ) , because:
k < k 0 < 2 k 0 < k , as the assumption was 1 k .
We conclude that if the number of consecutive dots above the baseline increases, the values of both s i g L e f t ( y , y ^ ) and s i g R i g h t ( y , y ^ ) increase as well.
Remark: A similar lemma could be proven: If the number of consecutive dots below the baseline increases, the values of both s i g L e f t ( y , y ^ ) and s i g R i g h t ( y , y ^ ) decrease. □

4. Characteristics of Sig Formula

Let us suppose a chart where the horizontal axis presents the actual values y i and the vertical axis stands for the expected values y i ^ . The baseline in such a chart would be a straight line with the equation y ^ = y :
  • If all predicted values are greater than all the respective actual values (i.e., all dots are above the baseline), then, according to Equation (11):
    s i g ( y , y ^ ) = 1 N i = 1 N a i = 1 N i = 1 N ( + 1 ) = 1
    which is the maximal possible value of s i g ( y , y ^ ) .
  • If all predicted values are smaller than all the respective actual values (i.e. all dots are below the baseline), then, according to Equation (11):
    s i g ( y , y ^ ) = 1 N i = 1 N a i = 1 N i = 1 N ( 1 ) = 1
    which is the minimal possible value of s i g ( y , y ^ ) .
  • If there is no difference between each predicted value and the respective actual value (i.e., all dots are on the baseline), then, according to Equation (11):
    s i g ( y , y ^ ) = 1 N i = 1 N a i = 1 N i = 1 N ( 0 ) = 0
  • In general, the predicted and actual values fluctuate. Some predicted values might be greater than and others might be smaller than the respective actual ones. The dots in Figure 1 then form a cloudy distribution around the baseline. The value of s i g ( y , y ^ ) lies then between –1 and +1.
  • Moreover, in the case of a uniform symmetry (named UniSym), i.e., if the values alternate around the baseline (one dot is above the baseline, the next one below, the next above, etc.), then with the increasing number N of the observations, the value of s i g ( y , y ^ ) approaches 0.
    lim N s i g ( y , y ^ ) = lim N 1 N i = 1 N a i = 0

5. Research Questions

These research questions (RQs) should be answered in this study:
  • RQ1: What are the difference of the s i g L e f t , s i g R i g h t , and s i g formulas?
  • RQ2: What is the importance of additional information related to the P R E D ( x ) and M M R E in measuring the performance of the predicted model?

6. Results and Discussion

In this section, we discuss the usefulness of additional information related to the P R E D ( x ) and M M R E based on twenty-eight projects (Dataset-1) collected from [30,31,32,33] with eight different predicted assumptions as given in Table 2. The column “Real_P20” represents real efforts in terms of use case points (UCPs) [30,34]. The other columns (Model 1–Model 8) contain the corresponding estimated efforts. The UCPs were originally designed by Karner [35] as a simplification of the functional points method [36]. Size estimation is based on the UML use case model. The UCPs are used in industry or in research. Many studies map effort methods, including UCPs and UCP modification [7,37,38,39,40]. Use case points faced some design issues, which were discussed by Ouwerkerk and Abran [41]. The UCP design flaws are mainly based on scale transformation when the values of the UCP components are calculated [42].
The presented models simulate the possible prediction model’s behavior based on real scenarios. The models’ behavior is configured as follows:
  • Model 1: The predicted efforts are random guessing such that their values are mostly greater than the real values, and the P R E D ( 0.25 ) reaches the maximum compared with Models 5, 6, 7, and 8.
  • Model 2: The predicted values are random guessing, as opposed to Model 1, where the P R E D ( 0.25 ) is equal to the P R E D ( 0.25 ) obtained from Model 1.
  • Model 3: The predicted efforts are random guessing such that the first half of the predicted values is mostly less than the real values, but the remaining are mostly greater than the real values, where the P R E D ( 0.25 ) is assumed equal to the P R E D ( 0.25 ) obtained from Model 1.
  • Model 4: It is assumed to be similar to Model 3 in the inverse sense. Furthermore, the predicted values in this model were purposefully chosen to minimize the M M R E .
  • Models 5, 6, 7, and 8: The predicted efforts are based on the rule that one or more initial predicted values are greater/less than the actual values, and one or more subsequent predicted values are greater/less than the actual values. The following predicted values follow the same sequence as the previous ones. The rule is repeated until the testing dataset is exhausted. Furthermore, we assumed their P R E D ( 0.25 ) are the same; they are greater than 0.7, but less than P R E D ( 0.25 ) , and their M M R E is greater than or equal to Models 1, 2, 3, and 4.
As for the predictions (estimated values):
  • The predicted values produced by Model 1 are mostly greater than the corresponding real values (all dots lie above the baseline), while Model 2 is the opposite.
  • In Model 3, the dots in the first half mostly lie below and the dots in the second half mostly lie above the baseline, and vice versa in Model 4.
  • The dots in Models 5 and 6 lie around the baseline, but the number of dots above the baseline is greater than the number of dots below the baseline (systematically overestimated). This is to demonstrate that the s i g value is positive.
  • The dots in Model 7 also lie around the baseline, but the number of dots above the baseline is smaller than the number of dots below the baseline. This is to demonstrate that the s i g value is negative (systematically underestimated). Moreover, Model 8 is an approximately UniSym model. The s i g value then reaches zero.
Figure 2 illustrates the scatter of predicted efforts obtained from Models 1 and 2 vs. actual efforts around the baseline. These scatters are generated by using the “scatter” function from the “matplotlib” library of Python. By default, this function automatically re-orders the values from the minimum to maximum values. Figure 3, Figure 4 and Figure 5 below are used in the same manner. As seen in these figures, most of the dots (a pair of actual and predicted values) in Model 1 are above the baseline (where the predicted and actual values are the same, y i ^ = y i ), resulting in s i g L e f t , s i g R i g h t , and s i g reaching +1.
In a similar manner in Model 2, all of the dots are below the baseline; therefore, s i g L e f t , s i g R i g h t , and s i g reach −1 (see Table 3). In practice, these assumptions are improbable because researchers are constantly looking for the most-efficient technique to obtain the best model. In some circumstances, the scatter between the real and predicted efforts might result in a case where the dots will lie around the baseline.
Another scenario is shown in Models 3 and 4 (see Figure 3). The s i g values calculated from both models reach zero. However, the scatter of the predicted and real values might not be realistic. In Model 3, the first half of the dots is mostly below the baseline, which results in s i g L e f t = 0.488 ( < 0 ) , while the second half of dots is mostly above the baseline s i g R i g h t = 0.345 ( > 0 ) (see Table 3). This outcome is contrary to Model 4. Therefore, besides using only the s i g value, we might consider adding s i g L e f t and s i g R i g h t to gain an insight. On the other hand, Figure 4 presents another case (Models 5 and 6) where the values of s i g L e f t and s i g R i g h t are useful to be used together with s i g , and the discussion below might be the answer for RQ1 regarding the difference of s i g L e f t , s i g R i g h t , and s i g . As can be seen, Models 5 and 6 are fairly random patterns around the baseline (see Table 4). Those scatters lead to a result that the absolute values of the s i g R i g h t obtained from Models 5 and 6 are larger than the absolute values of s i g L e f t (see Table 3). Furthermore, Model 6 is more symmetrical than Model 5, and the predicted efforts obtained from the projects P10, P11, and P12 in Model 5 show the different positive/negative errors compared with Model 6 (see Table 4), which leads to the fact that the absolute value of the s i g L e f t obtained from Model 5 is larger than that obtained from Model 6. Moreover, the s i g R i g h t obtained from Model 5 is greater than the s i g R i g h t obtained from Model 6, which might reveal that the number of dots on the same side of the baseline in the second half of Model 5 is greater than in the second half of Model 6 (see Lema 2). This information might be interesting because, if we use Model 5 as the predicted model, we should consider when we want to predict effort estimation with high values, as these may be overestimated.
Moreover, as shown in Figure 5, Models 7 and 8 randomly scatter between the predicted efforts against the actual efforts, whereas Model 8 is slightly more randomly scattered than Model 7. It is clear that the predicted efforts from P1 to P12 and P17 to P28 in both models have the same trend, but there are slight differences from P13 to P16 in Model 7 compared with Model 8 (see Table 4). The dots obtained from those efforts in Model 7 are below the baseline, which leads to the absolute value of s i g L e f t , s i g R i g h t attained from Model 7 being higher than those attained from Model 8. In addition, the number of dots above the baseline in both models is less than the number of dots below it. As a result, the calculated s i g values in both models are less than zero, while the absolute value of s i g obtained from Model 7 is less than that obtained from Model 8.
Table 3 summarizes the statistical metrics from the eight models, including s i g L e f t , s i g R i g h t , s i g , the M M R E , and the P R E D ( 0.25 ) . As can be seen, the M M R E and P R E D ( 0.25 ) obtained from Model 4 gain the best performance due to its M M R E reaching the minimum and its P R E D ( 0.25 ) reaching the maximum compared with the other models. Based on these metrics, this model might be the most suitable among the remaining models. Unfortunately, its specific information s i g L e f t / s i g R i g h t values (obtained 0.527/−0.384) are farther from zero in comparison to the corresponding values obtained in Models 5 to 8 (see Lema 2). Thus, if we want to choose the suitable predictive model among the discussed eight models, we might consider choosing Model 5, 6, 7, or 8 due to the additional information related to the M M R E and P R E D ( 0.25 ) obtained from those models being slightly closer to zero. This scenario might be a good example to answer RQ2; the values of s i g L e f t , s i g R i g h t , and s i g might be the useful additional information related to the M M R E and P R E D ( 0.25 ) .
On the other hand, we suppose that we are considering two predictive models, namely Models 7 and 8, and we want to decide which one is better. We noticed that the M M R E and P R E D ( 0.25 ) obtained from both are the same, 0.175 and 0.76, respectively. As discussed above, Model 8 has a fairly random scatter than Model 7 due to the s i g L e f t , s i g R i g h t , and s i g values obtained from Model 8 being smaller than those obtained from Model 7. Based on these findings, we might conclude that Model 8 outperforms Model 7.
Last but not least, s i g L e f t , s i g R i g h t , and s i g have a weakness, that is, if out of 28 predictions, 14 are underestimated by 1% and the remaining 14 are overestimated by 100%, the method proposed will indicate the prediction errors as symmetrical. This issue might raise confusion for software project practitioners. However, as mentioned in the purpose of these formulas, they are only additional information related to the M M R E and P R E D ( x ) , we still rely on the criteria of the M M R E and P R E D ( x ) with more information obtained from these formulas as useful information. As discussed above, if we omit s i g L e f t , s i g R i g h t , and s i g , this might lead to the selection of an improper model.
Case study: To be more specific, we verified the benefit of the s i g L e f t , s i g R i g h t , and s i g values based on the models obtained from XGBoost regression. XGBoost is a gradient-boosted decision tree development created for a highly efficient and accurate model. It is a Python open-source library, and it provides a framework such as the “fit” function to build a model based on the training dataset and the “predict” function to predict new values based on a new dataset [43]. In our scenario, we adopted XGBoost with a group of parameters. They include the learning rate (0.01), booster (“gblinear”), and n_estimators chosen by experimentation. We used Dataset-1 (twenty-eight projects) [30,31,32,33] and Dataset-2 (seventy-one projects) [31] as the historical datasets. Those datasets measure effort estimation in terms of use case points (UCPs) [30,34]. Both datasets are described by the unadjusted actor weight (UAW), unadjusted use case weight (UUCW), technical complexity factor (TCF), environmental complexity factor (ECF), and Real-P20. Real-P20 is considered as the dependent variable, and the UAW, UUCW, TCF, and ECF are considered as the independent variables.
Dataset-1/Dataset-2 is divided into two datasets, 80% of projects for training and the remaining projects for testing. As mentioned in Table 1, the P R E D ( x ) an M M R E are used as the criteria to measure the accuracy of effort estimation in UCPs. In addition, the values of s i g L e f t , s i g R i g h t , and s i g are shown.
Table 5 shows the evaluation criteria obtained from the XGBoost models. As can be seen, the s i g value obtained from the proposed models based on Dataset-1 and Dataset-2 is −0.556 and 0.286, respectively. Criterion s i g shows a model scatter (variance). If the variance reaches zero, that model is better. In those two datasets, we can see that model XGBoost has a tendency to underestimate Dataset-1 due to the values of s i g L e f t , s i g R i g h t , and s i g being less than and so far from zero. The scatter of the model for Dataset-2 might be smaller than for Dataset-1 due to the absolute value of the s i g obtained from the model for Dataset-1 is larger than that obtained from the model for Dataset-2. As can be seen, the absolute values of s i g L e f t are similar for Dataset-1 and Dataset-2. However, the | s i g R i g h t | obtained from Dataset-2 is smaller than that obtained from Dataset-1, which demonstrates that the scatter of the model for Dataset-2 might be smaller than for Dataset-1. These findings reveal that the s i g L e f t , s i g R i g h t , and s i g values might be used as reference criteria to consider choosing an appropriate predictive model. Compared to the M M R E and P R E D ( 0.20 ) , which is an absolute sig value, we might be confident to state that the XGBoost model for Dataset-2 outperforms that for Dataset-1.

7. Conclusions and Future Work

s i g L e f t , s i g R i g h t , and s i g refer to the sign of prediction error, not the magnitude. These values’ usefulness might be considered a contribution to the M M R E and P R E D ( x ) . Using the M M R E and P R E D ( x ) together with them as performance indicators could be beneficial to validate the symmetry of the predictive model values around the baseline. On the other hand, since those values indicate under-/over-estimation in terms of effort estimation independent of the used sizing unit, these criteria might be useful when validating the accuracy of the predicted model along with the P R E D ( x ) and M M R E .
Moreover, as mentioned in Lemma 2, the lower the absolute value of s i g L e f t and s i g R i g h t , the more homogeneous the distribution of the predicted and actual values around the baseline. Using s i g L e f t , s i g R i g h t , and s i g might detect whether the predictive model is under or over the baseline. Although this study proposes signals to identify the estimated model as under or over the baseline based on the limited dataset, based on the proof of Lemmas 1 and 2, the proposed formulas are completely adopted for the larger dataset. Last but not least, as discussed in Section 4, based on the adjusted values of s i g L e f t , s i g R i g h t , and s i g , the predictive model might be adjusted by increasing or decreasing the intercept, which might lead to a higher accuracy of the predictive model.
Limitations: Based on the s i g L e f t , s i g R i g h t , and s i g values, we just stated whether the model is over or under the baseline. It may be more helpful when the quantification of this over-/under-estimation is discussed further in the future.

Author Contributions

Conceptualization, H.H.T., P.S. and R.S.; methodology, H.H.T., P.S., M.F., Z.P. and R.S.; software, H.H.T.; validation, H.H.T. and M.F.; investigation, H.H.T., P.S., M.F., Z.P. and R.S.; resources, H.H.T., P.S., M.F., Z.P. and R.S.; data curation, H.H.T. and P.S.; writing—original draft preparation, H.H.T. and M.F.; writing—review and editing, H.H.T., P.S., R.S. and Z.P.; visualization, H.H.T., P.S., M.F., Z.P. and R.S.; supervision, P.S., Z.P. and R.S.; project administration, P.S., Z.P. and R.S.; funding acquisition, P.S., Z.P. and R.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Faculty of Applied Informatics, Tomas Bata University in Zlin, under Project No. RVO/FAI/2021/002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This investigation collected datasets from Ochodek et al [32], Subriadi and Ningrum [33], and Silhavy et al. [30].

Acknowledgments

We would like to thank Hoang Lê Minh, Hông Vân Lê, and Nguyen Tien Zung for helpful comments. We would also like to express our very great appreciation to Susely Figueroa Iglesias for her insightful suggestions and careful reading of the manuscript. We would also like to extend our thanks to the Torus Actions company (https://torus.ai (accessed on 1 November 2022)) for allowing us to spend time completing the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Azzeh, M.; Nassif, A.B. Project productivity evaluation in early software effort estimation. J. Softw. Evol. Process. 2018, 30, e21110. [Google Scholar] [CrossRef]
  2. Braz, M.R.; Vergilio, S.R. Software effort estimation based on use cases. In Proceedings of the 30th Annual International Computer Software and Applications Conference (COMPSAC’06), Chicago, IL, USA, 17–21 September 2006; Volume 1, pp. 221–228. [Google Scholar]
  3. Mahmood, Y.; Kama, N.; Azmi, A.; Khan, A.S.; Ali, M. Software effort estimation accuracy prediction of machine learning techniques: A systematic performance evaluation. Softw. Pract. Exp. 2022, 52, 39–65. [Google Scholar] [CrossRef]
  4. Munialo, S.W.; Muketha, G.M. A review of agile software effort estimation methods. Int. J. Comput. Appl. Technol. Res. 2016, 5, 612–618. [Google Scholar] [CrossRef]
  5. Silhavy, R.; Silhavy, P.; Prokopova, Z. Evaluating subset selection methods for use case points estimation. Inf. Softw. Technol. 2018, 97, 1–9. [Google Scholar] [CrossRef]
  6. Trendowicz, A.; Jeffery, R. Software project effort estimation. Found. Best Pract. Guidel. Success Constr. Cost Model. 2014, 12, 277–293. [Google Scholar]
  7. Azzeh, M.; Nassif, A.B.; Attili, I.B. Predicting software effort from use case points: Systematic review. Sci. Comput. Program. 2021, 204, 102596. [Google Scholar] [CrossRef]
  8. Conte, S.D.; Dunsmore, H.E.; Shen, Y.E. Software Engineering Metrics and Models; Benjamin-Cummings Publishing Co., Inc.: San Francisco, CA, USA, 1986. [Google Scholar]
  9. Praynlin, E. Using meta-cognitive sequential learning neuro-fuzzy inference system to estimate software development effort. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 8763–8776. [Google Scholar] [CrossRef]
  10. Fadhil, A.A.; Alsarraj, R.G.H.; Altaie, A.M. Software cost estimation based on dolphin algorithm. IEEE Access 2020, 8, 75279–75287. [Google Scholar] [CrossRef]
  11. Bilgaiyan, S.; Mishra, S.; Das, M. Effort estimation in agile software development using experimental validation of neural network models. Int. J. Inf. Tecnol. 2019, 11, 569–573. [Google Scholar] [CrossRef]
  12. Mustapha, H.; Abdelwahed, M. Investigating the use of random forest in software effort estimation. Procedia Comput. Sci. 2019, 148, 343–352. [Google Scholar]
  13. Ullah, A.; Wang, B.; Sheng, J.; Long, J.; Asim, M.; Riaz, F. A Novel Technique of Software Cost Estimation Using Flower Pollination Algorithm. In Proceedings of the 2019 International Conference on Intelligent Computing, Automation and Systems (ICICAS), Chongqing, China, 6–8 December 2019; pp. 654–658. [Google Scholar]
  14. Sethy, P.K.; Rani, S. Improvement in cocomo modal using optimization algorithms to reduce mmre values for effort estimation. In Proceedings of the 2019 4th International Conference on Internet of Things: Smart Innovation and Usages (IoT-SIU), Ghaziabad, India, 18–19 April 2019; pp. 1–4. [Google Scholar]
  15. Effendi, Y.A.; Sarno, R.; Prasetyo, J. Implementation of bat algorithm for cocomo ii optimization. In Proceedings of the 2018 International Seminar on Application for Technology of Information and Communication, Semarang, Indonesia, 21–22 September 2018; pp. 441–446. [Google Scholar]
  16. Khan, M.S.; ul Hassan, C.A.; Shah, M.A.; Shamim, A. Software cost and effort estimation using a new optimization algorithm inspired by strawberry plant. In Proceedings of the 2018 24th International Conference on Automation and Computing (ICAC), Newcastle upon Tyne, UK, 6–7 September 2018; pp. 1–6. [Google Scholar]
  17. Desai, V.S.; Mohanty, R. Ann-cuckoo optimization technique to predict software cost estimation. In Proceedings of the 2018 Conference on Information and Communication Technology (CICT), Jabalpur, India, 26–28 October 2018; pp. 1–6. [Google Scholar]
  18. Jørgensen, M.; Halkjelsvik, T.; Liestøl, K. When should we (not) use the mean magnitude of relative error (mmre) as an error measure in software development effort estimation? Inf. Softw. Technol. 2022, 143, 106784. [Google Scholar] [CrossRef]
  19. Foss, T.; Stensrud, E.; Kitchenham, B.; Myrtveit, I. A simulation study of the model evaluation criterion mmre. IEEE Trans. Softw. Eng. 2003, 29, 985–995. [Google Scholar] [CrossRef]
  20. Myrtveit, I.; Stensrud, E.; Shepperd, M. Reliability and validity in comparative studies of software prediction models. IEEE Trans. Softw. Eng. 2005, 31, 380–391. [Google Scholar] [CrossRef] [Green Version]
  21. Kitchenham, B.A.; Pickard, L.M.; MacDonell, S.G.; Shepperd, M.J. What accuracy statistics really measure [software estimation]. IEEE-Proc. Softw. 2001, 148, 81–85. [Google Scholar] [CrossRef] [Green Version]
  22. Shepperd, M.; MacDonell, S. Evaluating prediction systems in software project estimation. Inf. Softw. Technol. 2012, 54, 820–827. [Google Scholar] [CrossRef] [Green Version]
  23. Villalobos-Arias, L.; Quesada-Lopez, C.; Guevara-Coto, J.; Martınez, A.; Jenkins, M. Evaluating hyper-parameter tuning using random search in support vector machines for software effort estimation. In Proceedings of the 16th ACM International Conference on Predictive Models and Data Analytics in Software Engineering, Virtual, 8–9 November 2020; pp. 31–40. [Google Scholar]
  24. Idri, A.; Hosni, M.; Abran, A. Improved estimation of software development effort using classical and fuzzy analogy ensembles. Appl. Soft Comput. 2016, 49, 990–1019. [Google Scholar] [CrossRef]
  25. Gneiting, T. Making and evaluating point forecasts. J. Am. Stat. Assoc. 2011, 106, 746–762. [Google Scholar] [CrossRef] [Green Version]
  26. Strike, K.; El Emam, K.; Madhavji, N. Software cost estimation with incomplete data. IEEE Trans. Softw. Eng. 2001, 27, 890–908. [Google Scholar] [CrossRef] [Green Version]
  27. Hamid, M.; Zeshan, F.; Ahmad, A.; Ahmad, F.; Hamza, M.A.; Khan, Z.A.; Munawar, S.; Aljuaid, H. An intelligent recommender and decision support system (irdss) for effective management of software projects. IEEE Access 2020, 8, 140752–140766. [Google Scholar] [CrossRef]
  28. Ali, A.; Gravino, C. A systematic literature review of software effort prediction using machine learning methods. J. Softw. Evol. Process. 2019, 31, e2211. [Google Scholar] [CrossRef]
  29. Gautam, S.S.; Singh, V. The state-of-the-art in software development effort estimation. J. Softw. Evol. Process. 2018, 30, e1983. [Google Scholar] [CrossRef]
  30. Silhavy, R.; Silhavy, P.; Prokopova, Z. Algorithmic optimisation method for improving use case points estimation. PLoS ONE 2015, 10, e0141887. [Google Scholar] [CrossRef] [PubMed]
  31. Silhavy, R.; Silhavy, P.; Prokopova, Z. Analysis and selection of a regression model for the use case points method using a stepwise approach. J. Syst. Softw. 2017, 125, 1–14. [Google Scholar] [CrossRef] [Green Version]
  32. Ochodek, M.; Nawrocki, J.; Kwarciak, K. Simplifying effort estimation based on use case points. Inf. Softw. Technol. 2011, 53, 200–213. [Google Scholar] [CrossRef]
  33. Subriadi, A.P.; Ningrum, P.A. Critical review of the effort rate value in use case point method for estimating software development effort. J. Theroretical Appl. Inf. Technol. 2014, 59, 735–744. [Google Scholar]
  34. Hoc, H.T.; Hai, V.V.; Nhung, H.L.T.K. Adamoptimizer for the Optimisation of Use Case Points Estimation; Springer: Berlin/Heidelberg, Germany, 2020; pp. 747–756. [Google Scholar]
  35. Karner, G. Metrics for Objectory; No. LiTH-IDA-Ex-9344; University of Linkoping: Linkoping, Sweden, 1993; p. 21. [Google Scholar]
  36. ISO/IEC 20926:2009; Software and Systems Engineering—Software Measurement—IFPUG Functional Size Measurement Method. ISO/IEC: Geneva, Switzerland, 2009.
  37. Azzeh, M.; Nassif, A.; Banitaan, S. Comparative analysis of soft computing techniques for predicting software effort based use case points. Iet Softw. 2018, 12, 19–29. [Google Scholar] [CrossRef]
  38. Santos, R.; Vieira, D.; Bravo, A.; Suzuki, L.; Qudah, F. A systematic mapping study on the employment of neural networks on software engineering projects: Where to go next? J. Softw. Evol. Process. 2022, 34, e2402. [Google Scholar] [CrossRef]
  39. Carbonera, E.C.; Farias, K.; Bischoff, V. Software development effort estimation: A systematic mapping study. IET Softw. 2020, 14, 328–344. [Google Scholar] [CrossRef]
  40. Idri, A.; Hosni, M.; Abran, A. Systematic literature review of ensemble effort estimation. J. Syst. Softw. 2016, 118, 151–175. [Google Scholar] [CrossRef]
  41. Ouwerkerk, J.; Abran, A. An evaluation of the design of use case points (UCP). In Proceedings of the International Conference On Software Process And Product Measurement MENSURA, Cádiz, Spain, 6–8 November 2006; pp. 83–97. [Google Scholar]
  42. Abran, A. Software Metrics and Software Metrology; John Wiley & Sons: Hoboken, NJ, USA, 2010. [Google Scholar]
  43. Geron, A. Ensemble Learning and Random Forests; O’reilly: Sebastopol, CA, USA, 2019; pp. 189–212. [Google Scholar]
Figure 1. A sample of the distribution of predicted and actual values around the baseline.
Figure 1. A sample of the distribution of predicted and actual values around the baseline.
Mathematics 10 04649 g001
Figure 2. The scatter of predictive efforts vs. actual efforts: Model 1 vs. 2.
Figure 2. The scatter of predictive efforts vs. actual efforts: Model 1 vs. 2.
Mathematics 10 04649 g002
Figure 3. The scatter of predictive efforts vs. actual efforts: Model 3 vs. 4.
Figure 3. The scatter of predictive efforts vs. actual efforts: Model 3 vs. 4.
Mathematics 10 04649 g003
Figure 4. The scatter of predictive efforts vs. actual efforts: Model 5 vs. 6.
Figure 4. The scatter of predictive efforts vs. actual efforts: Model 5 vs. 6.
Mathematics 10 04649 g004
Figure 5. The scatter of predictive efforts vs. actual efforts: Model 7 vs. 8.
Figure 5. The scatter of predictive efforts vs. actual efforts: Model 7 vs. 8.
Mathematics 10 04649 g005
Table 1. Summary of reviewed related papers in recent years, where the MMRE was adopted.
Table 1. Summary of reviewed related papers in recent years, where the MMRE was adopted.
AuthorsProposal ModelCompare WithCriteria
Mahmood et al. [3]Machine-learning-based ensemble techniquesMachine-learning-based solo techniques M M R E , P R E D ( x )
Praynlin [9]Metacognitive neuro-fuzzyParticle swarm optimization, genetic algorithm, and backpropagation network M M R E , P R E D ( x ) , and other evaluation criteria
Fadhil et al. [10]DolBatCOCOMO II M M R E , P R E D ( x )
Bilgaiyan, S. et al. [11]Feedforward backpropagation NNCascade correlation NN, Elman NN M M R E , P R E D ( x ) , M S E
Mustapha et al. [12]RFClassical regression tree M M R E , P R E D ( x ) , M d M R E
Ullah et al. [13]Flower pollination algorithmCOCOMO-II M M R E
Sethy and Rani [14]TLBOBailey, COCOMO 2, Halstead, SEL BCO M M R E
Effendi et al. [15]Optimization of COCOMO II constantsCOCOMO II M M R E
Khan et al. [16]Optimization of COCOMOCOCOMO M M R E
Desai and Mohanty [17]ANN-COAOther neural network-based techniques M M R E , R M S E
Table 2. The twenty-eight projects and eight different predicted assumptions.
Table 2. The twenty-eight projects and eight different predicted assumptions.
EFFORT (Person-Hours)
No. Estimation ( y ^ )
Real_P20Model 1Model 2Model 3Model 4Model 5Model 6Model 7Model 8
P1151.85157.2135.9108.4157148149133145
P295.851019077.2102101102105.9101.9
P358.6564485160.3555339.849
P437.1393530474165.24739.9
P530.754.6242834172518.220
P624.6468215531.9332848
P713.851556158988.2
P8179.65186171175.3191.9206200189196
P984.058382.368.210378747276
P1067.27252528681849697
P116180.541466464495844
P123642161639.252585844.9
P1325.731138.73422.6241817
P1419.852522.81925283115.318
P15184.2193182197170157176.3181179
P1699110931149610310095103
P17197.5203180207183192169193178
P1896.2510281.710183101111105109
P19108.75115931099372789393.1
P20111.312999.2115113.8135.8118116118
P21132145118145.9113105103109113
P22128.4135.4119129111132148.1135.7143
P23152.1172151155138123151137149
P2484.811474108.6828181105.9105
P25183.5195177230.2171193.2200157167.2
P26143172126172128.4161161148152
P27137156.1121.6154117148148123132
P28168175150165160192179181174
Table 3. The summary of the s i g values among the eight models.
Table 3. The summary of the s i g values among the eight models.
MetricsModel 1Model 2Model 3Model 4Model 5Model 6Model 7Model 8
s i g L e f t 0.901−0.926−0.4880.5270.059−0.03−0.172−0.108
s i g R i g h t 0.956−0.9310.345−0.3840.2270.172−0.113−0.034
Sig0.929−0.929−0.0720.0720.1430.071−0.143−0.071
MMRE0.1630.1790.1750.1560.1630.1860.1750.175
PRED(0.25)0.800.800.800.800.760.760.760.76
Table 4. The difference between real efforts compared with predicted models.
Table 4. The difference between real efforts compared with predicted models.
No.Model 1Model 2Model 3Model 4Model 5Model 6Model 7Model 8
P1−5.3515.9543.45−5.153.852.8518.856.85
P2−5.155.8518.65−6.15−5.15−6.15−10.05−6.05
P3−5.3510.657.65−1.653.655.6518.859.65
P4−1.92.17.1−9.9−3.9−28.1−9.9−2.8
P5−23.96.72.7−3.313.75.712.510.7
P6−21.416.63.6−30.4−7.3−8.4−3.4−23.4
P7−1.158.857.85−1.155.854.855.855.65
P8−6.358.654.35−12.25−26.35−20.35−9.35−16.35
P91.051.7515.85−18.956.0510.0512.058.05
P10−4.815.215.2−18.8−13.8−16.8−28.8−29.8
P11−19.52015−3−312317
P12−62020−3.2−16−22−22−8.9
P13−5.312.717−8.33.11.77.78.7
P14−5.15−2.950.85−5.15−8.15−11.154.551.85
P15−8.82.2−12.814.227.27.93.25.2
P16−116−153−4−14−4
P17−5.517.5−9.514.55.528.54.519.5
P18−5.7514.55−4.7513.25−4.75−14.75−8.75−12.75
P19−6.2515.75−0.2515.7536.7530.7515.7515.65
P20−17.712.1−3.7−2.5−24.5−6.7−4.7−6.7
P21−1314−13.91927292319
P22−79.4−0.617.4−3.6−19.7−7.3−14.6
P23−19.91.1−2.914.129.11.115.13.1
P24−29.210.8−23.82.83.83.8−21.1−20.2
P25−11.56.5−46.712.5−9.7−16.526.516.3
P26−2917−2914.6−18−18−5−9
P27−19.115.4−1720−11−11145
P28−71838−24−11−13−6
Table 5. Evaluation criteria obtained from XGBoost.
Table 5. Evaluation criteria obtained from XGBoost.
Dataset-1Dataset-2
R-Squared0.8840.974
MMRE0.3180.136
PRED(0.20)0.2220.714
s i g L e f t −0.4220.429
s i g R i g h t −0.6890.143
s i g −0.5560.286
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huynh Thai, H.; Silhavy, P.; Fajkus, M.; Prokopova, Z.; Silhavy, R. Propose-Specific Information Related to Prediction Level at x and Mean Magnitude of Relative Error: A Case Study of Software Effort Estimation. Mathematics 2022, 10, 4649. https://doi.org/10.3390/math10244649

AMA Style

Huynh Thai H, Silhavy P, Fajkus M, Prokopova Z, Silhavy R. Propose-Specific Information Related to Prediction Level at x and Mean Magnitude of Relative Error: A Case Study of Software Effort Estimation. Mathematics. 2022; 10(24):4649. https://doi.org/10.3390/math10244649

Chicago/Turabian Style

Huynh Thai, Hoc, Petr Silhavy, Martin Fajkus, Zdenka Prokopova, and Radek Silhavy. 2022. "Propose-Specific Information Related to Prediction Level at x and Mean Magnitude of Relative Error: A Case Study of Software Effort Estimation" Mathematics 10, no. 24: 4649. https://doi.org/10.3390/math10244649

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop