Next Article in Journal
Investigations on the Complex Band Diagram of Flexural Wave through the Fluid-Loaded Phononic Plate
Previous Article in Journal
Estimation and Adjustment of Interface Stiffnesses for Machine-Tool
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Software Reliability Growth Model with Dependent Failures and Uncertain Operating Environments

1
Department of Computer Science and Statistics, Chosun University, 146, Chosundae-gil, Dong-gu, Gwangju 61452, Republic of Korea
2
Department of Industrial and Systems Engineering, Rutgers University, 96 Frelinghuysen Road, Piscataway, NJ 08855-8018, USA
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(23), 12383; https://doi.org/10.3390/app122312383
Submission received: 2 November 2022 / Revised: 23 November 2022 / Accepted: 1 December 2022 / Published: 3 December 2022

Abstract

:
Software is used in various industries, and its reliability has become an extremely important issue. For example, in the medical industry, software is used to provide medical services to underprivi-leged individuals. If a problem occurs with the software reliability, incorrect medical information may be provided. The software reliability is estimated using a software reliability growth model. However, most software reliability growth models assume that the failures are independent. In addition, it is assumed that the test and operating environments are the same. In this study, we propose a new software reliability growth model that assumes that software failures are dependent and uncertain operating environments. A comparison of the proposed model against existing NHPP SRMEs using actual datasets shows that the proposed model achieves the best fit.

1. Introduction

Since software is used in various industries, software reliability is an extremely important issue. For example, in the medical industry, software is used to provide medical services to underprivileged individuals. If a problem with software reliability occurs, incorrect medical information may be provided. Thus, the importance of software reliability cannot be overemphasized.
The software used in various industries has many different functions and its structure is complex. In addition, the reliability (stability) of both software and hardware is extremely important. Software reliability issues can cause financial and human losses. Thus, many researchers have been studying software reliability for decades. To conduct software reliability studies, software reliability growth models (SRGM) should be understood. A SRGM is a tool used to estimate the quality and reliability of software products and provide information to developers and consumers. Consumers can refer to such information to select products for purchase, and developers can judge the reliability of a product and efficiently manage the development process.
In most existing SRGMs, the mean value function is considered to follow a non-homogeneous Poisson process (NHPP). Each model has a unique mean value function that considers the failure intensity, detection rate, number of remaining failures, and various environments (e.g., development, testing, and operation). In particular, the form and parameters of the mean value function (m(t)) are significantly affected by the environments and assumptions made.
Previous studies have considered various SRGMs when discussing software reliability. The model proposed by Goel and Okumoto [1] is the most fundamental approach used in research on SRGMs and has become the focal point of various studies. Yamada et al. [2] proposed a model using an S-shaped curve. There have also been studies on models that reflect not only the shape of the function, but also various environments. Pham et al. [3,4] introduced models that consider an error introduction rate that follows an exponential function and a generalized SRGM that reflects the testing coverage. Song et al. [5] proposed a model for considering the test coverage. Pham [6] discussed a model that considers a logistic time-dependent fault-detection rate function. In addition, Pham [7] introduced a V-tub-shaped rate function. Pradhan et al. [8] presented an improved model for a generalized inflection S-shaped testing effort function. Erto et al. [9] also developed a generalized inflection S-shaped SRGM. Saxena et al. [10] proposed a SRGM that assumes imperfect debugging by a two-step process that considers fault observation and fault removal. Haque [11] presented a new Software Reliability Growth Model (SRGM) with the structure of a Logistic Growth Model in which the fault detection rate increases as the test department’s skill improves as the test progresses. Nafreen et al. [12] developed a SRGM with a bathtub-shaped fault detection rate function and proposed the conditional maximization algorithms to fit the models.
There are also SRGM studies that introduce different assumptions from previous studies. For example, the controlled test environment and actual operating environment of the consumers can vary significantly. Song et al. [13,14] discussed models that consider the uncertainty of the operating environments using new parameters as random variables. Zaitseva et al. [15] discussed a method to evaluate reliability analysis considering various types of uncertainty. Most SRGMs assume that failures are independent. However, a failure is occasionally dependent. Lee et al. [16] proposed a method for efficiently judging the software reliability using the model development, assuming a dependent failure and applying a sequential probability ratio test. Kim et al. [17] developed a new model that assumes the presence of dependent defects.
Recently, research on software reliability has been conducted with various approaches. Saxena et al. [18], Kumar et al. [19], and Garg et al. [20] developed criteria for judging goodness of fit for SRGMs. The criteria are calculated based on the combination of the entropy principle and various existing criteria. Most studies discussed the individual reliability of software and hardware. Yaghoobi [21] also developed two multi-criteria decision-making methods for comparison of SRGMs. Zhu [22] introduced complex reliability that considers both hardware and software, and proposed maintenance policies applicable to these systems. Recently, as the use of open-source software and cloud services increases, research on their reliability is being actively conducted [23,24,25]. Finally, several software reliability studies have recently been conducted using machine and deep learning [26,27,28,29].
In this study, we propose a new SRGM with two assumptions that software failures are dependent and the operating environments of the software are uncertain. The superiority of the proposed model is explained by the criteria for goodness of fit. The goodness of fit is used to determine which model fits the failure dataset for a given software product. A model with a good fit can predict the number of failures well and provide meaningful information for evaluating product reliability and establishing release policies. In Section 2, the concept of system reliability is provided. Section 3 describes the existing NHPP SRGM and the proposed model. Section 4 introduces the criteria and data used in the experiments, then discusses the results. Finally, Section 5 provides some concluding remarks.

2. Software Reliability Growth Model

2.1. Poisson Processes

The counting process is represented by N t , i.e., the total number of events up to time t . The counting process satisfies the following conditions [30]:
(1)
N t 0 .
(2)
N t is an integer.
(3)
If s t ,   N s N t .
(4)
When s t , N t N s represents the number of events within time interval s ,   t .
The homogeneous Poisson process (HPP) is given by
Pr N t = n = λ t n n ! e λ t , n = 0 ,   1 ,   2 ,  
where N t is the number of failures up to time t , and λ is a constant. Thus, the failure rate λ does not change over time t . The number of failures N t has a Poisson distribution with the mean λ t .
Most software reliability models assume that N t follows a non-homogeneous Poisson process (NHPP). When N t follows a/n NHPP, it is given by
Pr N t = n = m t n n ! e m t , n = 0 ,   1 ,   2 ,    
where m(t) is used instead of the constant λ , unlike with a HPP. This means that the failure rate m t depends on time t . Here, m(t) can be written as follows:
m t = E N t = 0 t λ s ds .

2.2. Reliability Function

Using m(t), the reliability function based on a NHPP can be expressed as follows [30]: The reliability function R t is defined as the probability that there will be no failures within the time interval 0 ,   t , and is given by
R t = P N t = 0 = e m t
Equation (4) indicates the probability that a software error will not occur within the interval [0,t]. If t + x is given, the software reliability can be expressed as a conditional probability R(x|t), as in Equation (5).
R x | t = P N t + x N t = 0 = e m t + x m t
Here, R(x|t) is the probability that a software error will not occur within the interval [t, t + x], where t ≥ 0 and x > 0. The density function of x is given by
f x = λ t + x e m t + x m t
where λ x = x m x .

3. NHPP Software Reliability Growth Model (SRGM)

3.1. Model Formulation

(1) In general, m t is obtained from a differential equation as follows [2,31]:
m t dt = b t a t m t
where a t is defined as the expected total number of initial failures plus the number of newly introduced errors that remain until testing, and b t is defined as the rate of detected failures per fault.
(2) The mean value function of the model when considering uncertainty of the operating environment can be obtained from the following differential equation [32]:
m t dt = η b t N m t
where b t is the fault detection rate function. Let η represent the uncertainty of the operating environment. Here, η is a random variable that follows gamma distribution g x with parameters α and β , which is given by
g x = β α x α 1 e β x Γ α .
(3) The mean value function of the model when considering dependent failures can be obtained from the following differential equation [16,17]:
m t dt = b t a t m t m t
where a t is defined as the expected total number of failures that remain prior to testing, and b t is the defined detection rate of failures per fault.

3.2. Proposed Model

Most models assume independent failures (see Section 3.3); however, this paper assumes dependent failures. For example, if an error occurs in a class in a certain program code, it may occur in chains in other classes that refer to this class. In this case, the failures are dependent, and the model proposed in this study also assumes the same type of failure.
We propose a model that considers an uncertain operating environment and dependent failures. We consider the second type owing to the fact that we need to assume an uncertain operating environment. When a t = N and b t = b 2 t 1 + bt are given, m t can be obtained from Equation (8) as follows:
m t = N 1 β α + b s ds α = N 1 β α + 0 t b 2 s 1 + bs ds α
= N 1 β α + bt log bt + 1 α
where b , α , β , and N are parameters of the proposed model. The parameters α and β in Equation (12) are the same as those in Equation (9). Therefore, parameters α and β of m t are dependent on the probability distribution of η , which indicates uncertain operating environments. This relationship leads to an assumption of dependent flaws in the model.

3.3. Existing NHPP Models

Table 1 summarizes the mean value functions for the existing NHPP models that are well known for their good performance and the proposed model. Dependent failure model 1 (DPF1) and dependent failure model 2 (DPF2) are models that assume dependent failures, whereas the other models assume independent failures. VTUB assumes the uncertain operating environments. In addition, the proposed model assumes two factors, the dependent failures and the uncertain operating environments. The parameters of each model can be estimated in various ways, including a Bayesian estimation, maximum likelihood estimation (MLE), and least square estimation (LSE). However, a Bayesian estimation is difficult to achieve owing to a lack of prior information, and a MLE estimation is occasionally difficult to apply because of the complexity of the mean value function of the model. Therefore, in this study, the parameters are estimated using the LSE method, which is discussed in Section 4.3.

4. Numerical Example

4.1. Datasets

We used two datasets to compare the goodness of fit of the model. Dataset 1, shown in Table 2, was collected from an IBM software package and has 40,000 lines of code [35]. The software was used to enter the data, and the time unit is in days. There is a total of 46 failures collected within the 21-day period.
Dataset 2 in Table 3 was collected from the ABC online communication system in 2000 [30]. The dataset was observed over a 12-week period, and the time unit is in weeks. A total of 55 failures were observed within the 12-week period.

4.2. Criteria

Various criteria have been proposed to discuss the goodness of fit of a model. Table 4 shows the criteria used to compare the goodness of fit for the SRGMs, which are listed in Table 1.
In Table 4, n is defined as the number of data, and m is defined as the number of parameters in the model (mean value function). The predicted value which means number of failures based on m t is presented as m ^ t i , and the actual data are presented as y i .
Most of the criteria mentioned measure the distance (or error) between predicted value and actual values. Therefore, it can be said that the shorter the distance, the better the mean value function of the model predicts the number of failures in the dataset.
The MSE measures the distance of predicted value by a model from the actual data with the consideration of the number parameters in the model and the number of data. The PRR measures the distance of predicted value from the actual data with the consideration the predicted value by a model. The PP measures the distance of predicted value from the actual data with the consideration the actual data. The SAE measures the sum of absolute error, which means the distance between the predicted value and the actual data.
The R 2 (coefficient of determination) is a measure of the fit of the regression equation. It represents the proportion of the regression sum of squares (SSR) to the total sum of squares (SST) of the model. The closer the value is to 0, the better the fit of the model.
The AIC measures a model’s ability to maximize a likelihood function ( L ). In general, the greater the number of parameters, the better the fit of the model. AIC mitigates the phenomenon by penalizing the number of parameters. In addition, log L of the AIC is given by
L = i = 1 n m ^ t i m ^ t i 1 y i y i 1 y i y i 1 !
log L = i = 1 n y i y i 1   log m ^ t i m ^ t i 1 m ^ t i m ^ t i 1 log y i y i 1 !
The BIC estimates the approximate value of the posterior probability of the model. In the second term of BIC, the number of parameters is penalized like AIC. Also, the term depends on the number of data.
The Bias measures the average distance between the predicted value and the actual data. The closer the value is to 0, the better the fit of the model. The Bias is also used to calculate PRV and RMSPE. The PRV calculates the standard deviation of the prediction bias, so the smaller the value, the better the fit of the model. The PRV is also called Variation or Variance. The RMSPE calculates the root mean square prediction error, which means the closeness of the predicted value with actual data using Bias and PRV. The MAE calculates the mean absolute error between the predicted value and the actual data. The MEOP calculates the sum of absolute error (SAE) with the consideration the predicted value by a model. The TS calculates the average deviation percentage over all data observation periods with regard to the actual data. The closer the value is to 0, the better the fit of the model.
To summarize the 13 criteria, the closer the value of R 2 is to 1, the closer the predicted value of the model is to the actual value. The closer the values of Bias and TS are to 0, the closer the predicted value of the model is to the actual value. For the other criteria—MSE, PRR, PP, SAE, AIC, BIC, PRV, RMSPE, MAE, and MEOP—the smaller the value, the closer the predicted value of the model is to the actual value, relative to other models on the same dataset.

4.3. Results

The estimated parameters of the models are listed in Table 5 and Table 6 using a least squares estimation (LSE).
Table 7 and Table 8 show the criteria obtained using the estimated parameters in Table 5 and Table 6, the best value for each criteria is shown in bold.
In Table 7, the proposed model shows the smallest MSE, the smallest PRR, the smallest PP, the smallest SAE, the largest R 2 , the closest-to-zero Bias, the smallest PRV, the smallest RMSPE, the smallest MAE, the smallest MEOP and the closest-to-zero TS at 1.3805, 0.0772, 0.0763, 15.9817, 0.9948, −0.0021, 1.0832, 1.0832, 0.9401, 0.8879 and 3.9233. Though the AIC and BIC are not the smallest values among all models, the proposed model displays the third= smallest AIC at 78.0915. In particular, the PRR value of DPF2 is the second-smallest value at 0.3333, but the values of other criteria are not good. The overall criteria values of VTUB are not good. When looking at the values of the overall criteria, it can be seen that the goodness of fit of the proposed model is excellent.
In Table 8, the proposed model shows the smallest MSE and the smallest AIC at 2.7776 and 57.2427, respectively. Though other criteria are not the best values among all models, the proposed model shows the second-largest R 2 , the second-smallest BIC, the second-closest-to-zero Bias, the second-smallest PRV, the second-smallest RMSPE, the second-closest-to-zero TS at 0.9920, 59.1823, −0.0070, 1.4213, 1.4213 and 3.6615, respectively. In particular, DPF1 shows the smallest PRR, the smallest PP, the smallest MAE and the smallest MEOP at 0.0276, 0.0270, 1.5924 and 1.4154, respectively. The VTUB shows the largest R 2 , the smallest SAE, the smallest PRV, the smallest RMSPE and the closest-to-zero TS at 0.9925, 12.6030, 1.3688, 1.3704 and 3.5306, respectively. When looking at the values of the overall criteria, it can be seen that the goodness of fit of the proposed model is excellent.
What is especially important about this result is that the fit of the proposed model is much better than DPF1, DPF2, and VTUB. In other words, DPF1 and DPF2 assume dependent failures, and VTUB assumes uncertain operating environments. It can be seen that the proposed model takes both assumptions into account and fits the dataset much better. Therefore, it is reasonable to predict the number of failures and conduct software reliability studies (reliability evaluation, release policy, etc.) based on the proposed model rather than DPF1, DPF2, or VTUB.
Among the 13 criteria in this paper, the most classic and well used are the MSE, PRR, and PP. Figure 1 and Figure 2 display the three criteria values of the top three models for each dataset. For the criteria—MSE, PRR, PP, AIC, and RMSPE—the smaller the value, the closer the predicted value of the model is to the actual value, relative to other models on the same dataset.
For dataset 1 (See Figure 1), the proposed model displays the smallest MSE, the smallest PRR and the smallest PP at 1.3805, 0.0772, and 0.0763, respectively. For dataset 2 (see Figure 2), the proposed model displays the smallest MSE and the smallest AIC at 2.7776 and 57.2427, respectively. The VTUB displays the smallest RMSPE at 1.3704.
Figure 3 and Figure 4 present the 95% confidence intervals of the datasets, the formula of which is given by
m ^ t ± z α m ^ t
where z α is defined as the 100 1 α / 2 percentile of the standard normal distribution [30].
Figure 5 and Figure 6 show the mean value functions of the existing NHPP models and the proposed model for datasets.

5. Conclusions

Most NHPP SRGMs have been developed with stand-alone failure considerations or are environment-based (controlled). Failures can occasionally occur, and the test and operating environments used by real consumers differ. As the main idea of this study, the proposed model considers both dependent failures and uncertain operating environments. When looking at the values of the overall criteria in Table 7 and Table 8, it can be seen that the goodness of fit of the proposed model is excellent. Among the 13 criteria, 11 criteria values show the best values for dataset 1. For dataset 2, 2 criteria values show the best values and 6 criteria values show the second-best values. In other words, the proposed model more closely predicts the number of failures than other models. Therefore, we conclude that it is reasonable to predict the number of failures based on the proposed model and conduct software reliability studies (reliability assessment, release policy, etc.).

6. Future Research

In the future, it is believed that various SRGMs can be developed if a t and b t are based on various assumptions, a dependent failure, and uncertain operating environments are taken into consideration. We will develop a model that expresses more robust dependent failures by extending the model proposed in this study. Also, we plan to convert software reliability evaluation and cost function into a single-objective function [44,45]. In addition, it can be extended to various studies [46,47,48,49]. This study can be extended into a release policy study. As the frequency of use of cloud services and open-source software is increasing, it is necessary to study the reliability of collecting datasets on them.

Author Contributions

Conceptualization, H.P.; funding acquisition, I.H.C.; software, D.H.L.; writing—original draft, D.H.L.; writing—review and editing, I.H.C. and H.P. These three authors contributed equally to the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF), funded by the Ministry of Education (NRF-2021R1F1A1048592 and 2021R1A6A3A01086716).

Institutional Review Board Statement

Not required.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository.

Acknowledgments

Many thanks to the reviewers for careful reading and valuable comments which improved the paper representation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goel, A.L.; Okumoto, K. Time-dependent error-detection rate model for software reliability and other performance measures. IEEE Trans. Reliab. 1979, 28, 206–211. [Google Scholar] [CrossRef]
  2. Yamada, S.; Ohba, M.; Osaki, S. S-shaped software reliability growth models and their applications. IEEE Trans. Reliab. 1984, 33, 289–292. [Google Scholar] [CrossRef]
  3. Pham, H.; Zhang, X. An NHPP software reliability model and its comparison. Int. J. Reliab. Qual. Saf. Eng. 1997, 4, 269–282. [Google Scholar] [CrossRef]
  4. Pham, H.; Zhang, X. NHPP software reliability and cost models with testing coverage. Eur. J. Oper. Res. 2003, 145, 443–454. [Google Scholar] [CrossRef]
  5. Song, K.Y.; Chang, I.H.; Pham, H. A testing coverage model based on NHPP software reliability considering the software operating environment and the sensitivity analysis. Mathematics 2019, 7, 450. [Google Scholar] [CrossRef] [Green Version]
  6. Pham, H. A logistic fault-dependent detection software reliability model. J. Univers. Comput. Sci. 2018, 24, 1717–1730. [Google Scholar]
  7. Pham, H. Distribution function and its applications in software reliability. Int. J. Performability Eng. 2019, 15, 1306–1313. [Google Scholar] [CrossRef]
  8. Pradhan, V.; Kumar, A.; Dhar, J. Enhanced growth model of software reliability with generalized inflection S-shaped testing-effort function. J. Interdiscip. Math. 2022, 25, 137–153. [Google Scholar] [CrossRef]
  9. Erto, P.; Giorgio, M.; Lepore, A. The Generalized Inflection S-Shaped Software Reliability Growth Model. IEEE Trans. Reliab. 2020, 69, 228–244. [Google Scholar] [CrossRef]
  10. Saxena, P.; Ram, M. Two phase software reliability growth model in the presence of imperfect debugging and error generation under fuzzy paradigm. Math. Eng. Sci. Aerosp. (MESA) 2022, 13, 777–790. [Google Scholar]
  11. Haque, M.A.; Ahmad, N. An effective software reliability growth model. Saf. Reliab. 2021, 40, 209–220. [Google Scholar] [CrossRef]
  12. Nafreen, M.; Fiondella, L. Software Reliability Models with Bathtub-shaped Fault Detection. In Proceedings of the 2021 Annual Reliability and Maintainability Symposium (RAMS), Orlando, FL, USA, 24–27 May 2021; pp. 1–7. [Google Scholar]
  13. Song, K.Y.; Chang, I.H.; Pham, H. A three-parameter fault-detection software reliability model with the uncertainty of operating environments. J. Syst. Sci. Syst. Eng. 2017, 26, 121–132. [Google Scholar] [CrossRef]
  14. Song, K.Y.; Chang, I.H.; Pham, H. NHPP software reliability model with inflection factor of the fault detection rate considering the uncertainty of software operating environments and predictive analysis malignant. Symmetry 2019, 11, 521. [Google Scholar] [CrossRef] [Green Version]
  15. Zaitseva, E.; Levashenko, V. Construction of a Reliability Structure Function Based on Uncertain Data. IEEE Trans. Reliab. 2016, 65, 1710–1723. [Google Scholar] [CrossRef]
  16. Lee, D.H.; Chang, I.H.; Pham, H. Software reliability model with dependent failures and SPRT. Mathematics 2020, 8, 1366. [Google Scholar] [CrossRef]
  17. Kim, Y.S.; Song, K.Y.; Pham, H.; Chang, I.H. A software reliability model with dependent failure and optimal release time. Symmetry 2022, 14, 343. [Google Scholar] [CrossRef]
  18. Saxena, P.; Kumar, V.; Ram, M. A novel CRITIC-TOPSIS approach for optimal selection of software reliability growth model (SRGM). Qual. Reliab. Eng. Int. 2022, 38, 2501–2520. [Google Scholar] [CrossRef]
  19. Kumar, V.; Saxena, P.; Garg, H. Selection of optimal software reliability growth models using an integrated entropy–Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS) approach. Math. Methods Appl. Sci. 2021. [Google Scholar] [CrossRef]
  20. Garg, R.; Raheja, S.; Garg, R.K. Decision Support System for Optimal Selection of Software Reliability Growth Models Using a Hybrid Approach. IEEE Trans. Reliab. 2022, 71, 149–161. [Google Scholar] [CrossRef]
  21. Yaghoobi, T. Selection of optimal software reliability growth model using a diversity index. Soft Comput. 2021, 25, 5339–5353. [Google Scholar] [CrossRef]
  22. Zhu, M. A new framework of complex system reliability with imperfect maintenance policy. Ann. Oper. Res. 2022, 312, 553–579. [Google Scholar] [CrossRef]
  23. Wang, L. Architecture-Based Reliability-Sensitive Criticality Measure for Fault-Tolerance Cloud Applications. IEEE Trans. Parallel Distrib. Syst. 2019, 30, 2408–2421. [Google Scholar] [CrossRef]
  24. Wang, H.; Wang, L.; Yu, Q.; Zheng, Z. Learning the Evolution Regularities for BigService-Oriented Online Reliability Prediction. IEEE Trans. Serv. Comput. 2019, 12, 398–411. [Google Scholar] [CrossRef]
  25. Wang, L.; He, Q.; Gao, D.; Wan, J.; Zhang, Y. Temporal-Perturbation Aware Reliability Sensitivity Measurement for Adaptive Cloud Service Selection. IEEE Trans. Serv. Comput. 2022, 15, 2301–2313. [Google Scholar] [CrossRef]
  26. Wang, J.; Zhang, C. Software reliability prediction using a deep learning model based on the RNN encoder–decoder. Reliab. Eng. Syst. Saf. 2018, 170, 73–82. [Google Scholar] [CrossRef]
  27. San, K.K.; Washizaki, H.; Fukazawa, Y.; Honda, K.; Taga, M.; Matsuzaki, A. Deep Cross-Project Software Reliability Growth Model Using Project Similarity-Based Clustering. Mathematics 2021, 9, 2945. [Google Scholar] [CrossRef]
  28. Li, L. Software reliability growth fault correction model based on machine learning and neural network algorithm. Microprocess. Microsyst. 2021, 80, 103538. [Google Scholar] [CrossRef]
  29. Banga, M.; Bansal, A.; Singh, A. Implementation of machine learning techniques in software reliability: A framework. In Proceedings of the 2019 International Conference on Automation, Computational and Technology Management (ICACTM), London, UK, 24–26 April 2019; pp. 241–245. [Google Scholar] [CrossRef]
  30. Pham, H. System Software Reliability; Springer: London, UK, 2006. [Google Scholar]
  31. Pham, H.; Nordmann, L.; Zhang, Z. A general imperfect-software-debugging model with S-shaped fault-detection rate. IEEE Trans. Reliab. 1999, 48, 169–175. [Google Scholar] [CrossRef]
  32. Pham, H. A new software reliability model with Vtub-shaped fault-detection rate and the uncertainty of operating environments. Optimization 2014, 63, 1481–1490. [Google Scholar] [CrossRef]
  33. Yamada, S.; Tokuno, K.; Osaki, S. Imperfect debugging models with fault introduction rate for software reliability assessment. Int. J. Syst. Sci. 1992, 23, 2241–2252. [Google Scholar] [CrossRef]
  34. Chang, I.H.; Pham, H.; Lee, S.W.; Song, K.Y. A testing-coverage software reliability model with the uncertainty of operation environments. Int. J. Syst. Sci. Oper. Logist. 2014, 1, 220–227. [Google Scholar] [CrossRef]
  35. Ohba, M. Software reliability analysis models. IBM J. Res. Dev. 1984, 28, 428–443. [Google Scholar] [CrossRef]
  36. Song, K.Y.; Chang, I.H.; Pham, H. A software reliability model with a Weibull fault detection rate function subject to operating environments. Appl. Sci. 2017, 7, 983. [Google Scholar] [CrossRef] [Green Version]
  37. Li, Q.; Pham, H. A testing-coverage software reliability model considering fault removal efficiency and error generation. PLoS ONE 2017, 12, e0181524. [Google Scholar] [CrossRef] [Green Version]
  38. Akaike, H. A new look at the statistical model identification. IEEE Trans. Automat. Contr. 1974, 19, 716–723. [Google Scholar] [CrossRef]
  39. Okamura, H.; Dohi, T.; Osaki, S. Software reliability growth models with normal failure time distributions. Reliab. Eng. Syst. Saf. 2013, 116, 135–141. [Google Scholar] [CrossRef]
  40. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  41. Zhu, M.; Pham, H. A two-phase software reliability modeling involving with software fault dependency and imperfect fault removal. Comput. Lang. Syst. Struct. 2018, 53, 27–42. [Google Scholar] [CrossRef]
  42. Pillai, K.; Sukumaran Nair, V.S. A model for software development effort and cost estimation. IEEE Trans. Softw. Eng. 1997, 23, 485–497. [Google Scholar] [CrossRef]
  43. Anjum, M.; Haque, M.A.; Ahmad, N. Analysis and ranking of software reliability models based on weighted criteria value. Int. J. Inf. Technol. Comput. Sci. 2013, 5, 1–14. [Google Scholar] [CrossRef] [Green Version]
  44. Gharaei, A.; Amjadian, A.; Shavandi, A. An integrated reliable four-level supply chain with multi-stage products under shortage and stochastic constraints. Int. J. Syst. Sci. Oper. Logist. 2021, 1–22. [Google Scholar] [CrossRef]
  45. Askari, R.; Sebt, M.V.; Amjadian, A. A Multi-product EPQ Model for Defective Production and Inspection with Single Machine, and Operational Constraints: Stochastic Programming Approach. In International Conference on Logistics and Supply Chain Management; Springer: Cham, Switzerland, 2021; pp. 161–193. [Google Scholar]
  46. Souza, R.L.C.; Ghasemi, A.; Saif, A.; Gharaei, A. Robust job-shop scheduling under deterministic and stochastic unavailability constraints due to preventive and corrective maintenance. Comput. Ind. Eng. 2022, 168, 108130. [Google Scholar] [CrossRef]
  47. Amjadian, A.; Gharaei, A. An integrated reliable five-level closed-loop supply chain with multi-stage products under quality control and green policies: Generalised outer approximation with exact penalty. Int. J. Syst. Sci. Oper. Logist. 2022, 9, 429–449. [Google Scholar] [CrossRef]
  48. Gharaei, A.; Almehdawe, E. Optimal sustainable order quantities for growing items. J. Clean. Prod. 2021, 307, 127216. [Google Scholar] [CrossRef]
  49. Gharaei, A.; Diallo, C.; Venkatadri, U. Optimal economic growing quantity for reproductive farmed animals under profitable by-products and carbon emission considerations. J. Clean. Prod. 2022, 374, 133849. [Google Scholar] [CrossRef]
Figure 1. Three criteria values of top three models for dataset 1. (a) MSE; (b) PRR; (c) PP.
Figure 1. Three criteria values of top three models for dataset 1. (a) MSE; (b) PRR; (c) PP.
Applsci 12 12383 g001
Figure 2. Three criteria values of top three models for dataset 2. (a) MSE; (b) AIC; (c) RMSPE.
Figure 2. Three criteria values of top three models for dataset 2. (a) MSE; (b) AIC; (c) RMSPE.
Applsci 12 12383 g002
Figure 3. Confidence intervals (95%) of dataset 1.
Figure 3. Confidence intervals (95%) of dataset 1.
Applsci 12 12383 g003
Figure 4. Confidence intervals (95%) of dataset 2.
Figure 4. Confidence intervals (95%) of dataset 2.
Applsci 12 12383 g004
Figure 5. Mean value function of the models for dataset 1.
Figure 5. Mean value function of the models for dataset 1.
Applsci 12 12383 g005
Figure 6. Mean value functions of the models for dataset 2.
Figure 6. Mean value functions of the models for dataset 2.
Applsci 12 12383 g006
Table 1. Mean value functions for existing NHPP models and the proposed approach.
Table 1. Mean value functions for existing NHPP models and the proposed approach.
No.Model m t
1DPF1 [16] a 1 + a h b + c c + b exp bt a b
2DPF2 [17] a 1 + a h 1 + c c + exp bt a
3DS [33] a 1 1 + bt exp bt
4GO [1] a 1 exp bt
5IS [2] a 1 exp bt 1 + β exp bt
6YID [33] a 1 exp bt 1 α b + α at
7PNZ [31] a 1 exp bt 1 α b + α at 1 + β exp bt
8PZ [3] c + a 1 exp bt ab b α exp at exp bt 1 + β exp bt
9TC [34] N 1 β β + at b α
10VTUB [32] N 1 β β + a t b 1 α
11NEW N 1 β α + bt log bt + 1 α
Table 2. Dataset 1.
Table 2. Dataset 1.
TimeCumulative
Failures
TimeCumulative
Failures
TimeCumulative
Failures
128121531
239191637
3410211738
4511221841
5712241942
6913262045
71114302146
Table 3. Dataset 2.
Table 3. Dataset 2.
TimeFailuresCumulative
Failures
TimeFailuresCumulative
Failures
110107440
22128343
34169144
462210650
562811151
683612455
Table 4. Criteria.
Table 4. Criteria.
NoCriteria
1Mean square error (MSE) [30]
i = 1 n m ^ t i y i 2 n m
2Predictive ratio risk (PRR) [30]
i = 1 n m ^ t i y i 2 m ^ t i
3Predictive power (PP) [30]
i = 1 n m ^ t i y i 2 y i
4Sum of absolute error (SAE) [36]
i = 1 n m ^ t i y i
5R-square ( R 2 ) [37]
1 i = 1 n m ^ t i y i 2 i = 1 n y i y ¯ i 2
6Akaike’s information criterion (AIC) [38]
2 log L + 2 m
7Bayesian information criterion (BIC) [39,40]
2 log L + m log n
8Bias [41]
i = 1 n m ^ t i y i n
9Predicted relative variation (PRV) [42]
i = 1 n m ^ t i y i Bias 2 n 1
10Root mean square prediction error (RMSPE) [42]
Bias 2 + PRV 2
11Mean absolute error (MAE) [43]
i = 1 n m ^ t i y i n m
12Mean error of prediction (MEOP) [43]
i = 1 n m ^ t i y i n m + 1
13Theil statistic (TS) [43]
100 i = 1 n m ^ t i y i 2 i = 1 n y i 2 %
Table 5. Estimated parameters of models for dataset 1.
Table 5. Estimated parameters of models for dataset 1.
Model a ^ b ^ α ^ β ^ N ^ c ^ h ^
DPF151.3500.001---0.2162.659
DPF251.3500.005---0.0762.629
DS77.2530.0966-----
GO14,264.0000.0002-----
IS58.9430.170-8.386---
YID1.4910.30681.7457----
PNZ29.8750.1920.0454.900---
PZ59.3160.1682128.10298.2581-0.0005-
TC0.01911.567839.1540221.173578.7859--
VTUB1.97010.68920.292819.852987.2519--
NEW-0.24702.3551.968126.140--
Table 6. Estimated parameters of models for dataset 2.
Table 6. Estimated parameters of models for dataset 2.
Model a ^ b ^ α ^ β ^ N ^ c ^ h ^
DPF155.8930.004---0.5487.274
DPF256.0580.008---0.0937.195
DS57.4780.344-----
GO94.3440.0733-----
IS65.7810.206-1.293---
YID5.74952.4150.756----
PNZ64.9220.2080.0011.286---
PZ7.6170.2100.0051.321-64.992-
TC0.0051.0752001.00084.68180.373--
VTUB5.06931.7930.01810.000457.6685--
NEW-0.3161.3261.14291.500--
Table 7. Criteria values for model comparison of dataset 1.
Table 7. Criteria values for model comparison of dataset 1.
ModelMSEPRRPPSAER2AICBICBiasPRVRMSPEMAEMEOPTS (%)
DPF12.01590.34340.602620.55850.992478.823183.00120.10651.30441.30881.20931.14214.7409
DPF22.00550.33330.575320.46060.992478.799982.97800.10061.30161.30541.20361.13674.7287
DS1.636526.32291.208121.03490.993178.118480.2075−0.23261.22391.24581.10711.05174.5159
GO6.60100.80801.860242.54360.972177.335179.42420.89132.33172.49622.23912.12729.0696
IS1.39520.69910.300317.44950.994476.692579.8261−0.03251.12011.12050.96940.91844.0584
YID1.70083.06560.608121.06060.993278.662481.7960−0.05481.23601.23721.17001.10854.4810
PNZ1.48440.95780.352117.90740.994478.941983.1200−0.05041.12211.12321.05340.99494.0683
PZ1.56970.69470.300317.52030.994480.711685.9342−0.02721.12031.12061.09501.03064.0586
TC1.72685.95900.758019.43420.993982.566087.7886−0.05481.17401.17531.21461.14324.2568
VTUB1.54380.56100.269917.51000.994580.601985.8245−0.04251.11051.11131.09441.03004.0250
NEW1.38050.07720.076315.98170.994878.091582.2696−0.00211.08321.08320.94010.88793.9233
Table 8. Criteria values for model comparison of dataset 2.
Table 8. Criteria values for model comparison of dataset 2.
ModelMSEPRRPPSAER2AICBICBiasPRVRMSPEMAEMEOPTS (%)
DPF12.82010.02760.027012.73890.991958.695860.63540.01031.43211.43211.59241.41543.6893
DPF22.79460.02830.027512.75150.991958.627460.56700.00781.42561.42561.59391.41683.6726
DS8.20967.36790.617720.95400.970469.625170.5950−0.70592.63052.72362.09541.90497.0377
GO4.02450.29320.162719.41700.985557.707658.6775−0.05221.91201.91271.94171.76524.9275
IS4.05550.48150.190517.05200.986860.145161.5998−0.17251.81261.82081.89471.70524.6926
YID7.75360.08930.102724.40960.974858.259359.71400.00002.51872.51872.71222.44106.4885
PNZ4.56320.48180.190617.05660.986862.138964.0786−0.17221.81281.82102.13211.89524.693
PZ5.21530.48900.191717.04590.986864.168966.5934−0.17581.81251.82102.43512.13074.6931
TC5.64200.43070.188818.37230.985764.251966.6765−0.12031.89061.89452.62462.29654.8813
VTUB2.95160.03200.029612.60300.992559.551461.9760−0.06641.36881.37041.80041.57543.5306
NEW2.77760.06060.051414.49140.992057.242759.1823−0.00701.42131.42131.81141.61023.6615
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, D.H.; Chang, I.H.; Pham, H. Software Reliability Growth Model with Dependent Failures and Uncertain Operating Environments. Appl. Sci. 2022, 12, 12383. https://doi.org/10.3390/app122312383

AMA Style

Lee DH, Chang IH, Pham H. Software Reliability Growth Model with Dependent Failures and Uncertain Operating Environments. Applied Sciences. 2022; 12(23):12383. https://doi.org/10.3390/app122312383

Chicago/Turabian Style

Lee, Da Hye, In Hong Chang, and Hoang Pham. 2022. "Software Reliability Growth Model with Dependent Failures and Uncertain Operating Environments" Applied Sciences 12, no. 23: 12383. https://doi.org/10.3390/app122312383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop