Next Article in Journal
An Efficient End-to-End Obstacle Avoidance Path Planning Algorithm for Intelligent Vehicles Based on Improved Whale Optimization Algorithm
Next Article in Special Issue
Causality-Driven Efficient Feature Selection for Deep-Learning-Based Surface Roughness Prediction in Milling Machines
Previous Article in Journal
Estimates for the Coefficients of Subclasses Defined by the Bell Distribution of Bi-Univalent Functions Subordinate to Gegenbauer Polynomials
Previous Article in Special Issue
Global Value Chains and Industry 4.0 in the Context of Lean Workplaces for Enhancing Company Performance and Its Comprehension via the Digital Readiness and Expertise of Workforce in the V4 Nations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AdaBoost Algorithm Could Lead to Weak Results for Data with Certain Characteristics

by
Olivér Hornyák
1 and
László Barna Iantovics
2,*
1
Institute of Information Engineering, University of Miskolc, 3515 Miskolc, Hungary
2
Department of Electrical Engineering and Information Technology, George Emil Palade University of Medicine, Pharmacy, Science and Technology of Targu Mures, 540142 Targu Mures, Romania
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(8), 1801; https://doi.org/10.3390/math11081801
Submission received: 12 March 2023 / Revised: 4 April 2023 / Accepted: 7 April 2023 / Published: 10 April 2023
(This article belongs to the Special Issue Industrial Big Data and Process Modelling for Smart Manufacturing)

Abstract

:
There are many state-of-the-art algorithms presented in the literature that perform very well on some evaluation data but are not studied with the data properties on which they are applied; therefore, they could have low performance on data with other characteristics. In this paper, the results of comprehensive research regarding the prediction with the frequently applied AdaBoost algorithm on real-world sensor data are presented. The chosen dataset has some specific characteristics, and it contains error and failure data of several machines and their components. The research aims to investigate whether the AdaBoost algorithm has the capability of predicting failures, thus providing the necessary information for monitoring and condition-based maintenance (CBM). The dataset is analyzed, and the principal characteristics are presented. Performance evaluations of the AdaBoost algorithm that we present show a prediction capability below expectations for this algorithm. The specificity of this study is that it indicates the limitation of the AdaBoost algorithm, which could perform very well on some data, but not so well on others. Based on this research and some others that we performed, and actual research from worldwide studies, we must outline that the mathematical analysis of the data is especially important to develop or adapt algorithms to be very efficient.

1. Introduction

Modern buildings integrate increasingly smarter applications; thus, a large amount of real-time sensor data can be collected and transmitted by the Internet of Things devices [1]. Oliveira et al. [2] presented evaluation procedures for performing forecasting based on complex spatiotemporal data. Leon and Gavrilescu [3] presented a survey on the problem of tracking and trajectory prediction methods that should be applied in autonomous driving. Haq et al. [4] studied an adapted consumption prediction model that can be applied in commercial and residential sectors.
Typical applications aim to conserve energy consumption: heating, lighting, air conditioning, ventilation, etc. Other applications belong to routine facility management such as electricity, water supply, sanitary, etc. There is immense potential to apply artificial intelligence and machine learning during the life cycle of buildings [5].
Measuring machine intelligence makes possible the comparison of the systems and applications based on their intelligence. Because the diversity of the systems is large, it is difficult to elaborate on universal intelligence metrics. The paper [6] presented the mathematical modeling of a universal black-box-based intelligence metric called MetrIntPairII, which is able to measure and compare a set of systems based on their intelligence, finally classifying the systems into intelligence classes. In [7], a novel universal method called ExtrIntDetect for the identification of intelligent systems with extremely low and extremely high intelligence was proposed. This allowed choosing the system that has statistically much lower or higher intelligence than others to solve problems. MetrIntPairII and ExtrIntDetect could help intelligent systems developers and at the same time users of intelligent systems in choosing systems based on problem-solving intelligence.
This paper focuses on the monitoring and condition-based maintenance (CBM) of construction facility equipment that is essential for providing flawless operation. CBM, which is defined as “preventive maintenance which includes assessment of physical conditions, analysis and the possible ensuing maintenance actions” [8] plays an important role in lifecycle engineering to improve machines’ availability and reduce maintenance costs. State-of-the-art CBM applications have been improved recently; they use data analytics processes on the data of the system being investigated. Very often, data analytics [9] use Artificial Intelligence techniques that can evaluate whether the machine operations follow a normal condition pattern or whether there is an anomaly [10]. Machine designers do not have experience with machine operational failures; the machines operate in different environments. Thus, the maintenance activities become increasingly significant [11]. A closely related term is prognostics and health management (PHM) which aims for the prediction of its reliability and the remaining useful lifetime (RUL) of machines and machines health [12].
Internet of Things (IoT) technology focuses on obtaining sensor information [13,14,15], then securely processing and storing it. Cloud-based data centers are widely used. The benefits of Big Data techniques [16] are also utilized in this process. Some research suggests using blockchain technology [17].
The monitoring of electricity and thermal energy systems, ventilation, and cooling/heating systems can be traced back to measuring electrical parameters. The following signals are typically measured: power consumption, and mechanical vibration. For example, increasing power consumption or higher vibration may indicate an upcoming failure. They are usually analyzed using methods such as [18] a fast Fourier transform and power spectrum analysis.
Collecting and monitoring these data can form a base to evaluate the condition of each piece of equipment to explore the characteristic and life curve state of a particular piece of equipment. The remaining useful life is the length of usable time left on an asset at a specific time [19]. The definition of what is ‘usable’ for the owner of the equipment may be specific to the asset.
The residual life is constantly decreasing during the operation of equipment. Properly scheduled maintenance and repairs can extend the lifetime. The advantage of condition-based maintenance is that the maintenance events of the equipment need to happen only when its status characteristics justify this. The investigation of the dynamics of physical signal characteristics and performance measurements may indicate the failure of parts of the equipment so that the maintenance time and cost can be minimized with the help of the accurate prediction of forthcoming failures. Another advantage is that purchasing and storing those parts can be more efficiently scheduled.
CBM may be based on offline and online measurements. In the offline case, the measurement and data acquisition are performed periodically, followed by recurring data processing. This is a cost-effective way for monitoring equipment whose lifetime degradation is known and nearly constant.
The other type is as follows: online, real-time measurement and monitoring is feasible to use in cases where the lifetime degradation is unknown or very uncertain. Furthermore, this strategy is valuable when the loss of the equipment is great, there is a potential personal injury, or the environmental damage is greater than the cost of real-time process monitoring.
Figure 1 shows the most typical CBM techniques presented by [20]. There are three major families of techniques: data processing, diagnostics, and prognostics. Data processing focuses on handling and analyzing data or signals and transforming them into an interpretable form. Diagnostics are responsible for fault detection, when and what kind of fault happened, and which components were affected. Prognostics attempt to predict upcoming faults.
Another comprehensive review of the available CBM models was given in [21]. The paper treats two distinct strategies: time-based maintenance and condition-based maintenance. CBM techniques are divided into prognostic methods, and machine learning approaches are also discussed such as artificial neural networks, fuzzy logic, and expert systems. Common data-driven methods include the principal component analysis, learning vector quantization, and hidden Markov models.
The structure of the rest of the paper is as follows: Section 2 describes the research work, and then a review is given of the scientific literature. Section 3 describes the characteristics of the dataset. Then, the AdaBoost algorithm is introduced to predict failures using the given dataset. Finally, in Section 4 the conclusions are formulated.

2. Materials and Methods

In this paper, a data-driven method is studied. The research hypothesis is that by investigating a database that contains records of machine errors and failures along with further telemetry data of those machines, this will allow the machine errors to be predicted with some expected probability. The research project focused on investigating a boosting algorithm, which will be discussed later. In our methodology, the dataset was divided into two parts. Within the available time range, the time of the investigation split the dataset into historical and future data. By using the historical data, predictions were made for the ‘future’, i.e., to the time after that point. The predictions were then compared with the real facts. The research workflow is depicted in Figure 2.

2.1. Estimations of Machine Failures

We considered a constraint when an estimation should be given on the expected time of a machine failure. The following sections give an overview of the estimation methods that have been investigated in this paper.

2.1.1. Estimation by Mean Operation Time

Let us assume that the data of some similar machines operating in a similar environment are available for investigation. It can be noticed that the machine failure will be close to the mean life (unless the usage of a given machine is different from the other machines). Let us have n machines. The start time of the machine is denoted by si. Let us denote the fi as the failure time of the ith machine. The operation time of the ith machine can be calculated as:
t i = f i s i
The mean operation time is:
t = i = 1 n t i n
Let us assume the kth machine starts at sk. The average time of the expected failure time can be calculated by (3):
s k + t
The precision of the estimation can be described by the distance between the expected and real failure time. The absolute value of the distance or the square of the distance is an appropriate indicator. It is expected that the more machines we have, the more precise the calculation of the expected value. The precision of the estimation can be improved incrementally. Let us assume there is an appropriate number of data samples to calculate the average operation time. When a new failure is noted, the value of the mean operation time can be updated by (2).

2.1.2. Estimation Based on the Relative Frequency

When the estimation targets a fixed time interval, we can count the occurrences of the failures of the previous periods, and the relative frequency can be considered as the probability of occurrence. This is a calculation that is not computationally expensive and provides an approximate value for future occurrences without considering the distribution of the failures. An advantage of the method is that adding new samples to the dataset will make the estimation more precise. The disadvantage is that the entire time interval must be considered. Properly selected aggregation methods can eliminate the linear growth of computing time. This method cannot ’forget’, i.e., all the data samples are taken into account. We may want the model not to consider failure data too old, or ’refresh’ memory after a repair of the machines. Applying a sliding time window can improve this method. Another characteristic of the model is that certain time windows contain no failure data at all; therefore, the predicted probability of failure is zero.

2.1.3. Estimation by the Expected Value

Since time is a continuous value, the likelihood is 0 that the failure will occur exactly at the estimated time. For this reason, it is expedient to estimate for a period. Let us utilize Markov’s inequality [22], which looks as follows:
P ( ξ > δ · E ξ ) 1 δ )
where ξ is a probability variable and δR. In our case, ξ refers to the lifetime of the equipment. The expected value E(ξ) can be predicted by the aforementioned t value. In the case of a homogeneous sampling, when n→∞ the value of t will approach the expected value. If there is an estimation for the expected value, any arbitrary δ > 0 value can be used to estimate whether a specific machine would fail within a specified period. Let us denote this period hR. For the estimation, let us use (5):
1 P ( ξ δ · E ξ ) 1 δ )
from which
1 P ( ξ h ) 1 δ )
as
δ = h E ( ξ )
Rearranging the equation results:
P ξ h 1 E ξ h 1 t h
Note that h E ( ξ ) must be true.
It is important to note that this estimation does not take into account how long the machine has been working. The estimated probability can be calculated when the machine is started. From a practical point of view, it is more advantageous to give an estimate for a specified interval such as one day, one week, or one month forward. A threshold for the estimated probability of the failure occurring in the interval can be calculated based on a metric. Such a metric can be the proportion of correct and incorrect failures. In order to calculate this, a known period must be investigated and the expected and obtained results should be compared. Frequencies and relative frequencies can be used. The advantage of using the frequency is that it contains the size of the samples as meta information. The use of relative frequencies makes the result easier to compare with other calculations. To evaluate the method by investigating the number of failures, we can use:
  • The absolute value of the difference between the observed and calculated results;
  • The square of the difference between the observed and calculated results;
  • And p-norm as a generic case. Let be the vector of differences, and pR, p ≥ 1. p-norm [23] is defined as:
    p = i = 1 n Δ i p 1 p
    where n is the number of discrete times considered for the estimation.
If the question is only whether the machine operates with or without failure during the period, then the problem can be considered as a classification problem. The simplest case is binary classification: to have one class for the failures and one for no failure. If the failure count is an important factor, then additional classes can be introduced to represent the number of failures.

2.1.4. Estimation by the Expected Value

With Markov inequality [22], we can only provide a very rough probability of failures. Let us have a ξ probability value, whose expected value is E and its standard deviation is D. For an arbitrary εR, ε > 0, the following is true:
P ( ξ E ( ξ ) ε ) D 2 ( ξ ) ε 2
Therefore, the inequality (10) shows the probability of the values that are likely to fall into the ε radius of the expected value. In practice, this can be used as follows:
  • Based on the available samples, give an estimation for the expected average value and the standard deviation;
  • Specify an ε value that will select a time interval of 2ε length on the time axis, with the expected value in the center;
  • Give an estimation of the probability of an item being inside/outside of the interval;
  • Often, the result can be interpreted as giving the probability of the failure to fall into the examined interval. With the rearrangement of the equation, this can be calculated as:
P ξ E ξ < ε 1 D 2 ( ξ ) ε 2
if
1 D 2 ξ ε 2 > 0 1 > D 2 ξ ε 2 ε 2 > D 2 ξ ε > D ( ξ )
It can also be noticed that the increase in the value of ε, the size of the period, and the probability of a failure to fall into the period will increase.
It is possible to use inequality in another form. If the probability p is known and the size of the period is the question, then the inequality can be rearranged as follows:
p = 1 D 2 ( ξ ) ε 2
To express ε :
ε = D 2 ( ξ ) 1 p
The higher the probability p, the higher ε will be. Note that the equalities refer to known theoretically expected values and standard deviations. In most cases, however, they are not available for the equipment.

2.2. Analysis of the Dataset

In this section, the dataset used is described. The target of the investigations made in this paper was acquired with a real dataset (see data availability statement at the end of the paper), which contained data from 100 machines, see Table 1. A brief overview of the information gained about the machine’s lifetime is as follows:
  • The lifetime of the machines is between 0 and 20 years;
  • No typical distribution is identified for the age of the machines;
  • The average lifetime of the machines is 11.33 years;
  • The age of the machines is approximately 5.8 years;
  • No 13-year-old machines are in the set—all other ages in the range of [0, 20] can be seen; see Figure 3;
  • The most frequent is the 14-year-old machines with the count 14. The second most common age is 10 years, with 10 instances.
Some remarks on model types:
  • The distribution of the number of similar machines is not uniform. The individual model counts are 16, 17, 35, and 32;
  • The most common machine of type #3 is 14 years old having 6 instances;
  • The average age of type #4 is outstanding: the average ages of other types are 12.25, 12.76, and 12.03, respectively, while type #4 is about 9.34 years old on average;
  • The highest standard deviation regarding ages is for type #4;
  • The maximum lifetime for each type is 20 years. The minimum lifetime is between 0 and 2 years.
Figure 3 depicts the histogram of the machine ages in the dataset.

2.3. Performed Statistical Analysis

In this section, we perform a specific statistical analysis.

Correlation Analysis

Initially, some quantitative indicators were determined. In the sample dataset, the nr_failures (count of failures), nr_errors (count of error statements), and nr_maintenance (count of maintenance activities) were identified for the analysis.
For the data analysis, we performed a verification of the outliers first, followed by the verification of the data normality.
An outlier that could appear in a dataset is an extreme that is statistically significantly much higher or lower than the other values from the same dataset. The appearance of outliers in datasets can influence to a high degree the evaluation results. For instance, if the mean must be calculated, a very high or low outlier influences the result to a high degree. We have verified applying the two-tailed Grubbs outliers detection test [24] at the significance level αGR = 0.05 for identifying outliers, in all the datasets, nr_errors, nr_failures, and nr_maintenance. The Grubbs test applied to the number of errors did not detect any outlier, just the value 60 that was statistically furthest from the rest. The Grubbs test applied to the number of failures did not detect any outlier, just the 19 value that was statistically furthest from the rest. The Grubbs test applied to the number of maintenances did not detect any outlier, just the 25 value that was statistically furthest from the rest. In the following analyses, we have considered even these farthest from the rest of the values.
There are diverse goodness-of-fit tests that can be applied for the verification of data normality. Among the most frequently used goodness-of-fit normality tests can be mentioned the Kolmogorov–Smirnov (KS) test [25,26]; Lilliefors (Lill) test [27], which is a Kolmogorov–Smirnov test with a specific Lilliefors correction; Anderson Darling test [28] and Shapiro–Wilk (SW) test [29]. According to [30], from the previously mentioned tests, the SW test has the highest power, but it has disadvantages as well. The SW test is recommended to be applied for small samples (where the sample size ≤ 30). In the case of a large sample size, we recommend the application of the Lill test. As the significance level of the Lill normality test, we considered in most of the cases αnorm = 0.05. The p-value of the normality test can be interpreted if p-value < αnorm, then the Null hypothesis (H0) should be rejected, and H1 accepted as the alternative hypothesis; the data failed to pass the normality assumption at the αnorm significance level. Elsewhere, if p-value ≥ αnorm, H0 could not be rejected, and the data passed the normality assumption at the αnorm significance level.
For additional visual validation of the result of the Lill normality test, we recommend the use of the quantile-quantile plot (Q-Q plot) visual representation [31]. The Q-Q plot is a scatterplot appropriate for normality visual appreciation. If the data are normally distributed data, the points should fall approximately along this reference line. The larger the departure from the reference line, the greater the evidence for the conclusion that the data failed the normality assumption.
We considered that the data sample normality was an influencing factor that must be analyzed for the data characterization. In the following, we have verified the assumption of normality of the variables nr_errors, nr_failures, and nr_maintenance.
For the verification of normality, based on the fact that the sample size was larger by 30, we applied the Kolmogorov-Smirnov test with the Lilliefors improvement (Lill test) [29] at the αnorm = 0.05 significance level. Table 2 presents the obtained results by applying the Lill test, with the considered αnorm = 0.05 significance level.
For additional visual validation of the numerical normality analysis results, we created the Q-Q plots corresponding to the variables, nr_errors (Figure 4), nr_failures (Figure 5), and nr_maintenance (Figure 6). Based on the obtained numerical results and the visual validations, it can be concluded that the number of failures passed the normality assumption. The number of errors failed to pass the normality assumption. The number of maintenances failed to pass the normality assumption. Figure 7 depicts the relation between error statements, failures, and maintenance count.
For calculating the correlation coefficient of two variables, we recommend the Pearson [7,32] or Spearman [32,33] correlation coefficient as being the most appropriate in many cases. We have defined the decision rule for choosing between the Pearson or Spearman correlation coefficient based on the normality of the variables whose correlation was studied. The Spearman correlation coefficient is more appropriate in the nonparametric case when the data (variable) fail to pass the normality assumption. When the data (variable) normality assumption passes, we recommend the Pearson correlation coefficient [32] to be used.
Based on the results of the normality analysis, we decided to use the Spearman correlation coefficient r [33,34] between the number of errors and the number of failures. The obtained Spearman r = 0.448 numerical value indicated a moderate correlation. For verification, if r was statistically significant, we have calculated the 95% confidence interval (CI) obtaining [0.2703, 0.5961]. We found that r > 0 and 0.2703 > 0, which proves that the correlation was statistically significant. Additionally, we have applied the statistical ANOVA test at the significance level αANOVA = 0.05 for verification of the hypothesis that r was statistically significantly different from 0. The obtained p-value of 0.0001 indicated that the difference was statistically significant (0.0001 < αANOVA).
We calculated the Spearman correlation coefficient r between the number of failures and the number of maintenances obtaining a very small r = −0.04585. For verification, if r was statistically significant, we have calculated the 95%CI obtaining [−0.2457, 0.1577]. Therefore, r < 0 and 0.1577 > 0 proved that there was no correlation. Additionally, we have applied the statistical ANOVA test at the αANOVA = 0.05 significance level for verification of the hypothesis that r was statistically significantly different from 0. The obtained p-value of 0.6506 indicated that the difference was not statistically significant (0.6506 > αANOVA).
We calculated the Spearman correlation coefficient r between the number of errors and the number of maintenances, obtaining a very small r = −3.028 × 10−6. For verification, if r was statistically significant, we have calculated the 95% CI obtaining [−0.2021, 0.2021]. r < 0 and 0.2021 > 0 proved that there was no correlation. Additionally, we have applied the statistical ANOVA test at the 0.005 significance level for verification of the hypothesis that r was statistically significantly different from 0. The obtained p-value of 0.9999 indicated that the difference was not statistically significant (0.9999 > αANOVA).
In conclusion, the correlation analysis was realized using the Spearman correlation coefficient on the number of error statements and the number of maintenance events. The correlation analysis has provided the following results:
  • There was a moderate correlation between error statements and failures, with a value of 0.448. This was only −3.028 × 10−6 between error statements and maintenance, while in the case of failures and maintenance, this was −0.04585;
  • Our initial expectation was that the more failures there were, the more maintenance there was. In fact, due to regular/predictive maintenance, the correlation between the two types of events was not significant.
  • There was no relationship between error statements and maintenance.

2.4. Discussion on Error Statements by Machines

It can be assumed that if there are many similar types of machines in the supervised system, the phenomena observed on a machine could happen with similar machines as well. If we can provide a good estimation of the distribution of the error statements per machine, the error statement itself can be predicted with greater reliability.
For a fixed period, the number of error statements can be counted. From the distribution of these (see Figure 8), the following conclusions can be drawn:
  • The expected value can be estimated by the average values in the period;
  • The standard deviation indicates how similar the machines are.
In the following, we have studied how the age of machines affected the count of error statements. This can be seen in Figure 9; the visual analysis indicates no correlation. This can be interpreted as the maintenance work performed on the machines reducing the significance of the machines’ ages.
The histogram of elapsed time between errors can be seen in Figure 10. To be visually more illustrative, we have considered classes.
The dataset contained telemetry data as well. According to our assumption, telemetry data may have a relationship between the signals and the failures/error statements. It is assumed that telemetry data may be used to forecast future errors of the machine, and thus it may indicate the need for (condition-based) maintenance. For example, vibration and resonance may cause degradation of the lifetime of a rotating component, thus leading to mechanical failures. This research aimed to verify if recognizing patterns in telemetry data can be used to predict forthcoming failures, and thus form a base of CBM.
As discussed before, the dataset investigated consisted of a telemetry dataset of various machines. For example, the telemetry data of machine 1, sample 0 are plotted in Figure 11 as follows:

3. The Study of the AdaBoost Algorithm for Prediction

In this section, the details of the study are presented. In the previous sections, the goal of the research and the available dataset were discussed. First, the AdaBoost algorithm will be shown. After that, an overview will be given of the existing application areas of the algorithm. Finally, two experiments will be evaluated on the dataset.

3.1. A Summary of the AdaBoost Algorithm

The AdaBoost algorithm [35,36] is appropriate for accelerating machine learning algorithms and increasing their performance by making a strong classification as a linear combination of weak classifications with appropriate weights. The AdaBoost algorithm belongs to the class of boosting algorithms [37]. The quality of the resulting classification is influenced to a high degree by the definition of initial weak classifiers. The studies [38,39] proved that combining weak learners may form a string learner. Some enhancements of the original algorithm have been evaluated when running the method. The authors of [39] proposed a weight allocation scheme to enhance the generalization effect. By selecting and combining information in the dataset, different prediction models can be created.
The AdaBoost algorithm is a special case of combined classification methods. Combined classifiers are created by mixing multiple classes of classifiers and methods for making forecasts. This way, they increase the accuracy of the classification. During the learning phase, base classifiers are created. The classification itself relies on voting. The algorithm in its general form is described by the following pseudo-code: Algorithm AdaBoost (Algorithm 1).
Algorithm 1: Algorithm AdaBoost
  Input: X: domain dataset; Y: label dataset; T: number of steps;
  Output: H = C, the final hypothesis
  Create D0 initial distribution
  for t = 1 to T do
     Compute Dt+1 from Dt;
     Construct a Ct classifier from Dt;
  end for
  for (all xiS test records) do
     C(xi) = Vote C1(x1),C2(x2),…,Cm(xm);
  end for
End AdaBoost
The explanation of the algorithm is as follows. The training dataset consists of suggestions: (x1, y1) … (xm, ym) where xi belongs to the domain space X, and yi is the corresponding label. In our research, we assumed Y = {−1, 1}. (xi, yi) is the experimental classification of phenomenon i, which can be inaccurate. Dt is the distribution on round t, in other words, a weight that describes the importance of the classification of xi. In this algorithm, the weight is increased so that it acquires more focus when the classification is incorrect. Initially, the weights are equal (see Equation (15)). At each iteration step, an error is calculated, see Equation (16), and the distribution is updated and normalized (see Equations (17)–(19)). The output is the final hypothesis, which is a result of a voting mechanism. In this paper, a linear combination was used (20), but there can be other voting algorithms as well.
Boosting algorithms [40] are sequential in the sense that the creation of consecutive classification models depends on the data of the previous model. In the first model, we took into account each object of the training dataset with the same weight to see whether the model classified each instance correctly or incorrectly. In the former case, we reduced the weights in the latter to pay more attention to the improperly classified objects in the next round. During the procedure, this step was repeated through several iterations, with the purpose that the algorithm at a certain iteration will be correctly classified as the appropriate class.
There are several variants of the general boosting algorithm depending on how the parameters are used (how to apply weights on instances, what kind of basic classifier to use, and how to combine the models obtained during the process). The most widespread of these is the state-of-the-art AdaBoost algorithm [41,42,43]. Ref. [44] compared the prediction capabilities of the AdaBoost algorithm with the backpropagation neural network (BPNN), regression classifier, support vector machine (SVM), and support vector regression (SVR).

3.2. Applications of the AdaBoost Algorithm

Various application fields of the AdaBoost algorithm can be found in the scientific literature. Ref. [45] used the AdaBoost algorithm in the road engineering field to predict performance indicators of asphalt concrete roads. Leaf nitrogen concentration was estimated by the AdaBoost-based machine learning algorithm in [46]. Vertical total electron content forecasting of the Earth’s ionosphere was discussed in [47]. COVID-19 significantly changed the financial performance of enterprises. An AdaBoost-based intelligent driving algorithm for heavy-haul trains was described in [48] to realize the intelligent control of the air brake, optimized based on two aspects: the extraction method of the training sample subset and the voting weight. A neurobiological application can be found in [49], which described a model to improve dementia prediction accuracy; in the paper, an intelligent learning system was described. AdaBoost was used for electroencephalogram epileptic signals investigation [50]. An energetical application was described in [51]: the state-of-charge prediction of lead-acid batteries by AdaBoost. An online sequential extreme learning machine model was proposed in [52] where AdaBoost and the recurrent neural network models were used for lithium batteries’ state-of-charge estimation. The architectural engineering application of the AdaBoost algorithm was reported in [53] to predict some mechanical properties of surrounding rock in tunneling. The authors of [54] presented a video-based fire smoke detection using robust AdaBoost. AdaBoost was used to increase the accuracy and reliability of a framework for daily activities and environment recognition using mobile device data [55].

3.3. Implemented and Evaluated Version of the AdaBoost Algorithm

3.3.1. AdaBoost Algorithm for Condition-Based Maintenance

In this paper, the AdaBoost algorithm will be applied to a binary classification problem in the field of condition-based maintenance, so this version is reviewed. The algorithm is capable of executing both binary classification and multiclass problems [56]. In this case, the input dataset consists of (x1, y1), (x2, y2),…,(xN, yN) pairs, where xi is the property vector of the specified entity, yi = {0, 1} is a label, and i = {1, 2,…, N}.
Suppose teaching consists of t turns. The AdaBoost algorithm calls ht a weak classification procedure in each step. Most often, the decision chunk algorithm is used, which divides the teaching patterns along a single attribute value and cuts at a threshold d. The procedure is similar to a single-level decision tree model, i.e., one leaf of the tree will have an attribute value smaller than d, while the other leaf is larger than d.
Let us denote the weight of the i-th learning dataset of step t Dt(i). Initially, all the weights are calculated by (15):
D t ( 0 ) = 1 2 m i f   y t = 0 1 2 l i f   y t = 1
where m and l are the counts of negative and positive examples, respectively. In each iteration, the algorithm re-divides the weight of the training samples by increasing the weight of poorly classified patterns. As a result, poorly performing learning classifiers will receive more attention. The error [57] is evaluated as follows. Let us denote X as the set of property vectors of the data also known as a weak hypothesis: hi: X→{−1, 1}. In step t, the error ε t of the hypothesis can be calculated as the weighted sum of the wrongly classified entities, in other words, where the predicted ht(xi) is not equal to the training sample yi:
ε t = i : h t ( x i ) y i N D t ( i )
To calculate the distribution in the next step, we need an update parameter. Ref. [40] suggested choosing an α update parameter as:
α = 1 2 · l n 1 ε t ε t
In step t + 1, the weights are calculated as
D t + 1 ( i ) = D t ( i ) · e s t ( i ) α t h t ( x i )
st(i) is a sign which is −1 if the hypothesis i is correct, +1 otherwise:
s t i = 1 i f   h t ( x i ) = y i + 1 i f   h t ( x i ) y i
As D t + 1 ( i ) is a distribution, we need to apply a normalization in each step using a normalization factor Zt.
D t + 1 ( i ) = D t + 1 ( i ) Z t
The final hypothesis is the linear combination of the individual hypotheses [40]:
H ( x i ) = s i g n t = 1 T α t · h t ( x i )

3.3.2. Experimental Evaluation of the AdaBoost Algorithm

To cover a larger diversity of situations, we have performed more representative experiments.
Experiment 1—using no iteration
In the first experiment, the AdaBoost algorithm was executed to predict error statements and failures. There were one hundred machines in the dataset, and four telemetry types for each, so a total of four hundred records were available. The goal was to process three-quarters of the data to predict the failures in the fourth quarter. The input of the algorithm contained the number of maintenance events in the known quarters. Two types of weak classifiers were used, and the experiments were run ten times as the weight calculation had a random factor. Table 3 presents a screenshot of the obtained experimental evaluation results.
In order to formulate accurate conclusions, we performed an in-depth statistical analysis of the results obtained. For both precision with decision stump (PDS) and precision with support vector classifier (PSV) as a first step, we have verified, using the Grubbs outliers detection test at the αGR = 0.05 significance level, whether outlier values could be detected that were statistically significantly different from those others. By applying the Grubbs test to the PDS, it was identified that the value 0.7583 was the furthest from the rest, but not a significant outlier (the obtained p-value, p-value > αGR). By applying the Grubbs test, at the αGR significance level, to PSV, it was identified that a value of 0.6917 was furthest from the rest, but not a significant outlier (the obtained p-value, p-value > αGR 0.05).
As a next step, we have made a descriptive statistic for both PSD and PSV. For measuring the variability, we calculated the standard deviation (SD). Furthermore, we calculated the minimum, maximum, range, mean, 95 CI% of the mean, and median. For measuring the data homogeneity-heterogeneity, the coefficient of variation (CV) C V = S D m e a n × 100 was calculated, where CV < 10 indicates a homogeneous dataset; CV ∈ [10, 20) indicates a relatively homogeneous dataset; CV ∈ [20, 30) indicates a relatively heterogeneous dataset; CV ≥ 30 indicates a heterogeneous dataset. Table 4 presents the obtained experimental evaluation results and descriptive statistical characterization.
Further, we analyzed the data normality of PDS and PSV. Based on the fact that the sample sizes were very small 10, 10 < 30 (where 30 can be indicated as a threshold), we have chosen the Shapiro–Wilk (SW) test of normality. The SW test has higher power [29] compared with other very frequently applied statistical tests such as the Kolmogorov–Smirnov, Lilliefors (Kolmogorov–Smirnov test with Lilliefors correction), and Anderson–Darling tests. Table 5 presents the results of the SW test applied with the αnorm = 0.05 significance level. The obtained results indicate that PDS did not pass the normality assumption, and just PSV passed the normality assumption.
For additional visual validation of the normality assumption analysis results, we have plotted the Q-Q plot for PSD (Figure 12) and PSV (Figure 13). The visual interpretation of Figure 12 and Figure 13 led to the same conclusions as the SW test results (Table 5).
Based on the results regarding the normality analysis in a nonparametric case, we have noticed that the median was a more appropriate indicator of the central performance tendency than the mean. For the verification of the null hypothesis (H0) that between the medians of PSD and PSV there was no statistically significant difference, we have applied the nonparametric Mann–Whitney test [58] at the αMW = 0.05 significance level. As the result of the test application, it was obtained that the p-value denoted pMW = 0.0002, with pMW < αMW indicating that the medians of PDS and PSV were statistically significantly different. (H0 should be rejected and H1, the alternative hypothesis, must be accepted). Based on the performed analyses, it can be concluded that the decision stump classifier runs better than the support vector classifier.
Experiment 2 with an iteratively increased dataset
In the second experiment, the training set was iteratively increased. It always contained those 3616 records, where the output was 1, and it was extended by 1 × 3616, 2 × 3616, …, 10 × 3616 other records randomly. This increased training data produced a better hit ratio (see Table 6). However, by looking at the confusion matrix of the 10th run (22), which provided the most precise result, you could see:
10,780 8 1143 2
The algorithm correctly classified 10,780 of the 11,923 test input values. However, only two failures of 1143 were classified correctly. This rate is very poor [59] for failure prediction capabilities, and its root is probably the lack of failure records in the database. However, there are domains where the sensitivity is extremely important, i.e., to not miss true positives. There are two reasons for this. On the one hand, the number of normal operation samples is almost ten times bigger than the number of failure samples. On the other hand, there might be no relationship between the measured signals and the observed failures. Some of the referenced applications for the AdaBoost algorithm reported better performances, but their dataset might be better for their purpose.
Further indicators have been calculated according to [60]. The true positive rate (TPR), also known as the sensitivity hit rate:
T P R = T P ( T P + F N ) = 0.9041
The true negative rate (TNR), also known as specificity or selectivity:
T N R = T N ( T N + F P ) = 0.2
The positive predictive value (PPV), also known as precision:
P P V = T P ( T P + F P ) = 0.9992
The negative predictive value (NPV):
N P V = T N ( T N + F N ) = 0.0017
The false negative rate (FNR), also known as the miss rate:
F N R = 1 T P R = 0.0959
The false positive rate (FPR), also known as fall-out:
F P R = 1 T N R = 0.8
The false discovery rate (FDR):
F D R = 1 P P V = 0.007
The false omission rate (FOR):
F O R = 1 N P V = 0.9982
The positive likelihood ratio (LR+):
L R + = T P R F P R = 1.1301
The negative likelihood ratio (LR):
L R = F N R T N R = 0.4793
The prevalence threshold (PT):
P T = T P R ( T P R + F P R ) = 0.4847
The threat score (TS) or critical success index (CSI):
T S = T P ( T P + F N + F P ) = 0.9035
The prevalence:
P r e v a l e n c e = ( T P + F N ) ( T P + F N + T N + F P ) = 0.9991
The accuracy (ACC):
A C C = ( T P + T N ) ( T P + F N + T N + F P ) = 0.9035
The balanced accuracy (BA):
B A = T P R + T N R = 0.5520
The F1 score, which is the harmonic mean of precision and sensitivity:
F 1 = 2 T P ( 2 T P + F P + F N ) = 0.9035
Matthew’s correlation coefficient (MCC):
M C C = T P + T N F P + F N T P + F P ( T P + F N ) ( T N + F P ) ( T N + F N ) = 0.0013
The Fowlkes–Mallows index (FM):
F M = ( P P V · T P R ) ) = 0.9505
Bookmaker informedness (BM):
B M = T P R + T N R 1 = 0.1041
Markedness (MK):
M K = P P V + N P V 1 = 0.0010
The diagnostic odds ratio (DOR):
D O R = L R P L R N = 2.3578
We have analyzed the presence of outliers in the training dataset size (Table 6). The value 7232 was identified as furthest from the rest, but no significant outlier was detected (p-value > 0.05).
The precision data from Table 7 was analyzed in-depth. The Grubbs identified that the value 0.5442 was furthest from the rest, but not a significant outlier (p-value > 0.05).
Based on the fact that the sample size [61] was 10, which is very small, we have chosen the application of the Shapiro–Wilk test (SW) of normality at the αnorm = 0.05 significance level. We obtained, as the results, the test statistics 0.851 and the p-value 0.06 (0.06 ≥ αnorm), which indicated that the data passed the normality assumption. Based on this finding, the best indicator of the central tendency was the mean = 0.7988. The results also indicated that the 95 CI% of the mean [0.7157, 0.8819] must be reported.

4. Conclusions and Future Directions

The main purpose of this study was to assess the performance of the AdaBoost algorithm for researchers who intend to use it for problem-solving for diverse data. The study presented an approach to using sensor data to predict the probability of machine failures to track the remaining useful life, so that the proposed algorithm applied machine learning in the field of condition-based maintenance. The AdaBoost algorithm can analyze sensor data from equipment to predict when a machine is likely to fail. This enables maintenance personnel to schedule maintenance in advance, reducing downtime and maintenance costs. By analyzing sensor data and predicting equipment failure, the AdaBoost algorithm can help operation management to move from reactive maintenance to proactive maintenance, thus improving productivity and reducing costs.
Two models were implemented and evaluated by experiments. In the first experiment (300 samples, random factor of weight calculation), two types of classifiers were used. In the decision stump method, the precision of the classification fell into 65–75%, while the support vector classifier’s basic classifiers generally were 5–10% worse. The forecast accuracy (especially for the first version) can be considered good, but not excellent. In the second experiment (the training set was iteratively increased), an attempt was made to predict error statements by the four physical telemetry signals. Since less than 0.42% of the dataset reported errors, only some proportion of the entire dataset was taken into account during the training.
The main purpose of this study was to show the limitations of the AdaBoost algorithm for researchers who intend to use it for diverse data. Unfortunately, the AdaBoost algorithm did not produce the expected forecasting capability on the data used, as proven by the confusion matrix: the overall accuracy was not satisfactory. The reason for this could be that the classifiers were not detailed enough or the fact that there was no relationship between the telemetry data and error statements in the dataset we investigated. The authors also ran the AdaBoost algorithm on simulated data, where it performed better. Unfortunately, there were no other real-world data in the field of CBM to test the performance of the algorithm against.
Our research included, among others, a correlation analysis for which we have proposed a decision rule for choosing between the Pearson and Spearman correlation based on data (variable) normality. For a comparison of the performance of the two algorithms, the decision rule for establishing the appropriate central performance tendency indicator was presented. For an accurate comparison of the central performance tendency of the two algorithms, an in-depth statistical analysis was presented that considered the experimental evaluation results’ variability and the existence of outlier experimental evaluation results.
As future work, the authors plan to build a sensor set with collective intelligence in a real environment able to collect and intelligently analyze data. The authors expect that more reliable estimations can be made using data whose collection is based on prior knowledge. Furthermore, sensor data may be used to show the opportunity for optimization of the processes. Another direction of future work could be to compare AdaBoost against other machine learning algorithms. Another research direction is to study the long short-term memory model using a diverse dataset, including the dataset used in the research presented in this paper and its comparison with the Adaboost algorithm.

Author Contributions

Conceptualization, O.H. and L.B.I.; methodology, O.H. and L.B.I.; software, O.H.; validation, O.H. and L.B.I.; formal analysis, O.H. and L.B.I.; investigation, O.H. and L.B.I.; resources, O.H. and L.B.I.; data curation, O.H.; writing—original draft preparation, O.H.; writing—review and editing, L.B.I.; visualization, O.H. and L.B.I.; supervision, O.H.; project administration, O.H.; funding acquisition, O.H. and L.B.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by National Research Development and Innovation Office, Hungary, grant number 2020-1.1.2-PIACI-KFI-2020-00147.

Data Availability Statement

This report is based on the data for predictive maintenance, which are available for download at https://www.kaggle.com/arnabbiswas1/microsoft-azure-predictive-maintenance/code, accessed on 10 December 2022.

Acknowledgments

This research was realized in consultancy with members of SOON project developed in the framework of the CHIST-ERA program supported by the Future and Emerging Technologies (FET) program of the European Union through the ERA-NET Cofund funding scheme under the grant agreements, title: Social Network of Machines (SOON). This research was partially supported by a grant of the Romanian National Authority for Scientific Research and Innovation, CCCDI-UEFISCDI, project number 101/2019, COFUND-CHISTERA-SOON, within PNCDI III. Also, we would like to thank the Research Center on Artificial Intelligence, Data Science, and Smart Engineering (Artemis) for the support.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

CBMCondition-Based Maintenance
IoTInternet of Things
PHMPrognostics and Health Management
RULRemaining Useful Lifetime
Q–Q plot Quantile-Quantile plot
CIConfidence Interval
ANOVAAnalysis of Variance
BPNNBackpropagation Neural Network
SVMSupport Vector Machine
SVRSupport Vector Regression
PDSPrecision with Decision Stump
PSVPrecision with Support Vector Classifier
SDStandard Deviation
CVCoefficient of Variation
SWShapiro–Wilk Test of Normality
MWMann–Whitney Test

References

  1. Vijayan, D.S.; Rose, A.L.; Arvindan, S.; Revathy, J.; Amuthadevi, C. Automation systems in smart buildings: A review. J. Ambient. Intell. Humaniz. Comput. 2020, 1–13. [Google Scholar] [CrossRef]
  2. Oliveira, M.; Torgo, L.; Costa, V.S. Evaluation Procedures for Forecasting with Spatiotemporal Data. Mathematics 2021, 9, 691. [Google Scholar] [CrossRef]
  3. Leon, F.; Gavrilescu, M. A Review of Tracking and Trajectory Prediction Methods for Autonomous Driving. Mathematics 2021, 9, 660. [Google Scholar] [CrossRef]
  4. Haq, I.U.; Ullah, A.; Khan, S.U.; Khan, N.; Lee, M.Y.; Rho, S.; Baik, S.W. Sequential Learning-Based Energy Consumption Prediction Model for Residential and Commercial Sectors. Mathematics 2021, 9, 605. [Google Scholar] [CrossRef]
  5. Alanne, K.; Sierla, S. An overview of machine learning applications for smart buildings. Sustain. Cities Soc. 2022, 76, 103445. [Google Scholar] [CrossRef]
  6. Iantovics, L.B. Black-Box-Based Mathematical Modelling of Machine Intelligence Measuring. Mathematics 2021, 9, 681. [Google Scholar] [CrossRef]
  7. Iantovics, L.B.; Kountchev, R.; Crișan, G.C. ExtrIntDetect-A New Universal Method for the Identification of Intelligent Cooperative Multiagent Systems with Extreme Intelligence. Symmetry 2019, 11, 1123. [Google Scholar] [CrossRef] [Green Version]
  8. BS EN 13306; Maintenance-Maintenance Terminology. British Standards Institution: London, UK, 2017.
  9. Hiruta, T.; Uchida, T.; Yuda, S.; Umeda, Y. A design method of data analytics process for condition based maintenance. CIRP Ann. 2019, 68, 145–148. [Google Scholar] [CrossRef]
  10. Ahmad, R.; Kamaruddin, S. An overview of time-based and condition-based maintenance in industrial application. Comput. Ind. Eng. 2012, 63, 135–149. [Google Scholar] [CrossRef]
  11. Gouriveau, R.; Medjaher, K.; Zerhouni, N. From Prognostics and Health Systems Management to Predictive Maintenance 1: Monitoring and Prognostics; John Wiley & Sons: Hoboken, NJ, USA, 2016; pp. 67–135. [Google Scholar]
  12. Muhonen, T. Standardization of Industrial Internet and Iot (Iot–Internet of Things)–Perspective on Condition-Based Maintenance; University of Oulu: Oulu, Finland, 2015. [Google Scholar]
  13. Jo, O.; Kim, Y.-K.; Kim, J. Internet of things for smart railway: Feasibility and applications. IEEE Internet Things J. 2017, 5, 482–490. [Google Scholar] [CrossRef]
  14. Xu, X.; Chen, T.; Minami, M. Intelligent fault prediction system based on internet of things. Comput. Math. Appl. 2012, 64, 833–839. [Google Scholar] [CrossRef] [Green Version]
  15. Fumeo, E.; Oneto, L.; Anguita, D. Condition based maintenance in railway transportation systems based on big data streaming analysis. Procedia Comput. Sci. 2015, 53, 437–446. [Google Scholar] [CrossRef] [Green Version]
  16. Kumar, A.; Shankar, R.; Thakur, L.S. A big data driven sustainable manufacturing framework for condition-based maintenance prediction. J. Comput. Sci. 2018, 27, 428–439. [Google Scholar] [CrossRef]
  17. Idé, T. Collaborative anomaly detection on blockchain from noisy sensor data. In Proceedings of the 2018 IEEE International Conference on Data Mining Workshops (ICDMW), Singapore, 17–20 November 2018; pp. 120–127. [Google Scholar]
  18. Cerna, M.; Harvey, A.F. The Fundamentals of FFT-Based Signal Analysis and Measurement; Application Note 041; National Instruments: Austin, TX, USA, 2000. [Google Scholar]
  19. Si, X.S.; Wang, W.; Hu, C.H.; Zhou, D.H. Remaining useful life estimation–A review on the statistical data driven approaches. Eur. J. Oper. Res. 2011, 213, 1–14. [Google Scholar] [CrossRef]
  20. Prajapati, A.; Bechtel, J.; Ganesan, S. Condition based maintenance: A survey. J. Qual. Maint. Eng. 2012, 18, 384–400. [Google Scholar] [CrossRef]
  21. Peng, Y.; Dong, M.; Zuo, M.J. Current status of machine prognostics in condition-based maintenance: A review. Int. J. Adv. Manuf. Technol. 2010, 50, 297–313. [Google Scholar] [CrossRef]
  22. Wilhelmsen, D.R. A Markov inequality in several dimensions. J. Approx. Theory 1974, 11, 216–220. [Google Scholar] [CrossRef] [Green Version]
  23. Gentile, C.; Littlestone, N. The robustness of the p-norm algorithms. In Proceedings of the Twelfth Annual Conference on Computational Learning Theory, Santa Cruz, CA, USA, 6–9 July 1999; pp. 1–11. [Google Scholar]
  24. Stefansky, W. Rejecting outliers in factorial designs. Technometrics 1972, 14, 469–479. [Google Scholar] [CrossRef]
  25. Lilliefors, H. On the Kolmogorov-Smirnov test for normality with mean and variance unknown. J. Am. Stat. Assoc. 1967, 62, 399–402. [Google Scholar] [CrossRef]
  26. Lilliefors, H. On the Kolmogorov-Smirnov test for the exponential distribution with mean unknown. J. Am. Stat. Assoc. 1969, 64, 387–389. [Google Scholar] [CrossRef]
  27. Dallal, G.E.; Wilkinson, L. An analytic approximation to the distribution of Lilliefors’s test statistic for normality. Am. Stat. 1986, 40, 294–296. [Google Scholar]
  28. Stephens, M.A. EDF Statistics for Goodness of Fit and Some Comparisons. J. Am. Stat. Assoc. 1974, 69, 730–737. [Google Scholar] [CrossRef]
  29. Shapiro, S.S.; Wilk, M.B. An analysis of variance test for normality (complete samples). Biometrika 1965, 52, 591–611. [Google Scholar] [CrossRef]
  30. Razali, N.; Wah, Y.B. Power comparisons of Shapiro-Wilk, Kolmogorov-Smirnov, Lilliefors and Anderson-Darling tests. J. Stat. Model. Anal. 2011, 2, 21–33. [Google Scholar]
  31. Wilk, M.B.; Gnanadesikan, R. Probability plotting methods for the analysis of data. Biometrika 1968, 55, 1–17. [Google Scholar] [CrossRef] [PubMed]
  32. Benesty, J.; Chen, J.; Huang, Y.; Cohen, I. Pearson correlation coefficient. In Noise Reduction in Speech Processing; Springer: Berlin/Heidelberg, Germany, 2009; pp. 1–4. [Google Scholar]
  33. Stigler, S.M. Francis Galton’s Account of the Invention of Correlation. Stat. Sci. 1989, 4, 73–79. [Google Scholar] [CrossRef]
  34. Iantovics, L.B.; Enăchescu, C. Method for Data Quality Assessment of Synthetic Industrial Data. Sensors 2022, 22, 1608. [Google Scholar] [CrossRef]
  35. Zhou, Z.H. Ensemble Methods: Foundations and Algorithms; Publisher CRC Press: Boca Raton, FL, USA, 2012; pp. 23–44. [Google Scholar]
  36. Schapire, R.E. Explaining adaboost. In Empirical Inference; Springer: Berlin/Heidelberg, Germany, 2013; pp. 37–52. [Google Scholar]
  37. Freund, Y.; Schapire, R.E.; Abe, N.A. A short introduction to boosting. J. Jpn. Soc. Artif. Intell. 1999, 14, 771–780. [Google Scholar]
  38. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef] [Green Version]
  39. Ding, Y.; Zhu, H.; Chen, R.; Li, R. An Efficient AdaBoost Algorithm with the Multiple Thresholds Classification. Appl. Sci. 2022, 12, 5872. [Google Scholar] [CrossRef]
  40. Schapire, R.E.; Freund, Y. Boosting: Foundations and algorithms. Kybernetes 2013, 42, 164–166. [Google Scholar] [CrossRef]
  41. Freund, R.M.; Grigas, P.; Mazumder, R. A new perspective on boosting in linear regression via subgradient optimization and relatives. Ann. Stat. 2017, 45, 2328–2364. [Google Scholar] [CrossRef] [Green Version]
  42. Freund, Y.; Schapire, R.E. Experiments with a new boosting algorithm. In Proceedings of the Machine Learning: Thirteenth International Conference, Bari, Italy, 3–6 July 1996; pp. 148–156. [Google Scholar]
  43. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  44. Tsai, J.K.; Hung, C.H. Improving AdaBoost classifier to predict enterprise performance after COVID-19. Mathematics 2021, 9, 2215. [Google Scholar] [CrossRef]
  45. Wang, C.; Xu, S.; Yang, J. AdaBoost Algorithm in Artificial Intelligence for Optimizing the IRI Prediction Accuracy of Asphalt Concrete Pavement. Sensors 2021, 21, 5682. [Google Scholar] [CrossRef]
  46. Wang, J.; Xue, W.; Shi, X.; Xu, Y.; Dong, C. AdaBoost-Based Machine Learning Improved the Modeling Robust and Estimation Accuracy of Pear Leaf Nitrogen Concentration by In-Field VIS-NIR Spectroscopy. Sensors 2021, 21, 6260. [Google Scholar] [CrossRef]
  47. Natras, R.; Soja, B.; Schmidt, M. Ensemble Machine Learning of Random Forest, AdaBoost and XGBoost for Vertical Total Electron Content Forecasting. Remote Sens. 2022, 14, 3547. [Google Scholar] [CrossRef]
  48. Wei, S.; Zhu, L.; Chen, L.; Lin, Q. An AdaBoost-Based Intelligent Driving Algorithm for Heavy-Haul Trains. Actuators 2021, 10, 188. [Google Scholar] [CrossRef]
  49. Javeed, A.; Dallora, A.L.; Berglund, J.S.; Anderberg, P. An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and AdaBoost Ensemble Learning. Life 2022, 12, 1097. [Google Scholar] [CrossRef]
  50. Al-Hadeethi, H.; Abdulla, S.; Diykh, M.; Green, J.H. Determinant of Covariance Matrix Model Coupled with AdaBoost Classification Algorithm for EEG Seizure Detection. Diagnostics 2021, 12, 74. [Google Scholar] [CrossRef]
  51. Sun, S.; Zhang, Q.; Sun, J.; Cai, W.; Zhou, Z.; Yang, Z.; Wang, Z. Lead–Acid Battery SOC Prediction Using Improved AdaBoost Algorithm. Energies 2022, 15, 5842. [Google Scholar] [CrossRef]
  52. Li, R.; Sun, H.; Wei, X.; Ta, W.; Wang, H. Lithium Battery State-of-Charge Estimation Based on AdaBoost.Rt-RNN. Energies 2022, 15, 6056. [Google Scholar] [CrossRef]
  53. Zhao, H.; Zhang, L.; Ren, J.; Wang, M.; Meng, Z. AdaBoost-Based Back Analysis for Determining Rock Mass Mechanical Parameters of Claystones in Goupitan Tunnel, China. Buildings 2022, 12, 1073. [Google Scholar] [CrossRef]
  54. Wu, X.; Lu, X.; Leung, H. A video based fire smoke detection using robust AdaBoost. Sensors 2018, 18, 3780. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Ferreira, J.M.; Pires, I.M.; Marques, G.; Garcia, N.M.; Zdravevski, E.; Lameski, P.; Flórez-Revuelta, F.; Spinsante, S. Identification of daily activities and environments based on the AdaBoost method using mobile device data: A systematic review. Electronics 2020, 9, 192. [Google Scholar] [CrossRef] [Green Version]
  56. Ying, C.; Qi-Guang, M.; Jia-Chen, L.; Lin, G. Advance and prospects of AdaBoost algorithm. Acta Autom. Sin. 2013, 39, 745–758. [Google Scholar]
  57. WANG, R. AdaBoost for feature selection, classification and its relation with SVM, a review. Phys. Procedia 2012, 25, 800–807. [Google Scholar] [CrossRef] [Green Version]
  58. Fay, M.P.; Proschan, M.A. Wilcoxon–Mann–Whitney or t-test? On assumptions for hypothesis tests and multiple interpretations of decision rules. Stat. Surv. 2010, 4, 1–39. [Google Scholar] [CrossRef]
  59. Wang, W.; Zhang, W. An asset residual life prediction model based on expert judgments. Eur. J. Oper. Res. 2008, 188, 496–505. [Google Scholar] [CrossRef]
  60. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  61. Bonett, D.G.; Wright, T.A. Sample size requirements for Pearson, Kendall, and Spearman correlations. Psychometrika 2000, 65, 23–28. [Google Scholar] [CrossRef]
Figure 1. Most common CBM techniques presented in [20].
Figure 1. Most common CBM techniques presented in [20].
Mathematics 11 01801 g001
Figure 2. The workflow followed in the research.
Figure 2. The workflow followed in the research.
Mathematics 11 01801 g002
Figure 3. The histogram of machine ages in the dataset.
Figure 3. The histogram of machine ages in the dataset.
Mathematics 11 01801 g003
Figure 4. Q-Q plot for the number of errors.
Figure 4. Q-Q plot for the number of errors.
Mathematics 11 01801 g004
Figure 5. Q-Q plot for the number of failures.
Figure 5. Q-Q plot for the number of failures.
Mathematics 11 01801 g005
Figure 6. Q-Q plot for the number of maintenances.
Figure 6. Q-Q plot for the number of maintenances.
Mathematics 11 01801 g006
Figure 7. Relation between error statements, failures, and maintenance count. (a) The number of failures; (b) the number of maintenances.
Figure 7. Relation between error statements, failures, and maintenance count. (a) The number of failures; (b) the number of maintenances.
Mathematics 11 01801 g007
Figure 8. Distribution of error counts.
Figure 8. Distribution of error counts.
Mathematics 11 01801 g008
Figure 9. Machine age analysis. (a) Error counts by the machine’s age; (b) failures by the machine’s age.
Figure 9. Machine age analysis. (a) Error counts by the machine’s age; (b) failures by the machine’s age.
Mathematics 11 01801 g009
Figure 10. Histogram of elapsed time between errors.
Figure 10. Histogram of elapsed time between errors.
Mathematics 11 01801 g010
Figure 11. Telemetry data for vibration and pressure.
Figure 11. Telemetry data for vibration and pressure.
Mathematics 11 01801 g011
Figure 12. Q-Q-plot for PSD.
Figure 12. Q-Q-plot for PSD.
Mathematics 11 01801 g012
Figure 13. Q-Q-plot for PSV.
Figure 13. Q-Q-plot for PSV.
Mathematics 11 01801 g013
Table 1. Machine model and age data.
Table 1. Machine model and age data.
Machine IDModelAge
1model318
2model47
3model38
98model220
99model114
100model45
Table 2. Verification of data normality assumption using the Lill test, αnorm = 0.05.
Table 2. Verification of data normality assumption using the Lill test, αnorm = 0.05.
nr_Errorsnr_Failuresnr_Maintenance
Test statistics0.0950.0870.120
p-value0.0260.0580.001
p-value ≥ αnormNoYesNo
Normality assumption passed (No reason for rejection of H0)NoYesNo
Table 3. Evaluation results for experiment 1.
Table 3. Evaluation results for experiment 1.
Nr. of Experimental EvaluationPDSPSV
10.6750.06083
20.6750.625
30.73330.6083
40.70830.6917
50.75830.6583
60.6750.6167
70.6750.65
80.73330.633
90.6750.65
100.6751.575
Table 4. Results of the performed descriptive statistics.
Table 4. Results of the performed descriptive statistics.
Type of CharacterizationPDSPSV
Minimum0.6750.575
Maximum0.75830.6917
Range0.08330.1167
Mean0.703330.6317
95 CI% of the mean[0.68, 0.7263][0.6083, 0.655]
SD0.0321890.032591
CV4.58
(4.58 < 10, homogeneous data)
5.16
(5.16 < 10, homogeneous data)
Median0.69170.6292
Table 5. Verification of data normality assumption using the SW test, αnorm = 0.05.
Table 5. Verification of data normality assumption using the SW test, αnorm = 0.05.
PDSPSV
Test statistics0.8110.979
p-value0.0190.962
p-value ≥ αnormNoYes
Normality assumption passedNoYes
Table 6. Experiment 2 evaluation results.
Table 6. Experiment 2 evaluation results.
Training Dataset SizePrecision
172320.5442
210,8480.6691
314,4640.7486
418,0800.7871
521,6960.8298
625,3120.8540
728,9280.8692
832,5440.8860
936,1600.8965
1039,7760.9035
Table 7. Descriptive statistics for precision (data from Table 6).
Table 7. Descriptive statistics for precision (data from Table 6).
NameValue
minimum0.5442
maximum0.9035
range0.3593
mean0.7988
95 CI% of the mean[0.7157, 0.8819]
SD0.11613
CV15.54 (relatively homogeneous data)
median0.8419
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hornyák, O.; Iantovics, L.B. AdaBoost Algorithm Could Lead to Weak Results for Data with Certain Characteristics. Mathematics 2023, 11, 1801. https://doi.org/10.3390/math11081801

AMA Style

Hornyák O, Iantovics LB. AdaBoost Algorithm Could Lead to Weak Results for Data with Certain Characteristics. Mathematics. 2023; 11(8):1801. https://doi.org/10.3390/math11081801

Chicago/Turabian Style

Hornyák, Olivér, and László Barna Iantovics. 2023. "AdaBoost Algorithm Could Lead to Weak Results for Data with Certain Characteristics" Mathematics 11, no. 8: 1801. https://doi.org/10.3390/math11081801

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop