Next Article in Journal
Friction Velocity and Aerodynamic Drag Coefficient Retrieval from Sentinel-1 IW Cross-Polarization C-SAR Images under Hurricane Conditions
Previous Article in Journal
Synergistic Use of Earth Observation Driven Techniques to Support the Implementation of Water Framework Directive in Europe: A Review
Previous Article in Special Issue
Accuracy Analysis and Appropriate Strategy for Determining Dynamic and Quasi-Static Bridge Structural Response Using Simultaneous Measurements with Two Real Aperture Ground-Based Radars
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Proposed Machine Learning Techniques for Bridge Structural Health Monitoring: A Laboratory Study

1
Institute of Innovation, Science and Sustainability, Federation University Australia, Brisbane, QLD 4001, Australia
2
Centre for Infrastructure Engineering, School of Engineering, Design and Built Environment, Western Sydney University, Sydney, NSW 2747, Australia
3
Centre for Infrastructure Engineering and Safety, University of New South Wales, Kensington, NSW 2052, Australia
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(8), 1984; https://doi.org/10.3390/rs15081984
Submission received: 12 December 2022 / Revised: 17 March 2023 / Accepted: 27 March 2023 / Published: 9 April 2023
(This article belongs to the Special Issue Bridge Monitoring Using Remote Sensors)

Abstract

:
Structural health monitoring for bridges is a crucial concern in engineering due to the degradation risks caused by defects, which can become worse over time. In this respect, enhancement of various models that can discriminate between healthy and non-healthy states of structures have received extensive attention. These models are concerned with implementation algorithms, which operate on the feature sets to quantify the bridge’s structural health. The functional correlation between the feature set and the health state of the bridge structure is usually difficult to define. Therefore, the models are derived from machine learning techniques. The use of machine learning approaches provides the possibility of automating the SHM procedure and intelligent damage detection. In this study, we propose four classification algorithms to SHM, which uses the concepts of support vector machine (SVM) algorithm. The laboratory experiment, which intended to validate the results, was performed at Western Sydney University (WSU). The results were compared with the basic SVM to evaluate the performance of proposed algorithms.

1. Introduction

Structural health monitoring (SHM) can potentially provide effective solutions to the continuous assessment of infrastructural health. In the United States, every two years, approximately 600,000 bridges are inspected, scaled and rated according to their condition [1]. According to this rating, released by Federal Highway Administration, fewer than 60% of the national bridges are functioning efficiently [1]. This clearly highlights the need for structural health monitoring. Concrete degradation, steel corrosion, changes in boundary conditions, and the weakening of connections in structures over time are major concerns, which gradually deteriorate the bridge’s structural integrity and service capability if not maintained [2].
Currently, the main method of physical and functional assessment of conditions in civil infrastructure is manual, and mostly visual, inspection. This method of inspection is time consuming, expensive, and subjective. Studies have declared that such inspections are not always accurate and involve some errors, as they are highly variable and dependent on the inspectors’ proficiency, knowledge and experience, and their emotional state and alertness. Several accidents have been reported that can be related to human error-related insufficient inspections and condition assessments [3,4]. For example, I-35W Highway Bridge in Minneapolis (MA, USA) collapsed in 2007, killing 13 people and injuring a further 145. The National Transportation Safety Board classified this collapse as the consequence of safety issues, such as the lack of proper technology to accurately assess bridge conditions [5]. To prevent further incidents, it is necessary to continuously inspect and assess the condition of civil infrastructure with appropriate techniques, verifying safety and serviceability.
Therefore, putting forward a robust paradigm with respect to the aforementioned safety and economical concerns is a major challenge introduced in this study. Automated condition assessment systems based on machine learning (ML) techniques have been known as one of the technologies that can interpret a large volume of inspection data and detect early-stage structural failure. Despite the vast number of past research studies in this field, few robust methods have been proven to effectively indicate an adverse condition of a structure in service. In addition, many researchers have concentrated on evaluating the application of traditional approaches on different structural experiments, but few studies have been conducted into the enhancement of ML-based methods in SHM. Hence, this is the motivation for this present research.
In this study, we have focused on different paradigms based on machine learning. The study’s main goal is to investigate and enhance the performance of ML algorithms for damage detection in SHM based on the concepts of misclassified points, combination of kernels, and by using ensemble classifiers. As per our knowledge, no studies have previously used the first algorithm referenced in this paper in SHM application, and no researcher has investigated the combination of chosen kernels and classifiers of this study. However, we limited this investigation to the field of damage detection studies in civil engineering. Therefore, we generate data through smart aggregate (SA)-based transducers and perform intelligent analysis. Four algorithms are suggested to determine the health state of the structure. The most important contributions of this study are as follows:
  • Suggesting four SVM-based algorithms to detect damages in SHM;
  • Analyzing and comparing our proposed algorithms with other algorithms using statistical tools;
  • Verifying the proposed algorithms’ effectiveness in detecting and monitoring flexural cracks in simple concrete beams using mounted smart aggregate transducers;
  • Verifying the proposed algorithms in detecting and monitoring flexural cracks in reinforced concrete (RC) beams using mounted smart aggregate transducers.
The remainder of the article is organized as follows: the literature is reviewed in Section 2. The classification concept and the proposed learning algorithms are then presented in Section 3. In Section 4, the performance indicators are described. In Section 5, experimental results are investigated to evaluate the validity of the proposed algorithms. Finally, in Section 6, the article is concluded with discussion of its findings’ significance.

2. Literature Review

A damage detection procedure can be considered as a pattern recognition (PR) problem [6]. The solution requires a classifier that can discriminate structures as either damaged or intact. Concerning AI techniques, the most commonly used classifiers are based on discriminant analysis, artificial neural network (ANN), k-nearest neighborhood, support vector machine (SVM), support vector regression, adaptive neuro-fuzzy system (ANFIS), fuzzy inference system (FIS) and decision trees. These algorithms have been successfully utilized to solve engineering problems [7,8,9,10,11,12,13]. However, the efficiency of the solution depends on both extracted descriptors and a selected classifier. The support vector machine has been extensively accepted as an efficient classifier for detecting damages [14,15,16]. It also outperforms other techniques in this area of research [17,18,19,20,21].
Radhika et al. presented a damage detection method based on wavelet analysis and machine learning approaches. They used the wavelet to extract the statistical features and performed classification through ANN and SVM. Their results showed the superiority of SVM (66%) over ANN (61%) when compared for accuracy in a damage classification problem [17]. In another study, researchers incorporated SVM with ANN to classify structural modifications in bridges through obtained vibration data [18]. They confirmed the effectiveness of the method in the continuous monitoring of bridges. Bo et al.’s study examined the early examples of intelligence-based algorithms used for detection of structural cracks in an offshore environment [19]. They used the strain model differences as input parameters, and two artificial intelligence (AI) techniques were used to identify cracks—namely, a support vector machine and neural network. The results show the dramatic changes in strain mode in crack areas, which could be identified through both methods. Moreover, Bo et al. compared the performance of both techniques, and indicated that SVM outperformed NN with fewer errors [19].
Another study. presented the SVM-based approach to identify damages in a long-span arch bridge [20]. The researchers treated the variation ratio of curvature mode as the property to train SVM for identifying the damage level. The obtained results came close to the expected outcome. However, the researchers compared the results with the RBF neural network and verified the precision of the SVM-based method. Satpal et al.’s approach investigated the appropriateness of SVM in beam-like health monitoring through the vibration-induced modal displacement data [22]. The support vector machine was used to predict the damage location and intensity through the displacements of the first mode shape. They simulated 12 different levels of damage intensities with added white Gaussian noise. They also performed 1008 simulations and used SVM to train 90% of data, while 10% was used for testing. Their results showed that SVM has errors varying from 0.28% to 4.57% and 0% to 20.3% in predicting the location and intensity of damages, respectively, without the presence of noise. They also indicated that the presence of noise could decrease the SVM performance. However, they introduced SVM as an effective tool for structural health monitoring. There are many other studies in various domains that report the use of SVM as an effective approach [23,24,25,26,27,28,29,30]. Table 1 summarizes some of other SVM-related studies in SHM, which investigated kernels to develop more effective algorithms.
This literature review aims to identify the gap in research regarding the algorithms and SHM domain. It firstly indicates that the SVM-based model is superior to the other ML models. Due to its strong theoretical statistical framework, it has proven to be more robust in various applications and when data are integrated with noise. It also possesses high adaptability, good generalization performance, and the capability for global optimization independent of the dimensionality of the input data. SVM provides the least error and shortest processing times [35,36]. The proposed algorithms in this paper are based on SVM models. SVM has successfully solved classification, forecasting, and regression problems. Secondly, despite the diversity of kernels in SVM, the kernel combination is application dependent; however, this hybridization can outperform the single kernels. However, more investigation needs to be conducted into the kernel combinations’ performance. The accuracy will enhance if the kernel is aligned with the target information, leading to low errors [37]. Thirdly, ensemble learning is an effective technique that can enhance the models’ performance; its accuracy is also generally consistent. This method can also provide a critical boost to different challenges and generate better outputs by combining multiple results [38].
In the following section, the details of the proposed algorithms are explained, preceded by a brief review of fundamental algorithms.

3. Learning Algorithms

In this section, we detail the study’s key concepts and proposed algorithms.

3.1. Support Vector Machine (SVM)

This algorithm [39] was essentially defined to distinguish between two classes, but it could also solve multi-class problems. SVM provided the optimal hyperplane by separating two classes with the maximum margin from the hyperplane. The algorithm assumed an SVM binary classification with N samples in the data training set x 1 , y 1 , , x i , y i , , x N , y N R n × ± 1 , where x i is a feature vector and y i is the label of its class. The purpose of SVM was to find the line in such a way that provided the largest minimum distance from the labelled training data, known as the maximum margin hyperplane. Figure 1 demonstrates the basic concept of SVM.
The hyperplane is mathematically defined as a pair (w, b) through < w, x > + b = 0 formula. The hyperplane should meet the following condition, as shown in Equation (1), when linearly separating the train data:
y i w × x i + b 1 , i = 1 , , N
The distance between the hyperplane and each of the training data xi is defined through Equation (2) as:
d i = w × x i + b w
By considering both Equations (1) and (2), the following result is obtained for all x i as:
y i d i 1 w
where the 1 w is considered as the lower distance bound between the hyperplane and the training data x i . The maximum margin of the hyperplane is obtained through minimising Equation (4):
z = 1 2 w × w
Therefore, the final decision function is given as [40]:
f ( x ) = s i g n ( α i > 0 y i α i < x ,   x i > + b )
where α 1 , α 2 , , α N indicates the non-negative Lagrange multipliers related to the constraints, as shown in Equation (1). However, most of these α were usually zero, leading the small proportion of the training data x being considered as vector w. Since these points were considered as the closest points to the hyperplane, they were named the support vectors (SV). SVs were the training patterns which fell on the margin boundaries. In general, a small portion of training sample SVs was used by SVM for the purpose of classification. However, if the SVs consisted of training data beyond the corresponding margins, they were referred to as misclassified data [41]. If the linear separation of data was not feasible, then the hyperplane was pointless. Therefore, N non-negative ξ i , i = 1 , , N variables are considered such that:
y i w × x i + b 1 ξ i
This variable, ξ i , included a very small quantity of misclassified points and would be zero if the Equation (1) condition was met; otherwise, the term ξ was added to Equation (1) and generates Equation (6). However, the tolerance parameter could also cause the ignorance of some training data to make the linear hyperplane. In this situation, the generalized separation hyperplane is expressed by minimising Equation (7):
z = 1 2 w × w + c i = 1 N ξ i
The c i = 1 N ξ i expression supervised the number of misclassified points. For the smaller value of c, the above solution maximized the minimum distance of 1/w, and for the large value of c, it minimised the number of misclassified data. It was noted that the use of misclassified data could help to enhance the performance of classifiers.
Furthermore, SVM could generate non-linear decision functions by projecting the training data to the higher dimensional feature space through the non-linear map x : R n R d . This mapping was performed through the kernels [42]. Kernels execute all the essential operations in the input space through k(xi,xj) = (∅(x),∅(x)). The k(xi,xj) demonstrated the inner product in feature space and had to meet the Mercer’s condition [42]. Equation (8) indicates the kernel-based decision function as:
f ( x ) = s i g n ( α i > 0 y i α i k ( x , x i ) + b )

3.2. Kernels

Learning systems used the kernel functions to enhance their computational power. In fact, these functions mapped the data into the feature space (high dimensional space) to linearly separate the data in the new space [43]. However, the performance of kernel-based classification and transformation methods highly depends on the selection of kernel function and its parameters. Therefore, there are numerous approaches to define a suitable kernel function for classification through machine learning algorithms. Kernel design includes two approaches to table design and kernel function design. In table design, the main focus is to create the kernel table and no kernel function design is required. The elements of the kernel table were generated through the training data and optimization function. The other approach was designing the suitable kernel function for the problem [43]. Although selecting the suitable kernel function was extremely important, the different compound had a different performance, were particular problem and application. Therefore, theoretical methods for kernel selection are not completely advisable, except when being evaluated for a particular problem. Table 2 demonstrates different types of kernel functions [43].
It is important to address the challenge of making the appropriate kernel functions, as the proper selection or construction of kernel functions can significantly affect the performance of learning systems.

3.3. Proposed Algorithms

The details of the four SVM-based algorithms proposed in the present study are explained in the following.

3.3.1. SVM Based on Misclassified Data (SVM-MD)

There has been a substantial increase in the utilization of advice sets in learning algorithms. However, difficulties remain regarding the application and expression of this knowledge in terms of its constraints. Moreover, these techniques require new parameters that can enhance the SVM computational cost. In this regard, this study suggested the non-iterative algorithm, in which the subsequent knowledge is extracted from the training phase to improve the performance of SVM. In the implementation of a basic SVM algorithm, the first type of support vectors or hyperplane position is the only information that is utilized in the test phase of SVM from the training phase; the purpose of this algorithm is utilizing the second type of support vectors to provide the subsequent knowledge by the purpose of enhancement in SVM performance. The second type of support vectors was selected, as there was a lot of misclassified data, even in the existence of the optimized hyperplane [41]. Two potential sources of this misclassified data were the outliers and the non-linear separable data when using kernels. Basic SVMs do not consider non-linear separable data during the training phase. This happens in defining the constraints and the tolerance parameters of objective functions. The main concern herein is that, if the data in the test set appear practically the same as these misclassified data in the training set, they would be classified incorrectly. This misclassification could reduce accuracy, as it happened because the SVM ignored the information in the training phase. Therefore, to obtain benefits from misclassified data, the outlier’s effect should be taken into account. Searching to find more similar data to the misclassified data, as proposed in this research, could enhance SVM performance. In this method, after determining the misclassified data in the training phase, the SVM provided the advice weights which were used in conjunction with the decision values in the testing phase. These defined advised weights helped SVM to improve its accuracy while eliminating the outlier data. Equation (9) expresses the obtained misclassified data of the training phase as [44]:
M D = i = 1 N x i y i s i g n ( α j > 0 y j α j k x i , x j + b )
The right side of the above equation may include any SVM decision function. Although there was the possibility that misclassified data be null, experimental results showed that the incident of misclassified data was prevalent. For each misclassified data in M d set, the length from the corresponding k-nearest neighbour (k = 10), which have been correctly classified, is computed as:
C L x i = M i n i m u m x j c l x i c l x j
c l x i = x i M d
c l x j = x j M d
M d = 1 N j = 1 N ( x i x j ) · N = 1 , 2 , , 10 ,   y i # y j
where C L ( x i ) is the disperse of misclassified data from the 10-nearest neighbour which has been correctly classified. M d represents the average of all distances within the k-nearest neighbour (k = 10) of xi. However, by mapping the training data into higher dimension, Equation (14) can be used with respect to kernel k to compute the length as:
θ c l ( x i ) θ c l ( x j ) ) = k ( c l x i , c l x i ) + k ( c l x j , c l x j ) 2 k c l x i , c l x j 0.5
By computing the CL for each of the xk in the test set, the self-weighting (SW) is assigned as:
S W = 0 x i MD ,   if MD is empty or | | c l x k c l x i | | > CL x i 1 x i | | c l x k c l x i | | x i C L x i x i MD ,   if MD is empty or ( | | c l x k c l x i | | ) CL   ( x i )
The obtained SWs show the closeness of the test data to the misclassified data in the training phase. Therefore, the processing flow of proposed self-weighting SVM could be written as:
  • In training phase, perform the SVM training;
  • Use Equation (9) to find the misclassified data (MD);
  • Investigate the existence of misclassified data maintaining in a MD structure. If the MD includes the data, the CL is computed through Equation (10) for each member of MD; otherwise, the normal procedure of SVM is continued;
  • In the testing phase, compute the self-advised weights of each xk in the test set;
  • For each xk in the test set, the absolute values of SVM decision values are computed and scaled to [0, 1];
  • The SVM labelling is followed based on the conditions in Equation (15). The normal SVM labelling is performed if SW (xk) < decision value (xk).
However, this algorithm required a large amount of memory to store all of the support vectors and was not suitable for datasets with a large number of features, as it could become computationally intractable. On the other hand, it could be used in a wide range of applications, from text classification to image classification.

3.3.2. SVM Based on Hybrid Kernels

Although the concept of hybrid kernels was not new, limited prior research had been performed in in that area. The main concern of this section was to enhance the models based on new hybrid kernels that outperformed SVM with traditional kernels. As illustrated in various studies [27,45], generating the new kernel functions through the use of existing kernel functions was more efficient. Therefore, in this section, two hybrid kernel functions were generated through the use of polynomial and sigmoid kernels. The proposed hybrid kernels enhanced the accuracy, while keeping the number of support vectors in the range of required support vectors. In this method, the operators were applied on multiple kernel functions to provide new kernel functions, which meet the properties of each kernel combined. These operations directly affected the kernel matrix and the operation result was the positive semi-definite matrix at all times. The polynomial kernel function was a global kernel function that provided a better dissemination capability and a weaker learning ability [33], while the sigmoid kernel function provided a better global performance [46].
Sigmoid   K x i x j = t a n h γ x i T x j + c
Polynomial   K x i x j = x i x j + c a
To provide a model that employed the advantages of both kernels, new kernels were generated based on K M 1 and K S , which could provide better dissemination capability and better generalization ability. The new kernels can be formed as:
K M 1 x , z = k 1 x , z k 1 x , z
K M 2 x , z = k 2 x , z k 2 x , z
K S x , z = α k M 1 x , z + α k M 2 ( x , z )
where k 1 is the sigmoid kernel function with slope gamma and intercept c, k 2 is the polynomial kernel over X × X   R n , α ϵ R + . Therefore, the new kernel functions are mathematically expressed as:
K M 1 = tan h ( γ x i T x j + c ) tan h ( γ x i T x j + c )
K M 2 = x i x j + c a x i x j + c a
K S = α ( tan h ( γ x i T x j + c ) tan h ( γ x i T x j + c ) ) + α ( x i x j + c a x i x j + c a )
In mathematics, a proof of Equation (20) to determine that “the sum of two kernel functions still a kernel function” is shown as follows. The equation assumed that S is a set of finite points x 1 , x 2 , , x n , while k 1 and k 2 were the corresponding kernels resulted from the k 1 and k 2 on the restricted points, respectively. As illustrated above, k is the positive semi-definite matrix for all α ϵ R + as:
α k α 0
then,
α k 1 + k 2 α = α k 1 α + α k 2 α 0
Therefore, k 1 + k 2 0 is the positive semi-definite matrix and still a kernel function.
The proposed SVMs, which are based on K M 1 and K S kernels, are called SVM-S2 and SVM-SP, respectively.
Therefore, the processing flow of the enhanced algorithm can be written as:
Algorithm 1 SVM-S2 and SVM-SP
  • Implement classic SVM.
  • Tuning SVM parameters.
  • Set the kernel function.
If SVM-S2:
Set the kernel function to Equation (21).
If SVM-SP:
Set the kernel function to Equation (23).
4.
SVM training process.
5.
K-fold cross validation.
6.
SVM forecasting process.
Compute the accuracy (Acc1) and F-score (F1-S).

3.3.3. SVM Based on Ensemble Classifiers (SVM-EN)

Ensemble classifiers were the machine learning algorithms that utilized multiple learning algorithms to enhance the predictive performance that could be achieved by learning algorithms alone. The hybrid classifiers, therefore, provided better performance and accuracy than the simple individual classifier [47,48,49]. However, to build an efficient hybrid classifier system, it was required to choose the generation model scheme as well as the classification methods. There are many schemes that perform the model’s combination [48,50,51]. The main challenge that arose here was choosing and combining the diverse methods and models. The reliability and effectiveness of the system could have been directly affected by these selections, depending on the application. Therefore, the appropriate combination strategy reduced the prognosis errors.
Figure 2 and Figure 3 show two types of approaches in the model generating paradigm [52]. Figure 2 is a homogeneous model [53,54], which generates the model by employing a single learning algorithm to the different partitions of the feature vectors in the dataset.
Figure 3 is a heterogeneous model [52], which generates models by employing various learning algorithms with the same dataset of feature vectors. The obtained outcome of classifiers could be combined through different techniques, such as cascading, voting, and stacking. It should be noted that the ensemble models can solve various potential issues by only using the single classifier. These issues were associated with the proper model selection, choosing the correct local minimum and the infeasibility of the search space expansion [55].
The present study followed the heterogeneous paradigm, as seen in Figure 3. Therefore, the aggregation of the results of multiple classifiers with the ultimate goal of accuracy enhancement was presented. This ensemble embraced a set of individually trained classifiers: the enhanced support vector machine (I), the enhanced support vector machine (II), and the k-nearest neighbour classification model [56]. The process of ensemble design in this study followed two main steps. The first step was the model training, in which the above models were trained with the same training set. The second step was the model combination, in which the outputs of all classifiers made the final result through the majority voting. The idea of majority voting [57] is that each classifier provides its vote on the specific class; the majority of votes are then considered as the final output. The processing flow of this algorithm is detailed as follows:
Algorithm 2 SVM-EN
Input: X: training data, Y: class labels of X, K: number of nearest neighbors.
Output: Class of a test sample x.
Start
1. Implement algorithm 1, Section 3.3.1.
2. Compute the accuracy (Acc1) and F-score (F1-S).
3. Specify class/label.
4. Implement algorithm 2, Section 3.3.2.
5. Compute the accuracy (Acc2) and F-score (F2-S)
6. Specify class/label.
7. Classify (X, Y, x) by implementing KNN.
7.1. For each sample x do.
                     Calculate the distance: d (x, X) = i = 1 n ( x i X i ) 2 .
                End for.
           Classify x in the majority class: C ( x i ) = a r g m a x k x j K N N C ( X j , Y k ) .
7.2. Compute the accuracy (Acc3) and F-score (F1-S3).
7.3. Specify class/label.
8. Use majority voting to specify final output based on steps 3, 6, 7.4.
End.

4. Performance Indications

This study used the confusion matrix, Table 3, to evaluate the performance of algorithms by computing the accuracy and F1-score [57].
A c c u r a c y = t p + t n t p + f p + f n + t n
F 1 S c o r e = 2 × ( R e c a l l × P r e c i s i o n ) R e c a l l × P r e c i s i o n
R e c a l l = t p t p + f n
P r e c i s i o n = t p t p + f p

5. Experimental Analysis

The main purpose of this section is to present experimental data analysis based on the proposed algorithms. This experiment was performed at Western Sydney University (WSU). MATLAB R2018b software was used to analyze the data.

5.1. Concrete Beam Preparation

The blend proportion of the ready-mixed concrete was utilized to cast the concrete beams. The aggregates with the maximum size of 10 mm, a slump of 70 mm, 28 day-compressive strength of 40 MPa, and a water-to-cement ratio of 0.48 were used based on the design created in prior research studies [58]. The Australian Portland cement type GB was utilized and the concrete and sand conformed to AS3600 [58]. The material properties are summarized in Table 4 [58].
To test the design requirements, a slum test was carried out before the concrete casting. The standard cylinders were cast with the size of 102 mm × 203 mm. The same environmental conditions were applied to cast and cure the cylinders [58]. Ten concrete beams were tested. Plywood was used to fabricate the molds. After casting, the specimens were vibrated and the surface finishing was performed by hand floating.

5.2. Data Collection and Measurement Setup

In this study, the data was collected through the active sensing approach, which measures the propagation of stress waves characterized in concrete.
Figure 4 shows our measurement setup for this study. From a hardware perspective, the PC and SA transducers were connected to the data acquisition board. The guided stress wave was generated in the concrete through the SA actuator and was partially received by the SA sensors. Mounted SA transducers, a newly developed arrangement [58], were used to detect and localize the damage on concrete and reinforced concrete specimens under loading. The distances of 20 mm and 40 mm were selected for SAs and the time-domain signal was recorded.
Figure 5 and Figure 6 show SA arrangements for three-point and four-point bending tests on concrete and RC beams at Center of Infrastructure Engineering (CIE), Western Sydney University. The difference between these two experiments is that, in three-point bending test, only one crack appeared in each beam, as the concrete beams were not reinforced. When the crack happened, the concrete beam lost its resistance instantly. In four-point bending test, the crack spread into the neutral axis location, and the expected cracks occurred at the top of the specimen due to the compression effect, which caused the varying transmission properties. However, after loading, multiple foreseen cracks become visible in the mid-span region of the beam due to the concrete beam reinforcement.
This paper employs MATLAB software for analytical parts and LabVIEW software for two purposes. Firstly, the software generates a swept sine wave which is considered as an excitation wave; secondly, the received signals through the SA sensors are processed through the program. The sinusoidal sweep signal of this experiment was generated through the actuator, ranged from 100 Hz to 150 kHz, and has a magnitude of 10 V. The sweep and recording periods are 1 s and 4 s, respectively. [58]. Therefore, during each measurement, at least three complete sweep periods were recorded. However, the continuous signal recording of transducer readings was performed by the rate of 10 channels/s through the automatic data-logger for about 60 min. Ten concrete beams (400 mm × 100 mm × 100 mm) and four RC beams (1700 × 150 × 250 mm) were tested. The test was carried out through the Instron universal test machine (UTM) with a loading capacity of 200 kN and 1000 kN for concrete and RC beams, respectively. The load changes on simple and RC beam specimens were recorded through the software. The loading cell and displacement movement were set to 0 and 0.01 mm/min for concrete beams and 0 and 0.009 mm/s for RC beams, respectively. This data was used for training and testing the proposed methods. The data logger was calibrated before the beginning of each test.
The received signals were saved in the format of technical data management streaming (TDMS) and extracted to .txt file for further analysis. Figure 7 shows the time-domain signal plotted in LabVIEW.
The purpose of this data recording is to monitor the health state of structures in real-time. The continued recording of this data allowed ML techniques to detect concrete failure due to the changes happen in the value of extracted features. In total, 698 signals were recorded.

5.3. Feature Extraction

Feature extraction is the second step in every automatic system. In SHM applications, this step extracts the damage indicative features from the raw signals, which represent the most considerable changes in fault occurrence. The main task of feature extraction is to identify the signal characteristics as good indicators of faults in the structure [59].
The main concern regarding feature extraction is missing information, as some knowledge may not be captured through one particular feature or should be classified as noise instead of a feature. Therefore, it is necessary to establish some complementarity features that can represent the signal more accurately; the main challenge of this section is to provide the optimal set of properties for a feature vector. The proper selection of the properties, to provide the highest accuracy in one set, is a high priority for researchers. The combination of several different properties can provide a closer representation of the underlying phenomenon; it is more difficult to establish this understanding with only one property.
In general, the performance of feature sets is relative and can provide different outcomes based on the application. The properties can be extracted in time-domain, frequency domain and time–frequency domain formats. In many signals, the frequency component contains some important information, which is hidden and helpful in distinguishing the patterns in data analysis.
According to the literature [60,61,62,63], the following features (shown in Table 5) were extracted from the signals for further processing:
The above features are extracted from the sensor data and considered as the feature set. For each feature vector, the healthy and cracked states were labelled as 0 and 1, respectively. In the following section, by training the models through the formed feature set, crack identification is performed when the features’ values change over the healthy state. The following section investigates the results of models using the feature set.

5.4. Classification

This section discusses the results of classification on the experimental data of a simple and RC beam. The obtained features were used to investigate the performance of proposed classification techniques. The classifier learnt from the training data was used to make a decision when testing data were presented. The performance accuracy was computed through the average of 4-fold cross-validation (the data is split into four groups. Three groups were set as the training and validation data and one group was set as the test data) for 100 runs and performed for 29 observations on 699 feature sets. The average of these 29 observations represents the final accuracy and F-score. For all classifiers, the gamma was set to scale, the tolerance was set to 0.001, and the C parameter was set to 1. The kernel for Basic SVM was set to polynomial.
The first analysis trained the SVM-MD algorithm, which was based on the length of misclassified data from the corresponding 10-nearest neighbour, which were then correctly classified. The kernel type was set to polynomial. The performance of this algorithm was obtained by computing the average accuracy. The other analysis is associated with the SVM-S2 and proposed SVM-SP algorithms, which are based on the hybrid kernels. For these two models, the kernel type is set to KM1 and KS, respectively. The SVM-EN was introduced as the ensemble classifier and worked through the hybrid learning algorithms and voting system. In this algorithm, for the k-nearest neighbour model, k was set to 5, the distance metric was set to Euclidean distance and weighting was set to inverse distance weighting. Table 6 shows a comparison of accuracy between the proposed algorithms and the basic SVM. The basic SVM is considered as the benchmark.
According to the above results, the analysis conducted by estimating the performance of SVM-MD obtained an accuracy of 87.22% and F-score of 0.80 for simple beam, compared to an accuracy of 84.72% and F-score of 0.58 for the basic SVM. The analysis was repeated by obtaining the average accuracy of 86.82%, 86.46%, and 87.2% and F-score 0.77, 0.73, and 0.79 for the SVM-S2, SVM-SP and SVM-EN, respectively. However, the SVM-MD and SVM-EN yielded the highest performances; in contrast, the SVM-SP yielded a weaker performance. For RC beam specimens, the SVM-MD model showed an accuracy of 86.29% and F-score of 0.73, which outperforms the other models. The result of this experiment confirms the results obtained through the experiment on a simple beam.
Further analyses of these performances were carried out by computing the p-value of the t-test to determine the average of accuracies (shown in Table 7 and Table 8) for simple and RC beams, respectively. The details of these statistics can be found in Appendix A.
The p-value should be less than 0.05 for the results of the test to be considered statistically significant. This means that if the calculated p-value is equal to or greater than 0.05, the two samples are not significantly different. The obtained p-values in Table 7 demonstrate that the difference between the accuracies of the basic SVM and the proposed classifiers is statistically significant.
We used the confusion matrix to appraise the performance of models. This matrix computed for the known and predicted labels using the “confusionmat” function, which shows the resulting classification for structural status. The “confusionchart” function is used to carry out plotting. The results of the models are presented in Figure 8 for both simple and RC beams. It shows the summary of prediction results based on the extracted features from recorded signals, Section 5.3. The rows and columns correspond to output/predicted and target/true classes, respectively. The diagonal/off-diagonal cells correspond to observations which have been correctly/incorrectly classified. The cells show the number of observations and associated percentage figures. The right column demonstrates the percentages of all correctly/incorrectly classifications (predicted), while the bottom row shows the percentages of all correctly/incorrectly classifications (true belongings to each class). The bottom right cell demonstrates the overall accuracy.
The above figure for confusion matrices shows the number of correctly classified and misclassified data items for each model for both simple and RC beams. In the figure, “b”, “c”, “d” and “e” correspond to the defect and healthy categories on simple beams using the models. Moreover, “f”, “g”, “h”, “i” and “j” correspond to the defect and healthy categories on RC beams using the models. It can be seen from the results that the prediction accuracies of validation signals by four models are all above 86%.
From these results, its is clear that SVM-MD and SVM-EN in S.B have the optimal performance with overall accuracy of 87.2%, while SVM-SP is the least accurate algorithm in terms of defect classification. For RC beams, SVM-MD outperforms the other algorithms.
For further details on the above results, according to confusion matrix (b), 398 data points known to be in group 1 (healthy structure) are classified correctly. This corresponds to 57% of all 698 signal data. For group 2 (cracked), 53 of the data points are misclassified into group 1. Moreover, 37 of the data points known to be in group 1 are misclassified into group 2. In comparison with the basic SVM, the number of correctly classified data points has been increased, while the number of misclassifications has been reduced. Similar interpretations are applicable for other confusion matrices, as shown in Figure 8.
To evaluate the performance of models in identifying defects in beams, receiver operator characteristic (ROC) is adopted by plotting true positive rate (TPR) vs. false positive rate (FPR) for SBs. The area under the ROC curve (AUC) is an essential metric for performance assessment in machine learning study. In general, the value of AUC ranges between 0 and 1, where “0” indicates a poor model with the worst performance of separability and “1” indicates the model with excellent measure of separability [64]. As we investigated four models in this study, there are four ROC curves of different colors displayed in the Figure 9, together with corresponding AUC values.
Figure 9 displays the ROC curves with AUC values of different models. According to the values, the SVM-MD framework possesses the highest AUC, signifying that it has optimal capacity to detect and classify defects.

6. Conclusions

The purpose of this work was to enhance ML-based algorithms for SHM. Four SVM-based approaches were proposed to enhance the performance of SVM for bridge structural health monitoring. In the first approach, additional weights were assigned to the data by computing the length of each misclassified data item from the corresponding k-nearest neighbour (k = 10) which had been correctly classified. The second and third approaches introduced the new kernels through a combination of kernels, as well as polynomial and sigmoid kernel functions. In the fourth approach, the hybrid classifier was proposed through aggregating (majority voting) the different classifiers trained with the same data set. This ensemble embraces a training of individual classifiers, including the SVM-MD, SVM-S2, and the k-nearest neighbour classification model. The results showed the SVM-MD had an accuracy of 87.2% for S.B and 86.3% for RC beams, respectively; it also provided better classification performance than a basic SVM, which had an accuracy of 84.7% for S.B and 85.4% for RC beams, respectively. Though the other algorithms were less accurate than SVM-MD, they also outperformed the basic SVM.
The findings of this paper will firstly help in the health monitoring of bridge structures. The sensors are installed on the structure and the signals received by the sensors are continuously recorded. In the occurrence of a crack, the value of features extracted from the sensor signals differs from the health state. Therefore, the proposed ML-based techniques will classify these variations as crack identification.
Secondly, the paper’s findings help researchers to investigate the different models and develop new ones based on their application.

Author Contributions

Data curation, M.R.; Writing—original draft, A.N.H.; Writing—review & editing, A.N.H. and Y.Y.; Supervision, B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available on request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. t-test: two-sample assuming unequal variances between basic SVM and SVM-MD, simple beam.
Table A1. t-test: two-sample assuming unequal variances between basic SVM and SVM-MD, simple beam.
Basic SVMSVM-MD
Standard deviation0.130.03
Observations2929
Hypothesized mean difference0
df34
t Stat−97.32
P(T ≤ t) one-tail1.75 × 10−3
t Critical one-tail1.69
P(T ≤ t) two-tail3.50 × 10−3
t critical two-tail2.03
Table A2. t-test: two-sample assuming unequal variances between basic SVM and SVM-S2, simple beam.
Table A2. t-test: two-sample assuming unequal variances between basic SVM and SVM-S2, simple beam.
Basic SVMSVM-S2
Standard deviation0.130.08
Observations2929
Hypothesized mean difference0
df45
t Stat−73.63
P(T ≤ t) one-tail7.45 × 10−3
t Critical one-tail1.68
P(T ≤ t) two-tail1.49 × 10−3
t critical two-tail2.01
Table A3. t-test: two-sample assuming unequal variances between basic SVM and SVM-SP, simple beam.
Table A3. t-test: two-sample assuming unequal variances between basic SVM and SVM-SP, simple beam.
Basic SVMSVM-SP
Standard deviation0.130.08
Observations2929
Hypothesized mean difference0
df49
t Stat−58.70
P(T ≤ t) one-tail2.25 × 10−3
t Critical one-tail1.68
P(T ≤ t) two-tail4.51 × 10−3
t critical two-tail2.01
Table A4. t-test: two-sample assuming unequal variances between basic SVM and SVM-EN, simple beam.
Table A4. t-test: two-sample assuming unequal variances between basic SVM and SVM-EN, simple beam.
Basic SVMSVM-EN
Standard deviation0.130.11
Observations2929
Hypothesized mean difference0
df55
t Stat−75.20
P(T ≤ t) one-tail1.92 × 10−3
t Critical one-tail1.67
P(T ≤ t) two-tail3.85 × 10−3
t critical two-tail2.00
Table A5. t-Test: two-sample assuming unequal variances between basic SVM and SVM-MD, RC beam.
Table A5. t-Test: two-sample assuming unequal variances between basic SVM and SVM-MD, RC beam.
Basic SVMSVM-MD
Standard deviation0.240.22
Observations2929
Hypothesized mean difference0
df35
t Stat−20.37
P(T ≤ t) one-tail2.72 × 10−3
t Critical one-tail1.69
P(T ≤ t) two-tail5.44 × 10−3
t critical two-tail2.03
Table A6. t-Test: two-sample assuming unequal variances between basic SVM and SVM-S2, RC beam.
Table A6. t-Test: two-sample assuming unequal variances between basic SVM and SVM-S2, RC beam.
Basic SVMSVM-S2
Standard deviation0.240.07
Observations2929
Hypothesized mean difference0
df35
t Stat−13.75
P(T ≤ t) one-tail5.57 × 10−3
t Critical one-tail1.69
P(T ≤ t) two-tail1.11 × 10−3
t critical two-tail2.03
Table A7. t-Test: two-sample assuming unequal variances between basic SVM and SVM-SP, RC beam.
Table A7. t-Test: two-sample assuming unequal variances between basic SVM and SVM-SP, RC beam.
Basic SVMSVM-SP
Standard deviation0.240.08
Observations2929
Hypothesized mean difference0
df56
t Stat−7.65
P(T ≤ t) one-tail1.44 × 10−3
t Critical one-tail1.67
P(T ≤ t) two-tail2.87 × 10−3
t critical two-tail2.00
Table A8. t-Test: two-sample assuming unequal variances between basic SVM and SVM-EN, RC beam.
Table A8. t-Test: two-sample assuming unequal variances between basic SVM and SVM-EN, RC beam.
Basic SVMSVM-EN
Standard deviation0.240.15
Observations2929
Hypothesized mean difference0
df42
t Stat−21.78
P(T ≤ t) one-tail8.35 × 10−3
t Critical one-tail1.68
P(T ≤ t) two-tail1.67 × 10−3
t critical two-tail2.01

References

  1. Pines, D.; Aktan, A.E. Status of structural health monitoring of long-span bridges in the United States. Prog. Struct. Eng. Mater. 2002, 4, 372–380. [Google Scholar] [CrossRef]
  2. Islam, A.K.M.; Li, F.; Hamid, H.; Jaroo, A. Bridge Condition Assessment and Load Rating using Dynamic Response; Youngstown State University: Youngstown, OH, USA, 2014. [Google Scholar]
  3. Heasler, P.G.; Taylor, T.T.; Spanner, J.C.; Doctor, S.R.; Deffenbaugh, J.D. Ultrasonic Inspection Reliability for Intergranular Stress Corrosion Cracks; Nuclear Regulatory Commission: Washington, DC, USA, 1990. [Google Scholar]
  4. Zhu, Z.; Paal, S.; Brilakis, I. Detection of large-scale concrete columns for automated bridge inspection. Autom. Constr. 2010, 19, 1047–1055. [Google Scholar] [CrossRef]
  5. Bourgeois, A. I-35W Highway Bridge Collapse; University of Iowa College of Engineering: Minneapolis, MN, USA, 2007. [Google Scholar]
  6. Alavi, A.H.; Hasni, H.; Lajnef, N.; Chatti, K.; Faridazar, F. An intelligent structural damage detection approach based on self-powered wireless sensor data. Autom. Constr. 2016, 62, 24–44. [Google Scholar] [CrossRef]
  7. Azamathulla, H.M.; Guven, A.; Demir, Y.K. Linear genetic programming to scour below submerged pipeline. Ocean Eng. 2011, 38, 995–1000. [Google Scholar] [CrossRef]
  8. Chou, J.-S.; Pham, A.-D. Hybrid computational model for predicting bridge scour depth near piers and abutments. Autom. Constr. 2014, 48, 88–96. [Google Scholar] [CrossRef]
  9. Das, S.K.; Samui, P.; Sabat, A.K. Application of Artificial Intelligence to Maximum Dry Density and Unconfined Compressive Strength of Cement Stabilized Soil. Geotech. Geol. Eng. 2011, 29, 329–342. [Google Scholar] [CrossRef]
  10. Flood, I.; Christophilos, P. Modeling construction processes using artificial neural networks. Autom. Constr. 1996, 4, 307–320. [Google Scholar] [CrossRef]
  11. Salehi, H.; Das, S.; Chakrabartty, S.; Biswas, S.; Burgueño, R. Structural Assessment and Damage Identification Algorithms Using Binary Data. In Proceedings of the ASME 2015 Conference on Smart Materials, Adaptive Structures and Intelligent Systems. Volume 2: Integrated System Design and Implementation; Structural Health Monitoring; Bioinspired Smart Materials and Systems; Energy Harvesting., Colorado Springs, CO, USA, 21–23 September 2015; p. 57304. [Google Scholar]
  12. Tran, H. A hybrid fuzzy inference model based on RBFNN and artificial bee colony for predicting the uplift capacity of suction caissons. Autom. Constr. 2014, 41, 60–69. [Google Scholar]
  13. Yuvaraj, P.; Murthy, A.R.; Iyer, N.R.; Samui, P.; Sekar, S.K. Prediction of fracture characteristics of high strength and ultra high strength concrete beams based on relevance vector machine. Int. J. Damage Mech. 2014, 23, 979–1004. [Google Scholar] [CrossRef]
  14. Bornn, L.; Farrar, C.R.; Park, G.; Farinholt, K. Structural Health Monitoring With Autoregressive Support Vector Machines. J. Vib. Acoust. 2009, 131, 021004–021009. [Google Scholar] [CrossRef] [Green Version]
  15. Worden, K.; Lane, A.J. Damage identification using support vector machines. Smart Mater. Struct. 2001, 10, 540. [Google Scholar] [CrossRef]
  16. Yeesock, K.; Jo Woon, C.; Ki, H.C.; JungMi, K. Wavelet-based AR–SVM for health monitoring of smart structures. Smart Mater. Struct. 2013, 22, 015003. [Google Scholar]
  17. Radhika, S.; Tamura, Y.; Matsui, M. Cyclone damage detection on building structures from pre- and post-satellite images using wavelet based pattern recognition. J. Wind Eng. Ind. Aerodyn. 2015, 136, 23–33. [Google Scholar] [CrossRef]
  18. Alves, V.; Cury, A.; Roitman, N.; Magluta, C.; Cremona, C. Structural modification assessment using supervised learning methods applied to vibration data. Eng. Struct. 2015, 99, 439–448. [Google Scholar] [CrossRef]
  19. Bo, Y.; Cui, Y.; Zhang, L.; Zhang, C.; Yang, Y.; Bao, Z.; Ning, G. Beam Structure Damage Identification Based on BP Neural Network and Support Vector Machine. Math. Probl. Eng. 2014, 2014, 850141 . [Google Scholar] [CrossRef] [Green Version]
  20. Liu, C.C.; Liu, J. Damage identification of a long-span arch bridge based on support vector machine. Zhendong Yu Chongji/J. Vib. Shock 2010, 29, 174–178. [Google Scholar]
  21. Hirokane, M.; Nomura, Y.; Kusunose, Y. Damage detection using support vector machine for integrity assessment of concrete structures. Doboku Gakkai Ronbunshuu A 2008, 64, 739–749. [Google Scholar] [CrossRef] [Green Version]
  22. Satpal, S.B.; Khandare, Y.; Guha, A.; Banerjee, S. Structural health monitoring of a cantilever beam using support vector machine. Int. J. Adv. Struct. Eng. 2013, 5, 2. [Google Scholar] [CrossRef] [Green Version]
  23. Cao, Y.F.; Wu, W.; Zhang, H.L.; Pan, J.M. Prediction of the Elastic Modulus of Self-Compacting Concrete Based on SVM. Appl. Mech. Mater. 2013, 357–360, 1023–1026. [Google Scholar] [CrossRef]
  24. Cha, Y.-J.; Buyukozturk, O. Modal Strain Energy Based Damage Detection Using Multi-Objective Optimization. In Structural Health Monitoring; Springer: Cham, Switzerland, 2014; Volume 5, pp. 125–133. [Google Scholar]
  25. Chen, B.-T.; Chang, T.-P.; Shih, J.-Y.; Wang, J.-J. Estimation of exposed temperature for fire-damaged concrete using support vector machine. Comput. Mater. Sci. 2009, 44, 913–920. [Google Scholar] [CrossRef]
  26. Gong, L.; Wang, C.; Wu, F.; Zhang, J.; Zhang, H.; Li, Q. Earthquake-Induced Building Damage Detection with Post-Event Sub-Meter VHR TerraSAR-X Staring Spotlight Imagery. Remote Sens. 2016, 8, 887. [Google Scholar] [CrossRef] [Green Version]
  27. Huanrui, H. New Mixed Kernel Functions of SVM Used in Pattern Recognition. Appl. Adv. Comput. Simul. Inf. Syst. 2016, 16, 5–14. [Google Scholar] [CrossRef] [Green Version]
  28. Li, Z.; Burgueño, R. Using Soft Computing to Analyze Inspection Results for Bridge Evaluation and Management. J. Bridge Eng. 2010, 15, 430–438. [Google Scholar] [CrossRef]
  29. Shuai, Y.; Fang, C.Q.; Yuan, Z.J. Study on Mechanical Properties of Corroded Reinforced Concrete Using Support Vector Machines. Appl. Mech. Mater. 2014, 578–579, 1556–1561. [Google Scholar] [CrossRef]
  30. Ying, Y.; Garrett James, H.; Oppenheim Irving, J.; Soibelman, L.; Harley Joel, B.; Shi, J.; Jin, Y. Toward Data-Driven Structural Health Monitoring: Application of Machine Learning and Signal Processing to Damage Detection. J. Comput. Civ. Eng. 2013, 27, 667–680. [Google Scholar] [CrossRef]
  31. Yan, K.; Xu, H.; Shen, G.; Liu, P. Prediction of Splitting Tensile Strength from Cylinder Compressive Strength of Concrete by Support Vector Machine. Adv. Mater. Sci. Eng. 2013, 2013, 597257. [Google Scholar] [CrossRef] [Green Version]
  32. Ghiasi, R.; Torkzadeh, P.; Noori, M. A machine-learning approach for structural damage detection using least square support vector machine based on a new combinational kernel function. Struct. Health Monit. 2016, 15, 302–316. [Google Scholar] [CrossRef]
  33. Jianhong, X. Kernel optimization of LS-SVM based on damage detection for smart structures. In Proceedings of the 2009 2nd IEEE International Conference on Computer Science and Information Technology, Beijing, China, 8–11 August 2009; pp. 406–409. [Google Scholar]
  34. Kasnavi, S.A.; Aminafshar, M.; Shariati, M.M.; Emam Jomeh Kashan, N.; Honarvar, M. The effect of kernel selection on genome wide prediction of discrete traits by Support Vector Machine. Gene Rep. 2018, 11, 279–282. [Google Scholar] [CrossRef]
  35. Raghavendra, N.S.; Deka, P.C. Support vector machine applications in the field of hydrology: A review. Appl. Soft Comput. 2014, 19, 372–386. [Google Scholar] [CrossRef]
  36. Otchere, D.A.; Arbi Ganat, T.O.; Gholami, R.; Ridha, S. Application of supervised machine learning paradigms in the prediction of petroleum reservoir properties: Comparative analysis of ANN and SVM models. J. Pet. Sci. Eng. 2021, 200, 108182. [Google Scholar] [CrossRef]
  37. Kandola, J.; Shawe-Taylor, J.; Cristianini, N. On the Extensions of Kernel Alignment; Project Report; University of Southampton: Southampton, UK, 2002. [Google Scholar]
  38. Seni, G.; Elder, J. Ensemble Methods in Data Mining: Improving Accuracy through Combining Predictions; Morgan & Claypool Publishers: San Rafael, CA, USA, 2010. [Google Scholar]
  39. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Lauer, F.; Bloch, G. Incorporating prior knowledge in support vector machines for classification: A review. Neurocomputing 2008, 71, 1578–1594. [Google Scholar] [CrossRef] [Green Version]
  41. Zhan, Y.; Shen, D. An adaptive error penalization method for training an efficient and generalized SVM. Pattern Recogn. 2006, 39, 342–350. [Google Scholar] [CrossRef]
  42. Campbell, C. An introduction to kernel methods. In Radial Basis Function Networks 1; Physica Verlag Rudolf Liebing KG: Heidelberg, Germany, 2001; pp. 155–192. [Google Scholar]
  43. Moghaddam, V.H.; Hamidzadeh, J. New Hermite orthogonal polynomial kernel and combined kernels in Support Vector Machine classifier. Pattern Recogn. 2016, 60, 921–935. [Google Scholar] [CrossRef]
  44. Maali, Y.; Al-Jumaily, A. Self-advising support vector machine. Knowl.-Based Syst. 2013, 52, 214–222. [Google Scholar] [CrossRef]
  45. Huang, F.; Yan, L. Combined Kernel-Based BDT-SMO Classification of Hyperspectral Fused Images. Sci. World J. 2014, 2014, 738250. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Li, M.; Lu, X.; Wang, X.; Lu, S.; Zhong, N. Biomedical classification application and parameters optimization of mixed kernel SVM based on the information entropy particle swarm optimization. Comput. Assist. Surg. 2016, 21, 132–141. [Google Scholar] [CrossRef] [Green Version]
  47. Dasarathy, B.V.; Sheela, B.V. A composite classifier system design: Concepts and methodology. Proc. IEEE 1979, 67, 708–713. [Google Scholar] [CrossRef]
  48. Dietterich, T.G. Machine-Learning Research—Four Current Directions. AI Mag. 1997, 18, 97–136. [Google Scholar]
  49. Ho, T. Multiple classifier combination: Lessons and next steps. In Hybrid Methods in Pattern Recognition; World Scientific: Singapore, 2002. [Google Scholar] [CrossRef]
  50. Duin, R.P.W. The combining classifier: To train or not to train? In Proceedings of the Object Recognition Supported by User Interaction for Service Robots, Quebec City, QC, Canada, 11–15 August 2002; Volume 762, pp. 765–770. [Google Scholar]
  51. Valentini, G.; Masulli, F. Ensembles of Learning Machines. In Proceedings of the Neural Nets: 13th Italian Workshop on Neural Nets, WIRN VIETRI 2002, Vietri sul Mare, Italy, 30 May–1 June 2002; Volume 2486, pp. 3–22. [Google Scholar]
  52. Bahler, D.; Navarro, L. Methods for Combining Heterogeneous Sets of Classiers. Artif. Intell. 2000. [Google Scholar]
  53. Briem, G.J.; Benediktsson, J.A.; Sveinsson, J.R. Boosting, Bagging, and Consensus Based Classification of Multisource Remote Sensing Data. In Multiple Classifier Systems; Springer: Berlin/Heidelberg, Germany, 2001; pp. 279–288. [Google Scholar]
  54. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  55. Aleksandra, P.; Michael, A.; Galina, M. Heterogeneous versus Homogeneous Machine Learning Ensembles. Inf. Technol. Manag. Sci. 2015, 18, 135–140. [Google Scholar] [CrossRef] [Green Version]
  56. Samworth, R.J. Optimal weighted nearest neighbour classifiers. Ann. Statist. 2012, 40, 2733–2763. [Google Scholar] [CrossRef]
  57. Sattar, A.; Kang, B.H. AI 2006: Advances in Artificial Intelligence; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  58. Taghavipour, S.; Kharkovsky, S.; Kang, W.H.; Samali, B.; Mirza, O. Detection and monitoring of flexural cracks in reinforced concrete beams using mounted smart aggregate transducers. Smart Mater. Struct. 2017, 26, 104009. [Google Scholar] [CrossRef]
  59. Sohn, H.; Farrar, C.; Hemez, F.; Shunk, D.; Stinemates, D.W.; Nadler, B. A Review of Structural Health Monitoring Literature: 1996–2001; Los Alamos National Laboratory: Los Alamos, NM, USA, 2004. [Google Scholar]
  60. Scott, S.; Landis, E.N.; Peterson, M.L.; Shah, S.P.; Achenbach, J.D. Ultrasonic investigation of concrete with distributed damage. ACI Mater. J. 1998, 95, 27–36. [Google Scholar]
  61. Dorfman, L.S.; Trubelja, M. Structural Integrity Associates San Jose, CA. Torsional monitoring of turbine-generators for incipient failure detection. In Proceedings of the 6th EPRI Steam Turbine/Generator Workshop, St. Louis, MI, USA, 17–20 August 1999. [Google Scholar]
  62. Aggelis, D.G. Classification of cracking mode in concrete by acoustic emission parameters. Mech. Res. Commun. 2011, 38, 153–157. [Google Scholar] [CrossRef]
  63. Tayfur, S.; Alver, N.; Abdi, S.; Saatcı, S.; Ghiami, A. Characterization of concrete matrix/steel fiber de-bonding in an SFRC beam: Principal component analysis and k-mean algorithm for clustering AE data. Eng. Fract. Mech. 2018, 194, 73–85. [Google Scholar] [CrossRef]
  64. Nahm, F.S. Receiver operating characteristic curve: Overview and practical use for clinicians. Korean J. Anesthesiol. 2022, 75, 25–36. [Google Scholar] [CrossRef]
Figure 1. Basic concept of SVM [39].
Figure 1. Basic concept of SVM [39].
Remotesensing 15 01984 g001
Figure 2. Homogeneous paradigm.
Figure 2. Homogeneous paradigm.
Remotesensing 15 01984 g002
Figure 3. Heterogeneous paradigm.
Figure 3. Heterogeneous paradigm.
Remotesensing 15 01984 g003
Figure 4. Measurement system.
Figure 4. Measurement system.
Remotesensing 15 01984 g004
Figure 5. The three-point bending test, Center of Infrastructure Engineering (CIE), Western Sydney University.
Figure 5. The three-point bending test, Center of Infrastructure Engineering (CIE), Western Sydney University.
Remotesensing 15 01984 g005
Figure 6. Reinforced concrete under four-point bending, Center of Infrastructure Engineering (CIE), Western Sydney University.
Figure 6. Reinforced concrete under four-point bending, Center of Infrastructure Engineering (CIE), Western Sydney University.
Remotesensing 15 01984 g006
Figure 7. TDMS file run in LABVIEW software.
Figure 7. TDMS file run in LABVIEW software.
Remotesensing 15 01984 g007
Figure 8. Confusion matrix of enhanced SVM models for S.B and RC beams. The rows correspond to the true class, while the columns correspond to the predicted class. Diagonal and off-diagonal cells correspond to correctly and incorrectly classified observations, respectively.
Figure 8. Confusion matrix of enhanced SVM models for S.B and RC beams. The rows correspond to the true class, while the columns correspond to the predicted class. Diagonal and off-diagonal cells correspond to correctly and incorrectly classified observations, respectively.
Remotesensing 15 01984 g008
Figure 9. ROC curves for the models for defect detection.
Figure 9. ROC curves for the models for defect detection.
Remotesensing 15 01984 g009
Table 1. Summary of SVM studies based on kernels.
Table 1. Summary of SVM studies based on kernels.
Ref No.Algorithm Domain and Outcome
[31]SVM
  • Concrete strength
  • The R-square errors are:
  • 0.8115 and 0.8227 (radial basis function (RBF) and polynomial kernel function-training)
  • 0.9422 and 0.9327 (RBF and polynomial kernel function -testing)
Conclusion: Successful performance of the both the SVM and the polynomial kernel.
[32]SVM-combination of kernels (spline and wavelet)
Four-story steel structure
Enhanced accuracy about (0–8%) than;
(1)
Simple wavelet
(2)
Gaussian RBF
(3)
Thin plate spline RBF
(4)
Morlet wavelet
(5)
Sinc wavelet
(6)
Shannon wavelet
(7)
Littlewood–Paley wavelet
Enhanced accuracy about (2–4%) than;
(8)
Gaussian RBF + polynomial
(9)
Gaussian RBF + linear
(10)
Gaussian RBF + sinc wavelet
Conclusion: Hybrid kernels can be helpful in enhancing the accuracy, No investigation has been provided on performance of sigmoid kernel alone in this area or combined with any other kernels. Based on our knowledge, no or few investigations have been conducted on the applicability of a sigmoid kernel for damage detection in civil area.
[33]Combination of kernels (Gaussian RBF and Polynomial)
Better dissemination ability than the Gaussian RBF. Gaussian RBF has 6.8% higher error than the combined kernel.
Conclusion: Hybrid kernels can be helpful in enhancing the accuracy. No investigation has been provided on performance of polynomial kernel rather than the combined kernels.
[34]Linear, radial, polynomial and sigmoid kernel-based Support vector machine (SVM)
Biomedical engineering
Radial- and sigmoid-based SVM outperformed the polynomial and linear kernels significantly (p < 0.05)
Conclusion: Sigmoid kernel has provided better accuracy than other kernels in biomedical engineering. Based on our knowledge, no or few investigations have been conducted on the applicability of sigmoid kernel for damage detection in a civil engineering context.
Table 2. Different types of kernel functions.
Table 2. Different types of kernel functions.
RBF k x i x j = e γ x i x j 2
Sigmoid k x i x j = t a n h γ x i T x j + c
Polynomial k x i x j = x i x j + c a
Wavelet k x , x = i = 1 n c o s 1.75 × x i x i σ e x p x x 2 2 σ 2
Chebyshev k x , z = j = 0 n U i ( x ) U j T a x , z
Gaussian radial basis k x , x = e x p x x 2 2 σ 2
Exponential radial basis k x , x = e x p x x 2 2 σ 2
Multi-layer Fourier series k x , x = t a n h ρ x , x + e
Fourier series k x , x = s i n N + 1 2 ( x x ) s i n 1 2 ( x x )
Splines k x , x = r = 0 k x T x T + r = 0 k x T s + k x T s + k
B splines k x , x = B 2 N + 1 ( x x )
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Predicted PositivePredicted Negative
Label positive tp: true positivefn: false negative
Label negativefp: false positivetn: true negative
Table 4. Material properties in concrete mix.
Table 4. Material properties in concrete mix.
MaterialsCharacteristicsValues
Portland cement type GBSpecific gravity3.15
Natural river sand (Fine aggregate)Specific gravity2.55
Size0.15 to 4.75 mm
Natural river gravel (Coarse aggregate)Specific gravity2.60
Maximum size10 mm
Tap waterDensity998–1000 kg/m3
Table 5. Extracted features for signal processing.
Table 5. Extracted features for signal processing.
Crest factor: The L infinity norm and RMS values are computed throughout the specified dimension. C F = X 1 N n = 1 N X n 2 ,
Root-mean-square level: where x is a vector, y indicates that the y is a real-valued scalar. x R M S = 1 N n = 1 N x n 2 ,
Sparse   filtering :   where ,   f j ( i ) is the jth feature value for the ith column. m i n i m i z e i = 1 M f ^ ( i ) 1 = i = 1 M f ~ ( i ) f ~ ( i ) 2 1
Average frequency: n = number of frequency bins in the spectrum; fi = frequency of spectrum at bin i of n;
Ii = Intensity (dB scale) of spectrum at bin i of n.
f m e a n = i = 0 n I i × f i i = 0 n I i
Energy: Z is the magnitude; Es is signal energy. E = E s Z = 1 Z x ( t ) 2 d t
Maximum-to-minimum difference: A p e a k t o p e a k = A a v e r a g e × π
Rise level:RL = value (Rwave) − value (Qwave)
Fall time:tf = Time lasts for the amplitude of a pulse to fall from a specified value to another specified value.
Fall levelFL = value (Rwave) − value (Swave)
Table 6. Comparison of accuracy between basic SVM, SVM-MD, SVM-S2, SVM-SP, and SVM-EN for simple and RC beams.
Table 6. Comparison of accuracy between basic SVM, SVM-MD, SVM-S2, SVM-SP, and SVM-EN for simple and RC beams.
Models1 S. BeamRC Beam
Acc (%)F1-SAcc (%)F1-S
Basic SVM84.720.5885.380.63
SVM-MD87.220.8086.290.73
SVM-S286.820.77860.70
SVM-SP86.460.7385.540.68
SVM-EN87.20.7986.080.71
1 S. Beam is contraction of simple beam.
Table 7. t-test: basic SVM and enhanced SVMs sample for variance for simple beam.
Table 7. t-test: basic SVM and enhanced SVMs sample for variance for simple beam.
ModelsP(T ≤ t) One-TailP(T ≤ t) Two-Tail
SVM-MD1.75 × 10−33.50 × 10−3
SVM-S27.45 × 10−31.49 × 10−3
SVM-SP2.25 × 10−34.51 × 10−3
SVM-EN1.92 × 10−33.85 × 10−3
Table 8. t-test: basic SVM and enhanced SVMs sample for variance for RC beam.
Table 8. t-test: basic SVM and enhanced SVMs sample for variance for RC beam.
ModelsP(T ≤ t) One-TailP(T ≤ t) Two-Tail
SVM-MD2.72 × 10−35.44 × 10−3
SVM-S25.57 × 10−31.11 × 10−3
SVM-SP1.44 × 10−32.87 × 10−3
SVM-EN8.35 × 10−31.67 × 10−3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Noori Hoshyar, A.; Rashidi, M.; Yu, Y.; Samali, B. Proposed Machine Learning Techniques for Bridge Structural Health Monitoring: A Laboratory Study. Remote Sens. 2023, 15, 1984. https://doi.org/10.3390/rs15081984

AMA Style

Noori Hoshyar A, Rashidi M, Yu Y, Samali B. Proposed Machine Learning Techniques for Bridge Structural Health Monitoring: A Laboratory Study. Remote Sensing. 2023; 15(8):1984. https://doi.org/10.3390/rs15081984

Chicago/Turabian Style

Noori Hoshyar, Azadeh, Maria Rashidi, Yang Yu, and Bijan Samali. 2023. "Proposed Machine Learning Techniques for Bridge Structural Health Monitoring: A Laboratory Study" Remote Sensing 15, no. 8: 1984. https://doi.org/10.3390/rs15081984

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop