1. Introduction
Earthquakes are calamitous events that can harm structures and human existence. The seriousness of earthquake-actuated structure damage relies upon factors like size, distance from the focal point, land conditions, and seismic structure execution. The rapid assessment of the distribution pattern and severity of damage to buildings is critical for post-event emergency response and recovery [
1]. According to Samuel Roeslin et al., earthquake information from the seismic tremor perception division shows that small seismic tremors often happen in Northern Thailand. On 5 May 2014, there was a 6.3 seismic tremor in Chiang Rai Province [
2]. The strike was a tremendous recorded seismic tremor in Thailand, damaging structures in a vast region. The time of a potential earthquake is unpredictable. Consequently, seismic danger evaluation is vital for readiness for proper tremor relief exercises. An exhibition-based quake-designing evaluation technique was created to give an essential dynamic seismic arrangement. According to the suggestions of Rao, M.V.V. et al., a detailed nonlinear relocation-based strategy is a complex computation for improving a structure’s seismic presentation. Machine learning strategies are computationally requested to evaluate enormous topographical regions [
3]. Tianyu Ci et al. stated that, after an earthquake, it is crucial to assess the level of endangerment sustained by buildings so that people can avoid being in unsafe structures. The post-earthquake structure of a building must be restored with a proper analysis of the damage. Appropriate technology must be utilized to categorize the damage to the building facilities. Building damage is visually observed and recorded. Manual assessment and categorization are labor-intensive and time-consuming and may take place for months after the disaster. Machine learning (ML) and deep learning (DL) techniques could be applied as a tool for the faster assessment of damage and early restoration, thus preventing loss of life and property [
4].
The main objective of this research was to investigate how much the performance of a deep, fully connected neural network could be improved by tuning the input features and parameters of the model compilation. For this purpose, this research is explained as follows.
Section 2 outlines the past studies related to damage grade classification using ML and DL techniques.
Section 3 describes the research workflow, the steps to frame the proposed ASR-DCNN model, and the module workflow with the architecture.
Section 4 explains the implementation steps of data preprocessing and model training.
Section 5 discusses the results of experimentation at various stages of development.
Section 6 concludes by summarizing the findings and the future scope for enhancements.
2. Literature Review
In recent years, ML and DL algorithms have been widely used in various domains. Specifically, in applications in which rapid and accurate analysis is required for large data, ML and DL methods are necessary. ML and DL algorithms significantly improve automation and accuracy enhancement in structural analysis. Based on the severity of the earthquake, there are five grades of damage to building structures. The damage level of the buildings can be assessed by the data gathered using regression methods [
5]. Xiao-Wei Zheng et al. stated that ML algorithms are instrumental when the damage-sensitive features obtained from the seismic response are influenced by operation, environmental variability, and damage. ML techniques generalize the normal structural behavior to detect deviations from the baseline period associated with destruction. To assess the risk of damage, high-rise building seismic data can be studied [
6,
7]. Although the researchers have worked on incorporating various types of ML and DL to assess the damage risk of earthquake-affected buildings, there is a vast need for a new methodology that effectively assesses the damage grade of a facility by considering all the parameters of the building damage details, according to the analysis of Kim, B et al. [
8]. According to Mangalathu S. et al., after an earthquake, it is crucial to assess the level of endangerment sustained by the buildings so that people can avoid being in unsafe structures after the calamity. The potential advantage of using ML methods, such as discriminant analysis, k-nearest neighbors, decision trees, and random forests, is that they support the faster prediction of earthquake-induced damage to structures. Using a subset of Napa earthquake data, a predictive model was developed and assessed using various ML methods, among which random forest achieved 66% accuracy in predicting the building damage [
9]. Based on the analysis of Chaparala A et al., building structural integrity can be monitored using hybrid machine learning techniques (HMLT). HMLT integrate improved classifiers such as support vector machines (SVM) and artificial neural networks (ANN), to predict the strength level of buildings. The HMLT efficacy yielded a 91% F1-score and 92% accuracy [
10]. Saengtabtim K et al. accessed the calamities of the great East Japan earthquake and tsunami of 2011. They initially checked whether the maximal values of the flow depth and flow velocity matched their critical values, and then determined a combination of the parameters that accurately estimated the structural damage level. Implementing the decision tree algorithm using critical flow depth and maximum flow velocity produced the highest accuracy for determining the degree of building damage [
11].
Kim B. et al. stated that combining convolutional neural network (CNN) with an ordinal regression model efficiently assessed the damages in buildings caused by earthquakes [
12]. Ci, T. et al. developed a deep learning model for examining the stages of building damages that produced 93.95 accuracy with hyperparameter tuning [
4]. The performance of the CNN-based model in transfer learning was promising, specifically in the geographical transferability of the trained network to imagery acquired in different locations, spatial resolutions, and platforms using the heterogeneous and substantial datasets gathered by Nex, F et al. The results demonstrate that the composition of training samples used in the network influences metrics quality. The pre-trained networks, optimized for satellite, airborne, and unmanned aerial vehicle (UAV) image spatial resolutions and viewing angles, are readily available to the scientific community to encourage their broader application [
13,
14]. According to the findings of Ahadzadeh S. et al., the cost and time of disaster management are significant. Also, the presence of the data acquired before the event is essential to evaluate the data after an earthquake. Since earthquake behavior is typically nonlinear, neural networks can be a suitable approach [
15].
P Debnath, P et al. found that, in geoscience, ML classifier algorithms could categorize the earthquake as fatal, mild, or moderate from the range of seismic magnitude. The model implemented seven machine learning classifiers using six datasets collected from India and neighboring regions [
16]. Artificial-intelligence-based (AI) solutions were frequently applied to solve civil engineering problems involving modeling and optimizing large structure systems and enormous computing resources. AI applications in civil engineering were developed using chaos theory, cuckoo search, firefly algorithm, knowledge-based engineering, evolutionary algorithms, simulated annealing, neural nets, fuzzy systems, optimization, reasoning, categorization, and learning techniques [
17].
The long short-term memory (LSTM) DL-based algorithm was applied to categorize the building damage based on the textual descriptions documented after the earthquake in 2014 in South Napa, California [
18]. The investigation of Kumari V. et al. shows that applying AI methods to civil engineering applications has produced encouraging outcomes and minimized human interaction, including uncertainty and biased judgment. A dataset containing various earthquakes explored many well-known nonparametric techniques toward rapid visual screening (RVS). The methodology also offers the potential to evaluate the sensitivity of the buildings considering their significance and exposure [
19]. Alvanitopoulos, P.F. et al. state that engineers must assess the safety of existing structure facilities and choose the appropriate course of action after an earthquake. ANN and neuro-fuzzy systems were applied to automatically classify building damages using 20 seismic parameters that characterize the basic information included in accelerograms [
20].
Lazaridis, P.C. et al. studied structural damages using actual and synthetic ground shaking sequences and reported two total damage indices. The comparison analysis yielded the most effective damage index and the best machine learning algorithm for predicting how a reinforced concrete building would respond structurally to a single or series of seismic occurrences [
21]. Morfidis K. et al. suggested that ANN-based methods for the seismic vulnerability assessment of structures could be used as alternatives to existing well-established methodologies. ANNs were not widely accepted as computational methods for predicting the seismic damage level of structures, as most civil engineers researching methods for assessing a structure’s seismic vulnerability needed prior knowledge regarding the capabilities and applications of ANN-based methods [
22]. Rashid, M. et al. discovered that the destruction to internal structures was determined for each intensity level and integrated throughout the structure with the cost of necessary repairs to determine the structure repair cost ratio (RCR). The seismic vulnerability curves can be used to estimate the economic loss (direct repairability cost) of SMRF structures and the structural RCR correlated with the seismic intensity [
23].
Rao, M.V.V. et al. analyzed the state of the structure and determined its potential risk was two critical goals of structural strength-monitoring frameworks. Investigating, identifying, and characterizing risk in complicated structures is crucial to additional strength checking. The capacities are explored as scaled variations of a fundamental Gaussian hypothesis task and a vocabulary of time-recurrence movement. Characterization is then completed by synchronizing the eliminated damaged components with the time frequency. Signals collected from sensors are disintegrated into direct blends of minimal Gaussian capabilities using the coordinated significance decay computation. The balanced scratch-off and high-pass sifting procedures are sufficiently combined to address challenges in numerical reconciliation. As opposed to earlier numerical integrators, the combination’s accuracy was improved. In contrast to fuzzy set methods, the rough set analysis uses internal knowledge and does not assume any prior models [
24].
The structural building health damage monitoring system (SBHDMS) is a powerful technology for predicting the health of civil building structures. Buildings in SBHDMS have undergone unusual alterations in terms of damage levels. Earthquakes, floods, and cyclones are natural disasters that have an extraordinary impact on structures. The sensors record the vibration data and are used to alter the construction of the building in the event of a natural disaster. The peculiar variations were examined according to the vibration data. The RAS approach forecasts the vibration data levels recorded by the sensors in the damaged structure [
25].
Gemma Cremen et al. claimed that the early detection of earthquakes reduces the risk factor and environmental hazards. Earthquake early warning (EEW) systems make real-time information about active earthquakes available, allowing people in far-off communities, governments, companies, and other settings to take prompt action to reduce the likelihood of harm or loss before the earthquake-induced ground shaking reaches them. The limitation of the existing methods is the need for engineering-related metrics to make early detection decisions [
26]. Natt, L. et al. showed that the outcomes of significant damage analysis parameters provide engineers, architects, and construction and disaster recovery management personnel with a practical understanding of the degree of damage experienced by buildings [
27]. According to Chaparala A. et al., performing feature reduction or choosing features from the dataset is crucial before classifying data to ensure accuracy. Examining the dataset using the rough set theory helped to streamline the classification process by reducing the complexity of the feature selection procedure. Rough set theory analyzes the essential parameters without added data [
28]. In order to forecast the flexural strength of ultra-high-strength concrete, Wang et al. [
29] examined different supervised ML techniques such as decision tree bagging, decision tree gradient boosting, decision tree AdaBoost, and decision tree XG boost. The findings proved that DT bagging is the method that most closely approximates experimental outcomes. In order to show how seismically vulnerable existing structures are, Ruggieri et al. [
30] published a vulnerability study utilizing machine learning (VULMA). The procedure comprised four modules that each offer specific and specialized services. Street VULMA begins by gathering raw data. Data VULMA offers a way to classify and store data. Bi VULMA rates the pictures, looks at them, and calculates the vulnerability index after using the gathered data to train multiple machine learning models for picture classification. The five most typical flaws in RC bridges may be classified and detected using convolutional neural network (CNN)-based Xception and Vanilla models for the picture categorization procedure. The two models were developed, tested, and compared using the concrete fault bridge images (CODEBRIM) dataset for multi-class and multi-target image classification. The results demonstrated the possible use of the Xception and Vanilla models to classify defects in concrete bridges and the superiority of the Xception model in highly accurate defect classification [
31].
The literature outlined that varied methodologies were used to study the damage grade caused by earthquakes. Though ML and DL technologies can be utilized to assess the damage risk of earthquake-affected buildings, several challenges exist during implementation. The research challenges lie in considering the recorded dataset’s significant features, encoding the feature values, outlier analysis, sampling the target feature, and freezing the number of hidden layers with the appropriate activation function and optimizers. The present work focuses on building an efficient neural-network-based model with a competent dataset. For this purpose, varied techniques such as the ANOVA test and 10-component PCA methods are utilized.
5. Results and Discussion
The dataset with 26 features was fitted to all the classifiers to grade the damage before and after feature scaling. The ML classifier models such as logistic regression (LReg), K nearest neighbors (KNN), kernel support vector machine (KSVM), Gaussian naive Bayes (GNB), decision tree (Dtree), extra tree (Etree), random forest (RFor), ridge classifier (Ridge), RidgeClassifierCV (RCV), stochastic gradient descent (SGD), passive aggressive (PAg), and bagging (Bagg) were implemented. The exploratory data analysis aided in understanding the pattern of data distribution. Further, an adequate dataset was needed to construct an efficient damage classification model. Thus, the performance of the classifiers was analyzed at the train–test split with various ratios (60:40, 70:30, and 80:20), with and without feature scaling, and feature selection using the ANOVA test and PCA. Initially, the performance analysis of the models was conducted on the original data considering the training and testing dataset in the 60:40, 70:30, and 80:20 ratios. The evaluation metrics obtained are tabulated in
Table 3,
Table 4 and
Table 5, and the graphical plots are shown in
Figure 23,
Figure 24 and
Figure 25. The results proved that the model’s classification accuracy and precision significantly improved after implementing feature scaling. The performance indicated that feature scaling eliminates feature biasing and enables the machine learning models to interpret the features on a similar scale. A comparison of the performance of models with various train–test split ratios showed that the model performed better when the dataset was divided into an 80:20 ratio. The accuracy of the models improved when the training data were increased.
The effectiveness of the dataset was improved by selecting the most prominent features that contributed to categorizing the building damage. A 10-component PCA was applied to select the essential features from the raw dataset before and after feature scaling. The resultant dataset was fitted to the classifiers, and performances were compared. Evaluation metrics were obtained using cross-sectioning training and testing data with 80:20, 70:30, and 60:40 ratios, as shown in
Table 6,
Table 7 and
Table 8 and
Figure 26,
Figure 27 and
Figure 28, respectively. The results verified that model performance was further improved when fitted with significant features than on the entire dataset.
5.1. ANOVA Test Analysis
The ANOVA test was used to determine the influence of the independent features on the target variable. The ANOVA test analyzes the dataset’s features by comparing the null and alternate hypotheses. The PR value implies the probability of obtaining the observed value believing that the null hypothesis is true. F denotes the ratio between the variability among the variables and within group variables. An F value less than 0.05 indicates that the variable highly influenced the target variable. The degree of freedom (df) represents the number of independent variables considered to check the variability among the group. The preprocessed dataset with 25 input variables was subjected to an ANOVA test, and the results are tabulated in
Table 9. The features plinth_area_sq_ft, position, has_superstructure_cement_mortar_stone, has_superstructure_1_mortar_brick, has_super_4_non_engineered, has_superstructure_ stone_flag, and has_superstructure_4 _ engineered have a PR(>F) >0.05, and determined insignificance. Thus, the dataset was refined by eliminating the irrelevant features to form an ANOVA-reduced building damage dataset.
5.2. PCA ANOVA-Reduced Predictive Analysis
The ANOVA-reduced building damage dataset contains 18 input features. The effectiveness of the dataset was further improved by applying a 10-component PCA, which filtered the top 10 significant features that intend to predict the building damage grade. The 10-component PCA ANOVA-reduced dataset was fitted to all the classifiers, and performances were recorded. The performance of the classifiers on feature scaling and with varied split ratios were listed in
Table 10,
Table 11 and
Table 12 and shown in
Figure 29,
Figure 30 and
Figure 31.
The extensive experimentation provided the following perceptions: 1. The ML model could perform a task more efficiently when the data used for training was considerable. The 80:20 split ratio fit well while dividing the dataset for training and testing. 2. The model interpretation of the input features would be unbiased if a similar scaling was adopted across all the features. 3. The model’s performance can be enriched if trained with relevant and significant features that determine the output. Thus, a 10-component PCA ANOVA-reduced dataset was created to build an earthquake damage prediction model.
5.3. Proposed ANOVA-Statistic-Reduced Deep Fully Connected Neural Network
The deep, fully connected neural network model that accepts 10 input components and outputs any of the five classes of damage grade was implemented using Python programming on an NVIDIA Tesla V100 GPU workstation. The proposed ASR-DFCNN was trained and tested with 80% and 20% of the data. The model was executed with a batch size of 64 for 30 epochs. The proposed model was validated by comparing the accuracy and R-squared values with all other classifiers on the reduced dataset. Also, the designed DFCNN model was implemented with the raw dataset. The observations in
Table 13 and
Figure 32 proved that the proposed ASR-DFCNN model attained a greater accuracy of 98% and R-squared 97% than other models. The DFCNN model, when trained with raw data, attained only 87% accuracy, thus proving the importance of feature selection for building an efficient model.
6. Conclusions
The presented research attempted to grade the type of building damage caused by an earthquake by analyzing the essential features. The main objective of this research is to investigate how well the performance of the deep, fully connected neural network can be improved by tuning the input features and parameters of model compilation. The contribution of the research was two-fold. The first was identifying essential data processing techniques on input features to build a competent dataset. The second focus was designing the ASR-DFCNN model-based architecture that efficiently classified the earthquake-affected building’s damage grade compared to existing ML classifiers. The challenges in building the proposed ASR-DFCNN were input feature selection and opting the best the activation and optimization functions to improve model accuracy. The earthquake damage dataset with 26 attributes describing 762,106 buildings was used to train the proposed ASR-DFCNN model, which fulfilled this research work’s requirements and outperformed the existing DFCNN and classifier models. Initially, the raw dataset without feature scaling and selection was directly fitted to the classifiers in data split ratios of 60:40, 70:30, and 80:20 for training and testing. The performance analysis concluded that the accuracy of the models without feature scaling lay between 50% to 70%. The performance increased with feature scaling and splitting data with an 80:20 ratio for training and testing. However, when the same classifiers were applied with a 10-component PCA-reduced dataset, the accuracy of the models showed better improvement. After the ANOVA test implementation, the features that produced the PR (>F) > 0.05 were considered insignificant and eliminated from the dataset. Thus, the size of the dataset was reduced from 25 to 18 input features. A 10-component PCA was applied to the ANOVA-reduced dataset to select the top 10 input features contributing significantly to damage prediction. The results exhibited that the bagging classifier with the reduced dataset produced the greatest accuracy of 83% among all the classifiers considering an 80:20 ratio of data for the training and testing phases. To enhance the performance of prediction, a deep fully connected convolutional neural network (DFCNN) was implemented with a reduced dataset (ASR).
The proposed ASR-DFCNN model was designed with the sequential keras model with four dense layers, with the first three dense layers fitted with the ReLU activation function and the final dense layer fitted with a tanh activation function with a dropout of 0.2. The ASR-DFCNN model was compiled with a NADAM optimizer with the weight decay of L2 regularization. The research model fitted with an appropriate activation function to the dense hidden layers and model optimizers reduced the loss and produced improved accuracy in damage grade classification. The ASR-DFCNN model was trained and tested with the resultant dataset and validated by comparing its performance with all other classifier models. The results proved that the ASR-DFCNN model outperformed other models by achieving 98% accuracy and 97% R-squared value. Despite the proposed ASR-DFCNN model’s remarkable performance, it is still challenging for researchers to fine-tune the sampling ratios of data features by experimenting with various oversampling or under-sampling methods. This research work could also be further enriched by extending the outlier analysis and extraction of the significant data features.