Next Article in Journal
Cryptanalysis of Two Recent Ultra-Lightweight Authentication Protocols
Previous Article in Journal
TwoViewDensityNet: Two-View Mammographic Breast Density Classification Based on Deep Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Activation Layer-Based Wall Quality Recognition Using Conv2D ResNet Exponential Transfer Learning Model

1
Department of Robot and Smart System Engineering, Kyungpook National University, 80, Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea
2
Department of Computer Science and Engineering, KPR Institute of Engineering and Technology, Coimbatore 641407, India
3
Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Chennai 600062, India
4
School of Architecture, Civil, Environment and Energy Engineering, Kyungpook National University, 80, Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea
*
Authors to whom correspondence should be addressed.
Mathematics 2022, 10(23), 4602; https://doi.org/10.3390/math10234602
Submission received: 12 October 2022 / Revised: 23 November 2022 / Accepted: 2 December 2022 / Published: 5 December 2022

Abstract

:
Crack detection is essential for observing structural health and guaranteeing structural safety. The manual crack and other damage detection process is time-consuming and subject to surveyors’ biased judgments. The proposed Conv2D ResNet Exponential model for wall quality detection was trained with 5000 wall images, including various imperfections such as cracks, holes, efflorescence, damp patches, and spalls. The model was trained with initial weights to form the trained layers of the base model and was integrated with Xception, VGG19, DenseNet, and ResNet convolutional neural network (CNN) models to retrieve the general high-level features. A transfer deep-learning-based approach was implemented to create a custom layer of CNN models. The base model was combined with custom layers to estimate wall quality. Xception, VGG19, DenseNet, and ResNet models were fitted with different activation layers such as softplus, softsign, tanh, selu, elu, and exponential, along with transfer learning. The performance of Conv2D was evaluated using model loss, precision, accuracy, recall, and F-score measures. The model was validated by comparing the performances of Xception, VGG19, DenseNet, ResNet, and Conv2D ResNet Exponential. The experimental results show that the Conv2D ResNet model with an exponential activation layer outperforms it with an F-score value of 0.9978 and can potentially be a viable substitute for classifying various wall defects.

1. Introduction

Cracks are the first signs of civil structure degradation, and they can arise for various reasons, including structural foundation displacement, shrinkage and extension, uneven mix, bloated soil, and overloaded environmental and manufacturing disasters. Crack detection and recognition activities can be conducted automatically or manually and are subjected to human experts visually analyzing and evaluating the structure [1]. Manual inspection involves fetching a schematic of a crack and recording the circumstances of the abnormalities. Usually, manual inspection methods take a long time, rely on the observer, are potentially sensitive to the inspector’s insight, and lack descriptive methodology [2]. Automatic inspection methods provide a coherent solution that reduces subjectivity and replaces manual observation with the human eye [3]. Automatic crack identification has been established considering slow, biased, and old human inspection processes for fast and efficient surface defect assessment [4]. An image vision computer system is introduced to address this shortcoming, with the goal of instantly and reliably transforming image or video input into meaningful intelligence. It contains structural components, recognizes modifications in a reference image, and quantifies local and global visual damage [5]. Automating the process may significantly reduce computational costs and allow regular inspection intervals. The detection of cracks and disintegration of a bridge is examined as a part of automation using the rapid Haar transform [6]. Crack detection filters of various sizes were developed to locate cracking spots in inspection imagery. A semi-automatic strategy incorporating Sobel and Laplacian operators was used to identify crack edges, and a graph search algorithm was used to obtain cracks depending on the user input [7]. A principal-component-analysis-(PCA)-based approach combined with linear structural analysis has been reported to identify linear structural fractures in concrete bridge decks with the highest classification accuracy [8]. One of the most critical structural inspection and maintenance methods is vision-based technology, which utilizes essential diagnostic imaging devices, such as sensors [9]. Miscommunication deflection assessment, steel corrosion recognition, and spalling diagnosis are some of the most recent advancements in vision-based inspection and testing [10]. As a result, this technique has some constraints in terms of real-world applications as creating an automated system that can encompass all unpredicted possibilities for fast perceptible damage remediation in the physical world is difficult [11]. Deep learning (DL) has recently been recognized as one of the most potent remedies to this challenge [12]. In addition, machine learning (ML) strategies for DL models based on neural networks containing numerous hidden units have evolved with numerous benefits for a better solution [13].
Structural inspections across long, nonintrusive distances based on high precision have been performed using optics and computer vision advancements. The fundamental shortcoming of image-processing-based approaches is the lack of consistency and vibration among the fracture pixels [14]. DeepLab V2 was used to detect several cracks in the images. Automated inspection has achieved an expected improvement owing to the rapid evolution of ML. ML algorithms can acquire feature representations and execute confidence intervals without requiring human model training, which is performed using conventional techniques [15]. Data collection, feature extraction, and categorization are all conventional machine learning algorithms for pavement crack detection. Shallow convolutional-neural-network-(CNN)-based architectures have been implemented to identify surface cracks and achieve greater accuracy with efficient computational costs [16,17]. Deep CNN systems use a multilayer neural net to retrieve significant characteristics from the input data [18]. Numerous analyses of ML-based crack detection methods have revealed that the classifier may not produce reliable results if the derived features do not identify actual cracks [19]. The Hessian matrix has been used both to accentuate cracks over blebs or staining and to modify the thickness fluctuation of cracks during image pre-processing [20]. Probabilistic relaxation is employed to determine cracks coarsely, eliminate noise, and accomplish adaptive thresholding for their operation [21]. Transfer learning methods allow CNNs to be used without incurring high computing costs or needing a prior understanding of the working functionality of CNN layers. Visual Geometry Group’s VGGNet [22], Microsoft’s ResNet [23], and Inception-V3 [24] are some of Google’s transfer learning design models that employ photographic data as input. Models built over the hybridization of ML algorithms have been found to increase the performance of vision-based systems over traditional algorithms. Hybridizing optimization techniques with support vector machines (SVM), such as fuzzy logic, k-nearest neighbours, artificial neural networks (ANN), and evolutionary algorithms, led to substantial improvements in recognition accuracy [25]. Pavement Crack Detection Net (PCDNet) eliminates more localized distortion, detects smaller cracks, and interprets information at a significantly faster rate than other methods [26]. Optimal intelligence technology was used to examine the wall quality parameters concerning many components of the dynamic situations of the retaining wall [27]. In addition, using ant colony optimization (ACO), these criteria resulted in the best design solution. A hybrid ANN-ACO algorithm utilizes various parameters under different structural conditions, and dynamic loads remarkably impact structural models [16]. Ensemble models built by stacking the best-performing ML algorithms have proven to produce efficient results in predictive models [28]. A pre-processed CNN classifier of VGG-16 with ResNet-50 detects fractures in image and inception models for object localization incorporated class actuation monitoring (CAM) for excellent optimal detection [29]. Optimizing the hyperparameters of CNN architectures, such as VGG16 and Resnet, proved to achieve greater accuracy in object identification and classification [30]. Unmanned aerial vehicles (UAVs) and camera systems were implemented to identify cracks and trigger U-Net in pixel-wise classification for feature and flaw identification with various feature sets. Texture-based video processing methods handle local binary patterns (LBP) using SVM and Bayes decision theories. For noisy and complicated bridge photographs, wavelet features were retrieved from the scene using a sliding-window-texture-analysis-based technique [31]. A deep CNN-based damage locating method used DenseNet to identify the damaged and undamaged steel frames from the images provided as inputs. The model outperformed the MobileNet and ResNet architectures [32]. Cracks are detected by the trained ConvNet, SVM, and boosting methods on a sliding window, where the SVM sometimes fails to distinguish the crack from the background. CNN using traditional Canny and Sobel edge detection methods can scan larger pixel images and exhibit excellent robustness during the training process. The AdaBoost classifier was used for pre-processing the crack image, and DL techniques were used for crack-detection in the image data [33]. An EfficientNet-based transfer learning model was developed to detect and classify surface cracks in high-rise buildings using UAV. Microcrack detection is achieved by solving the binary classification problem of cracks using autoencoders and softmax regression [34]. CNN-based order forensics framework for detecting picture operator chains is described. The two-stream Framework captures both local noise residual data and manipulating artifacts proof. The model may automatically detect alternated detection features straight from picture data and is suggested explicitly for forensically recognizing a chain consisting of two image operators [35]. To directly extract features from the photos, dual-filtering CNN base was designed. It treats each resampling parameter as a separate class, followed by the formulation of resampling parameter estimation and reformulates it as a multi-classification problem [36]. A reliable blind watermarking system based on 3D convolutional neural networks that can extract and integrate watermark images into animated GIF files was proposed [37].
Literature studies reviewed that manual inspection of structures is challenging, time-consuming and provides biased results. Several research studies have implemented different image-based methods, machine learning and deep learning algorithms to enable the automatic monitoring of building structures. The performance of ML algorithms on image data could have been more appreciated as it depends on many features and encompasses complex feature engineering tasks. The DCNN-based models were employed for damage detection and classifications, but not all models performed efficiently due to insufficient data, overfitting, and vanishing gradient problems. The models’ competency was enhanced by customising the convolutional layers, hybridising, ensembling, and transfer-learning techniques. The ResNet, DenseNet, VGG, and Xception models were observed to perform efficiently on detecting the structural damages such as concrete cracks, and steel bar damages.
To overcome the issues discussed in previous studies, this study proposes an automatic vision-based crack-identification system based on DL to identify crack portions from a large dataset of images acquired in the environment. The main breakthrough is the establishment of neural network-based classification models for various structural environments. Using various cameras and visual equipment, such as drones, this technology intends to ease routine inspections of concrete buildings and increase the speed of the diagnosis process of precise crack distribution while retaining accuracy. First, a series of images captured under a combination of structural, meteorological, and photographic circumstances were gathered, allowing for easy classification of the images using search keywords. A transfer learning approach was established to minimize time and money while constructing a DL model.

2. Research Methodology

The Conv2D ResNet exponential model was fitted with a dataset of images for each wall defect, such as cracks, holes, efflorescence, damp patches, and spalls. The dataset was collected through publicly available dataset repositories, such as the kaggle dataset or Structural Defects Network (SDNET) 2018 [38]. The model was trained with 80% of the 5000 images and tested with 20% of the images. The research methodology used in this study is shown in Figure 1. In stage 1, existing crack detection methods for building walls were explored and analysed. In Stage 2, a novel Conv2D ResNet exponential model was designed to detect the damage class of the building wall. The dataset consisted of a collection of wall images with different defects, such as cracks, holes, efflorescence, damp patches, and spalls. Training and testing of the proposed model were performed using an 80:20 wall quality dataset. Stage 3 evaluated the proposed model using the test data and compared Conv2D with existing models, such as DenseNet, VGG19, and Xception, of several activation layers, such as softplus, softsign, Relu, elu, and tanh. In stage 4, the performance of the proposed model was analysed using metrics such as precision, recall, F-score, and accuracy. The architecture of the DL activation layer is illustrated in Figure 2.
The workflow of the Conv2D ResNet exponential model is illustrated in Figure 3. An open access dataset (SDNET2018) published by Utah State University is used for implementation. In this proposed research work, the Snip&Sketch annotation tool was used for extracting the region of interest from the images available in the dataset. The Conv2D ResNet exponential model was trained with 4000 images and tested with 1000 images. The initial base model was pre-trained with initial weights using ImageNet, and Conv2D was designed with CNN models such as Xception, VGG19, DenseNet, and ResNet to extract the general high-level features. The base model was added to the custom layers, developed using a transfer-based DL approach, and the performance was analysed on wall quality prediction. The Xception, VGG19, DenseNet, and ResNet models were fitted with different activation layers, such as softplus, softsign, tanh, selu, elu, and exponential, and were evaluated using model loss, precision, accuracy, recall, and F-score measures.
As an overview of novelty, Conv2D ResNet exponential model was built over the Conv2D ResNet base model, along with the transfer learning custom activation layers that classify the wall defects more effectively with high accuracy. The wall quality dataset was fitted with the Conv2D ResNet model, which acted as a base model, thereby learning the types of wall defects from 90% of the training data. The acquired knowledge from Conv2D ResNet was transferred to refit the model by integrating the layer with the exponential activation function, which identified the wall defects in a single image, thereby validating the transfer learning by transporting the knowledge from the base model to the custom layer that enhanced the accuracy.

3. Implementation Setup

A dataset with 1000 images was collected for various defects, such as cracks, holes, efflorescence, damp patches, and spalls. Wall cracks signify trouble with the living area’s groundwork. Once wall cracks are detected in a residence, it generally means that the foundation is transitioning. Cracks in walls are caused by the contraction and relaxation of construction materials due to temperature and water content oscillations. A hole in a wall is also a significant defect that affects the quality of a building structure over time. Efflorescence is merely the deposition of salts on the surface of the aggregate, which is usually white. Once dry, the efflorescence consists of a white cover outside the concrete wall. When absorbed water and salts vaporize, they appear on the walls as crystallized patches or a layer of white powder. Damp patch condensation occurs when warm humid air within the same room is exposed to a cold interior wall or surface. Then, it quickly compresses the air back into the water. This evaporation then condenses on the interior wall surface, causing damp patches on the wall. A spall in the wall refers to the discoloration, clamping, fading, imploding, or flaking of concrete or brickwork, especially where the surface components have been destroyed. A spall can occur because of moisture absorption, combustion, or mechanical processes. The wall quality dataset contains 1000 images for each defect, such as holes, cracks, efflorescence, damp patches, and spalls, and is represented as follows in Equations (1)–(6):
W a l l = H 1000 + C 1000 + E 1000 + D 1000 + S 1000 ,
H 1000 = U h = 1 1000 { i = 1 255 j 1 255 H i j h } ,
C 1000 = U c = 1 1000 { i = 1 255 j 1 255 C i j c } ,
E 1000 = U e = 1 1000 { i = 1 255 j 1 255 E i j e } ,
D 1000 = U d = 1 1000 { i = 1 255 j 1 255 D i j d } ,
S 1000 = U s = 1 1000 { i = 1 255 j 1 255 S i j s } ,
where H i j h represents the wall images with holes, C i j c represents the wall images with cracks, E i j e represents the wall images with efflorescence, D i j d represents the wall images with damp patches, and S i j s represents the wall images with spall. The wall quality dataset images used for implementation are shown in Figure 4.
The wall quality dataset was pre-trained with the ImageNet by substituting the weights and trained with convolutional neural network models, such as Conv2D Xception, DenseNet, VGG19, and ResNet models, to extract the essential features from the image. Equation (7) represents the Gaussian function applied for feature extraction. The parameter “r” denotes the variance of the Gaussian function.
F e a t u r e ( i , j , r ) = 1 2 π r exp ( i 2 + j 2 ) 2 r 2 ,
The gaussian orientation function used for image filtering is shown in Equation (8).
O r i e n t ( i , j , r ,   θ ) = F e a t u r e i i c o s 2 ( θ ) + 2   F e a t u r e i j c o s ( θ )   s i n ( θ ) + F e a t u r e j j s i n 2 ( θ ) ,
where   F e a t u r e i i , F e a t u r e i j , and F e a t u r e j j   denote the second derivatives of the Gaussian function, as represented in Equations (9)–(11).
F e a t u r e i i   ( i ,   j ,   r ) = ( i 2 r 2 ) exp ( ( ( i 2 + j 2 ) ) 2 r 2 ) 2 π r 5 ,
F e a t u r e j j   ( i ,   j ,   r ) = ( j 2 r 2 ) exp ( ( ( i 2 + j 2 ) ) 2 r 2 ) 2 π r 5 ,
F e a t u r e i j   ( i ,   j ,   r ) = i j exp ( ( ( i 2 + j 2 ) ) 2 r 2 ) 2 π r 5 .
The input images in the wall quality dataset were processed with four-layer Conv2D layers—convolution filtering, sigmoid filter, linear transformation, and linear sigmoid—to generate the final output and are denoted in Equations (12)–(14).
Z = F e a t u r e i j F i l t e r ,
C o n [ R , C ] = F e a t u r e i j K e r n e l [ R , C ] ,
C o n [ R , C ] = j k k e r n e l [ j , k ]     F e a t u r e [ R j ] [ C k ] ,
where “R, C” represents the rows and columns of the input image matrix.
The input images in the wall quality dataset were trained with Conv2D and designed with CNN models, such as Xception, VGG19, DenseNet, and ResNet, to extract the general high level features, which are represented in Equations (15)–(18).
D e n s e N e t = ( 5   C P + ( T L ( 6 + 12 + 48 ) + C L ( 32 ) ) 2   D B )  
V G G 19 = ( 5   C P + ( C R M ( 2 + 3 + 4 + 5 ) + F C L ( 3 ) )  
X c e p t i o n = ( ( E n t r y ( 2   C P + 3 ( 2   S C + 1   M P ) ) ) + M i d d l e (   8 ( 3 R e l u + 3 S C ) )     +   E x i t ( ( 1 C P + 2 R e l u + 2 S C   ) + 1 F C L )  
R e s N e t = ( 62   C P + ( C L ( 2 + 4 + 8 + 16 ) ) 3   P L )  
where CP, TL, CL, DB,   C R M , FCL, SC, MP, and PL represent the convolution and pooling layer, transition layer, classification layer, dense block layer, convolution relu max pooling layer, fully connected layer, separable convolution layer, max pooling layer, and normal pooling layer, respectively.
Apply the sigmoid function to the above equation, as shown in (19) and (20).
A S i g = S i g m o i d ( Z ) ,
S i g m o i d ( Z ) = 1 ( 1 + e Z ) .
The linear transformation was applied to the above layer to process the third layer of the CNN, as in (17).
Z L i n e a r = W e i g h t T × A S i g + B i a s .
The final output is given by applying a linear sigmoid, as given in (18).
O u t p u t = S i g m o i d ( Z L i n e a r ) .
The proposed Conv2D ResNet exponential model is built with exponential activation layer and is denoted in (23).
C o n v 2 D R e s N e t e x p o n e n t i a l m o d e l = E x p o n e n t i a l { R e s n e t ( O u t p u t ) }  

4. Prescriptive and Predictive Data Analysis of Wall Quality Defects

The pre-training of the wall quality dataset was performed with the ImageNet by substituting the weights and trained with CNN models, such as Conv2D Xception, DenseNet, VGG19, and ResNet models, for extracting the features. The designed base model was fitted to custom layers with various activation layers, such as softplus, softsign, tanh, selu, elu, and exponential, to analyse the performance of identifying the defects in the wall quality. Activation functions play an essential role in developing neural networks. The activation function determines the frequency with which the network structure acquires a training dataset. The activation function at the output layer determines a model’s prediction. An activation function is a unit placed at the end or middle of a neural network that determines whether a neuron will be activated. The activation function is a complex nonlinear transformation applied to an input signal. The signal is subsequently processed and provided as an input to the next layer of neurons. Softsign is the activation function of neural networks. Equation (24) provides the mathematical notation for the softsign activation function.
s o f t s i g n ( x ) = ( x ) ( 1 + | x | ) .
The softplus function is a soft equivalent of the ReLU activation function, occasionally used instead of the ReLU in neural networks. Softplus is related to the sigmoid function and is represented by the following mathematical Equation (25).
s o f t p l u s ( x ) = l o g ( 1 + e x ) .
The tanh activation function is a hyperbolic tangent activation function that reflects the sigmoid activation function. The tanh function takes the input as any real value and outputs the value from −1 to 1. It is represented by the following mathematical Equation (26).
t a n h ( x ) = ( e x e x ) ( e x + e x ) .
The scaled exponential linear unit (selu) activation function is implicitly induced with normalization properties. Normalization with selu is conducted as the input value x is less than zero, and the output is the product of x and lambda. When x is zero, the output is equal to 0. If x is less than zero, then the output is the product of lambda and alpha by the x-value minus the alpha value’s exponential, multiplied by the lambda value. Equation (27) represents the selu activation function.
s e l u ( x ) =   λ { x                                     i f   x > 0   α e x α               i f   x 0 .
The exponential linear unit (Elu) activation function mainly focuses on the positive values, and the alpha value is selected from 0.1 to 0.3, as represented in Equation (28).
e l u ( x ) =   λ { x                                         i f   x > 0   α ( e x 1 )             i f   x < 0 .
The exponential activation function indicates the positive-valued function of a real input variable and is represented by Equation (29).
e x p ( x ) = 1 + x + x 2 2 ! + x 3 3 ! + + x n n ! .
The model is examined with the activation function discussed above and compiled to analyse performance indices such as step loss, accuracy, validation loss, and validation accuracy. The loss function computes the difference between the actual output of the model and the target outcome. Step loss indicates the loss incurred at each iteration. The validation loss specifies the loss for the validation dataset, which is part of the training dataset. The wall quality falls under the classification problem, and the loss function is given by Equation (30).
L o s s = i = 1 n | y a c t u a l y t a r g e t | .
Accuracy is the ratio of the number of correct predictions to the total number of predictions and is denoted by Equation (31).
A c c u r a c y = T r u e   P o s i t i v e s + T r u e   N e g a t i v e s T r u e   P o s i t i v e s + T r u e   N e g a t i v e s + F a l s e   P o s i t i v e s + F a l s e   N e g a t i v e s .
The model was compiled with ten epochs, and prescriptive data analysis was performed by analysing the performance indices for each epoch, as shown in Table 1, Table 2, Table 3 and Table 4.
The wall quality dataset was pre-trained with the images from ImageNet by substituting the weights and then trained with convolutional neural network models such as Conv2D Xception, DenseNet, VGG19, and ResNet models to extract the essential features from the image. The designed base model was fitted to custom layers with various activation layers, such as softplus, softsign, tanh, selu, elu, and exponential, to analyse the performance defect identification in the wall quality. The performance was analysed in terms of model loss, model accuracy, precision, recall, and F-score, as shown in Table 5, Table 6, Table 7 and Table 8 and Figure 5, Figure 6 and Figure 7. It was observed that the performance of ResNet was comparatively better, with greater accuracy and F-score values, followed by DenseNet, Xception, and VGG19 models. The model performance across different activation layers was studied. It was noted that the models’ performance metrics were better with the exponential activation layer than with the other layers. Thus, the Conv2D ResNet model implemented with the exponential activation layer provided an enhanced model for wall quality detection.
The performance of the Conv2D Xception, DenseNet, VGG19, and ResNet models using the exponential activation layer was studied using learning curves during the training and validation phases. Figure 5 shows the accuracy and loss curves of all models over each epoch during training phase. The results revealed that accuracy gradually increased over epochs and stabilized after a threshold. The learning loss of the Xception, DenseNet, and ResNet models was more significant during early epochs and dropped as the epoch increased. Figure 6 and Figure 7 showcase the learning curves of the models during the validation phase. It is noted the performance of ResNet exponential model had a consistent increase in accuracy and decrease in loss at each epoch. The other models were found to have declining accuracy at certain epochs.
The wall quality dataset was fitted with the Conv2D ResNet model, which acted as a base model, thereby learning the types of wall defects from 90% of the training data. Now the model had learnt the exact region of interest through which it categorized the wall defect class as cracks, holes, efflorescence, damp patches, or spalls from the more significant number of images. The acquired knowledge from Conv2D ResNet was transferred to refit the model by integrating the layer with the exponential activation function, which identified the wall defects of a single image, thereby validating the transfer learning. When the test image was fitted with the Conv2D ResNet exponential model, the given image was validated with the wall defect class of the actual value and predicted value of the model, which was termed as the True and Guess label, respectively shown in Figure 8, Figure 9, Figure 10 and Figure 11.
Implementation results portrayed that the proposed Conv2D ResNet model with exponential activation layer outperforms with an accuracy of 0.9147, precision of 0.9212, recall of 0.9134, and F-Score of 0.9978 compared with other Conv2D models, such as Xception, VGG19, and DenseNet. The comparative study of learning accuracy during the training and validation phases of the implemented models is represented in Figure 12a,b. The results illustrated the accuracy of each model at every epoch, and it is noticeable that the proposed Conv2D ResNet exponential model steadily gained accuracy at every epoch, whereas the VGG19 model struggled for accuracy during both phases. The DenseNet and Xception models showed instabilities with epochs. The evaluation loss attained by the models is depicted in Figure 12c, which shows that the Conv2D ResNet exponential model had minimal losses at each epoch compared with the DenseNet, Xception and VGG19 models in that order. The Conv2D ResNet exponential model performed efficiently, producing a greater accuracy with minimal loss in wall quality detection.

5. Conclusions

This study attempted to analyse the performance of DL models for evaluating the quality of wall structures. This work’s main contribution is designing the Conv2D ResNet exponential model-based architecture that classifies wall defects, such as cracks, holes, efflorescence, damp patches, and spalls. A dataset with 5000 images was used to train the proposed model, which achieved the requirements of this research work and outperformed the other Conv2D models. The Conv2D ResNet model with 48 convolution layers, one maxpool, and an average pool layer was implemented in this study. This model served as the base and integrated with the exponential activation layer, improving the classifier’s performance in detecting wall defects. The proposed Conv2D ResNet exponential model was further investigated using the performance metrics precision, recall, F-score, and accuracy. The Conv2D ResNet exponential model classified the wall defect type through transfer learning and was also used to analyse the performance of the other CNN model with several activation layers. The wall quality dataset was fitted with the Conv2D ResNet model, which acted as a base model, thereby learning the types of wall defects from 90% of the training data. The acquired knowledge from Conv2D ResNet was transferred to refit the model by integrating the layer with the exponential activation function, which identified the wall defects of a single image, thereby validating the transfer learning by transferring the knowledge from the base model to the custom layer that enhances the accuracy.
This research provides a proper fitting of residual networks to reduce the loss, thereby improving the accuracy, of wall classification with other Conv2D models. The performances of the Xception, VGG19, DenseNet, and ResNet models were fitted with different activation layers such as softplus, softsign, tanh, selu, elu, and exponential, along with transfer learning and analysed using performance evaluation metrics. The dataset used for the proposed Conv2D ResNet exponential model can be used for classifying the defect type in the wall. The same dataset could also be used to identify the defect’s depth through object detection methods. However, categorizing the class of defects, such as cracks, holes, efflorescence, damp patches, and spalls in the walls directly related to the characteristics of the wall quality available in the dataset. Once the wall defect class is identified, the respective maintenance procedure can easily be conducted. Implementation results portrayed that the proposed Conv2D ResNet model with exponential activation layer outperforms with an F-Score of 0.997826 compared with other Conv2D models, such as Xception, VGG19, and DenseNet. The potential findings of the proposed Conv2D ResNet exponential model are identifying the appropriate activation layer function that provides the highest accuracy in predicting the type of wall defect. The proposed Conv2D ResNet exponential model improved the overall effectiveness of classifying the wall defects compared with the other deep learning techniques. As an overview of novelty, the Conv2D ResNet exponential model was built over the Conv2D ResNet base model and extended the transfer learning using custom activation layers that effectively classify the wall defects with high accuracy. Despite the Conv2D ResNet exponential model’s impressive performance, it is still challenging for researchers to fine-tune the base model hyper-parameters by integrating them with various optimizers. This research’s future enhancement focused on validating the accuracy of the wall defect prediction for various combinations of convolutional layers and probabilistic loss functions.

Author Contributions

Conceptualization, B.K. and Y.N.; methodology, K.R.S.P.; software, S.D.M. and K.R.S.P.; validation, B.K., Y.N. and D.-E.L.; formal analysis, A.R. and G.W.; investigation, S.D.M.; resources, A.R. and G.W.; data curation, S.D.M.; writing—original draft preparation, A.R. and K.R.S.P.; writing—review and editing, K.R.S.P. and D.-E.L.; visualization, Y.N. and K.R.S.P.; supervision, Y.N.; project administration, B.K. and D.-E.L.; funding acquisition, B.K. and K.R.S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Brain Pool program funded by the Ministry of Science and ICT through the National Research Foundation of Korea (NRF-2022H1D3A2A02082296). This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No.2021R1I1A1A01044447).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lacidogna, G.; Piana, G.; Accornero, F.; Carpinteri, A. Multi-technique damage monitoring of concrete beams: Acoustic Emission, Digital Image Correlation, Dynamic Identification. Constr. Build. Mater. 2020, 242, 118114. [Google Scholar] [CrossRef]
  2. Liu, P.; Lim, H.; Yang, S.; Sohn, H. Development of a “stick-and-detect” wireless sensor node for fatigue crack detection. Struct. Health Monit. 2016, 16, 153–163. [Google Scholar] [CrossRef]
  3. Zhao, S.; Sun, L.; Gao, J.; Wang, J. Uniaxial ACFM detection system for metal crack size estimation using magnetic signature waveform analysis. Measurement 2020, 164, 108090. [Google Scholar] [CrossRef]
  4. Yang, X.; Zhou, Z. Design of crack detection system. In Proceedings of the 2017 International Conference on Network and Information Systems for Computers, Shanghai, China, 14–16 April 2017; pp. 147–152. [Google Scholar] [CrossRef]
  5. Zhang, X.; Wang, K.; Wang, Y.; Shen, Y.; Hu, H. Rail crack detection using acoustic emission technique by joint optimization noise clustering and time window feature detection. Appl. Acoust. 2020, 160, 107141. [Google Scholar] [CrossRef]
  6. Amhaz, R.; Chambon, S.; Jerome, I. Automatic Crack Detection on Two-Dimensional Pavement Images: An Algorithm Based on Minimal Path Selection. Trans. Intell. Transp. Syst. 2016, 17, 2718–2729. [Google Scholar] [CrossRef] [Green Version]
  7. Amhaz, R.; Chambon, S.; Jerome, I.; Baltazart, V. A new minimal path selection algorithm for automatic crack detection on pavement images. In Proceedings of the 2014 International Conference on Image Processing, Paris, France, 27–30 January 2014. [Google Scholar] [CrossRef] [Green Version]
  8. Albishi, A.; Ramahi, O.M. Detection of Surface and Subsurface Cracks in Metallic and Non-Metallic Materials Using a Complementary Split-Ring Resonator. Sensors 2014, 14, 19354–19370. [Google Scholar] [CrossRef] [Green Version]
  9. Cheon, M.H.; Hong, D.G.; Lee, D. Surface Crack Detection in Concrete Structures Using Image Processing. In Advances in Intelligent Systems and Computing, Proceedings of the International Conference on Robot Intelligence Technology and Applications 5, Daejeon, Korea, 14–15 December 2017; Springer: Cham, Switzerland, 2019; p. 751. [Google Scholar] [CrossRef]
  10. Tedeschi, A.; Benedetto, F. A real-time automatic pavement crack and pothole recognition system for mobile Android-based devices. Adv. Eng. Inform. 2017, 32, 11–25. [Google Scholar] [CrossRef]
  11. Sun, H.; Liu, Q.; Fang, L. Research on Fatigue Crack Growth Detection of M(T) Specimen Based on Image Processing Technology. J. Fail. Anal. Prev. 2018, 18, 1010–1016. [Google Scholar] [CrossRef]
  12. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Hu, G.; Lee, D.-E. Wind-Induced Pressure Prediction on Tall Buildings Using Generative Adversarial Imputation Network. Sensors 2021, 21, 2515. [Google Scholar] [CrossRef]
  13. Wang, Y.; Huang, Y.; Huang, W. Crack junction detection in pavement image using correlation structure analysis and iterative tensor voting. IEEE Access 2019, 7, 138094–138109. [Google Scholar] [CrossRef]
  14. Li, W.; Ju, H.; Susan, L.; Ren, Q. Three-dimensional pavement crack detection algorithm based on two-dimensional empirical mode decomposition. J. Transp. Eng. Part B Pavements 2017, 143, 2573–5438. [Google Scholar] [CrossRef]
  15. Kim, B.; Lee, D.B.; Sri Preethaa, K.R.; Hu, G.; Natarajan, Y.; Kwok, K.C.S. Predicting wind flow around buildings using deep learning. J. Wind. Eng. Ind. Aerodyn. 2021, 219, 104820. [Google Scholar] [CrossRef]
  16. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Arun Pandian, R. Surface crack detection using deep learning with shallow CNN architecture for enhanced computation. Neural Comput. Appl. 2022, 33, 9289–9305. [Google Scholar] [CrossRef]
  17. Perez, H.; Tah, J.H.M.; Mosavi, A. Deep Learning for Detecting Building Defects Using Convolutional Neural Networks. Sensors 2019, 19, 3556. [Google Scholar] [CrossRef] [Green Version]
  18. Zhou, T. Analysis on Construction Quality Control Technology of Reinforced Concrete Shear Wall Structure. Front. Res. Archit. Eng. 2018, 1, 117–121. [Google Scholar] [CrossRef]
  19. Spencer, B.F., Jr.; Hoskerea, V.; Narazaki, Y. Advances in Computer Vision-Based Civil Infrastructure Inspection and Monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  20. Kumar, P.; Batchu, S.; Swamy, S.N.; Kota, S.R. Real-Time Concrete Damage Detection Using Deep Learning for High Rise Structures. IEEE Access 2021, 9, 112312–112331. [Google Scholar] [CrossRef]
  21. Hoang, N.D. Detection of Surface Crack in Building Structure Using Image Processing Technique with an Improved Otsu Method for Image Thresholding. Adv. Civ. Eng. 2018, 2018, 1–15. [Google Scholar] [CrossRef] [Green Version]
  22. Lee, B.Y.; Kim, Y.Y.; Yi, S.-T.; Kim, J.-K. Automated image processing technique for detecting and analysing concrete surface cracks. Struct. Infrastruct. Eng. 2016, 9, 567–577. [Google Scholar] [CrossRef]
  23. Alam, S.Y.; Loukili, A.; Grondin, F.; Rozière, E. Use of the digital image correlation and acoustic emission technique to study the effect of structural size on cracking of reinforced concrete. Eng. Fract. Mech. 2015, 143, 17–31. [Google Scholar] [CrossRef]
  24. Sri Preethaa, K.R.; Sabari, A. Intelligent video analysis for enhanced pedestrian detection by hybrid metaheuristic approach. Soft Comput. 2020, 24, 12303–12311. [Google Scholar] [CrossRef]
  25. Talab, A.M.A.; Huang, Z.; Xi, F.; Hai Ming, L. Detection crack in image using Otsu method and multiple filtering in image processing techniques. Opt.—Int. J. Light Electron Opt. 2016, 127, 1030–1033. [Google Scholar] [CrossRef]
  26. Mohan, A.; Poobal, S. Crack detection using image processing: A critical review and analysis. Alex. Eng. J. 2018, 57, 787–798. [Google Scholar] [CrossRef]
  27. Xu, C.; Gordan, B.; Koopialipoor, M.; Armaghani, D.J.; Tahir, M.M.; Zhang, X. Improving Performance of Retaining Walls Under Dynamic Conditions Developing an Optimized ANN Based on Ant Colony Optimization Technique. IEEE Access 2019, 7, 94692–94700. [Google Scholar] [CrossRef]
  28. Kim, B.; Lee, D.-E.; Hu, G.; Natarajan, Y.; Preethaa, S.; Rathinakumar, A.P. Ensemble Machine Learning-Based Approach for Predicting of FRP–Concrete Interfacial Bonding. Mathematics 2022, 10, 231. [Google Scholar] [CrossRef]
  29. Yamaguchi, T.; Hashimoto, S. Fast crack detection method for large-size concrete surface images using percolation-based image processing. Mach. Vis. Appl. 2010, 21, 787–809. [Google Scholar] [CrossRef]
  30. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Santhosh, R.; Sabari, A. Enhanced pedestrian detection using optimized deep convolution neural network for smart building surveillance. Soft Comput. 2020, 24, 17081–17092. [Google Scholar] [CrossRef]
  31. Sharma, S.; Gupta, N.K. A Genetic Approach to Segment and Detect Crack in Rail Track. In Proceedings of the 2019 International Conference on Computing Methodologies and Communication, Erode, India, 27–29 March 2019. [Google Scholar]
  32. Kim, B.; Yuvaraj, N.; Park, H.W.; Sri Preethaa, K.R.; Pandian, R.A.; Lee, D.E. Investigation of steel frame damage based on computer vision and deep learning. Autom. Constr. 2021, 132, 103941. [Google Scholar] [CrossRef]
  33. Yusof, N.; Osman, M.; Hussain, Z.; Noor, M. Automated Asphalt Pavement Crack Detection and Classification using Deep Convolution Neural Network. In Proceedings of the 2019 International Conference on Control System, Penang, Malaysia, 29 November–1 December 2019. [Google Scholar]
  34. Yuvaraj, N.; Kim, B.; Sri Preethaa, K.R. Transfer Learning Based Real-Time Crack Detection Using Unmanned Aerial System. Int. J. High-Rise Build. 2020, 9, 351–360. [Google Scholar] [CrossRef]
  35. Liao, X.; Li, K.; Zhu, X.; Liu, K.J.R. Robust Detection of Image Operator Chain with Two-Stream Convolutional Neural Network. IEEE J. Sel. Top. Signal Process. 2020, 14, 955–968. [Google Scholar] [CrossRef]
  36. Peng, L.; Liao, X.; Chen, M. Resampling parameter estimation via dual-filtering based convolutional neural network. Multimed. Syst. 2021, 27, 363–370. [Google Scholar] [CrossRef]
  37. Liao, X.; Peng, J.; Cao, Y. GIFMarking: The robust watermarking for animated GIF based deep learning. J. Vis. Commun. Image Represent. 2021, 79, 103244. [Google Scholar] [CrossRef]
  38. Utah State University, University Libraries Home Page. Available online: https://digitalcommons.usu.edu/all_datasets/48/ (accessed on 2 July 2022).
Figure 1. Research methodology flow.
Figure 1. Research methodology flow.
Mathematics 10 04602 g001
Figure 2. Deep learning activation layer model architecture.
Figure 2. Deep learning activation layer model architecture.
Mathematics 10 04602 g002
Figure 3. Workflow of Conv2D ResNet exponential model.
Figure 3. Workflow of Conv2D ResNet exponential model.
Mathematics 10 04602 g003
Figure 4. Wall quality dataset with cracks, holes, efflorescence, damp patches, and spall defects.
Figure 4. Wall quality dataset with cracks, holes, efflorescence, damp patches, and spall defects.
Mathematics 10 04602 g004
Figure 5. Model loss and accuracy of Conv2D models during training. (a) Xception; (b) VGG19; (c) DenseNet; (d) ResNet.
Figure 5. Model loss and accuracy of Conv2D models during training. (a) Xception; (b) VGG19; (c) DenseNet; (d) ResNet.
Mathematics 10 04602 g005aMathematics 10 04602 g005b
Figure 6. Model accuracy and loss of Conv2D models during cross validation. (a) Model accuracy of Xception; (b) model loss of Xception; (c) model accuracy of VGG19; (d) model loss of VGG19.
Figure 6. Model accuracy and loss of Conv2D models during cross validation. (a) Model accuracy of Xception; (b) model loss of Xception; (c) model accuracy of VGG19; (d) model loss of VGG19.
Mathematics 10 04602 g006
Figure 7. Model loss and accuracy of Conv2D models during cross validation. (a) Model accuracy of DenseNet; (b) nodel loss of DenseNet; (c) model accuracy of ResNet; (d) model loss of ResNet.
Figure 7. Model loss and accuracy of Conv2D models during cross validation. (a) Model accuracy of DenseNet; (b) nodel loss of DenseNet; (c) model accuracy of ResNet; (d) model loss of ResNet.
Mathematics 10 04602 g007
Figure 8. Prediction of wall defects by Conv2D Xception model.
Figure 8. Prediction of wall defects by Conv2D Xception model.
Mathematics 10 04602 g008aMathematics 10 04602 g008b
Figure 9. Prediction of wall defects by Conv2D VGG19 base model.
Figure 9. Prediction of wall defects by Conv2D VGG19 base model.
Mathematics 10 04602 g009
Figure 10. Prediction of wall defects by Conv2D DenseNet base model.
Figure 10. Prediction of wall defects by Conv2D DenseNet base model.
Mathematics 10 04602 g010
Figure 11. Prediction of wall defects by Conv2D ResNet base model.
Figure 11. Prediction of wall defects by Conv2D ResNet base model.
Mathematics 10 04602 g011aMathematics 10 04602 g011b
Figure 12. Learning curve comparisons of Conv2D models. (a) Training accuracy; (b) validation accuracy; (c) evaluation loss.
Figure 12. Learning curve comparisons of Conv2D models. (a) Training accuracy; (b) validation accuracy; (c) evaluation loss.
Mathematics 10 04602 g012aMathematics 10 04602 g012b
Table 1. Prescriptive data analysis for Conv2D Xception model compilation.
Table 1. Prescriptive data analysis for Conv2D Xception model compilation.
EpochsStep LossAccuracyValidation LossValidation Accuracy
1/101.43660.43190.89060.6316
2/100.72440.73010.66960.7474
3/100.39340.85600.80880.7895
4/100.24610.91520.86540.7789
5/100.16430.95630.65660.7780
6/100.14950.96740.67660.8001
7/100.15720.95890.54160.8632
8/100.06990.97940.75790.8211
9/100.06360.97880.73900.8632
10/100.06910.98201.05620.8316
Table 2. Prescriptive data analysis for Conv2D DenseNet model compilation.
Table 2. Prescriptive data analysis for Conv2D DenseNet model compilation.
EpochsStep LossAccuracyValidation LossValidation Accuracy
1/101.45460.42190.88060.6234
2/100.74530.72010.67660.7572
3/100.39440.83500.80880.7895
4/100.34410.92620.85640.7569
5/100.17830.96730.65560.7770
6/100.14850.97740.64360.8341
7/100.16720.96890.56660.8772
8/100.07590.98840.75590.8311
9/100.08660.98980.72800.8732
10/100.05610.99201.04620.8446
Table 3. Prescriptive data analysis for Conv2D VGG19 model compilation.
Table 3. Prescriptive data analysis for Conv2D VGG19 model compilation.
EpochsStep LossAccuracyValidation LossValidation Accuracy
1/101.45560.42110.87890.6324
2/100.74630.74440.67330.7452
3/100.39540.83440.81120.7555
4/100.34610.95670.83450.7669
5/100.17440.95630.64350.7760
6/100.14650.97860.64450.8871
7/100.16780.98780.53450.8342
8/100.07660.98760.76540.8321
9/100.08820.98990.73420.8652
10/100.05550.99431.04350.8896
Table 4. Prescriptive data analysis for Conv2D ResNet model compilation.
Table 4. Prescriptive data analysis for Conv2D ResNet model compilation.
EpochsStep LossAccuracyValidation LossValidation Accuracy
1/101.54460.23190.88060.6834
2/100.45530.57010.67660.7872
3/100.94440.63500.80880.7995
4/100.43410.88620.85640.7579
5/100.72830.86730.65560.7870
6/100.42850.87740.64360.8221
7/100.63720.86890.56660.8552
8/100.73590.93840.75590.8321
9/100.02660.97980.72800.8432
10/100.03610.98801.04620.8226
Table 5. Performance analysis for Conv2D VGG19 model with various activation layers.
Table 5. Performance analysis for Conv2D VGG19 model with various activation layers.
Conv2D VGG19 ModelPrecisionRecallF-ScoreAccuracy
Conv2D VGG19—softsign0.81010.81790.81270.8179
Conv2D VGG19—softplus0.81960.80680.81570.8169
Conv2D VGG19—selu0.82550.83500.82110.8350
Conv2D VGG19—elu0.84960.83680.83570.8469
Conv2D VGG19—tanh0.86960.86840.85680.8690
Conv2D VGG19—exponential0.89610.88380.88680.8990
Table 6. Performance analysis for Conv2D Xception model with various activation layers.
Table 6. Performance analysis for Conv2D Xception model with various activation layers.
Conv2D Xception ModelPrecisionRecallF-ScoreAccuracy
Conv2D Xception—softsign0.86340.86560.86570.8689
Conv2D Xception—softplus0.84660.84890.84780.8499
Conv2D Xception—selu0.85660.85590.85290.8550
Conv2D Xception—elu0.94060.93680.93570.9469
Conv2D Xception—tanh0.91230.91350.91680.9190
Conv2D Xception—exponential0.90710.90380.90560.9090
Table 7. Performance analysis for Conv2D DenseNet model with various activation layers.
Table 7. Performance analysis for Conv2D DenseNet model with various activation layers.
Conv2D DenseNet ModelPrecisionRecallF-ScoreAccuracy
Conv2D DenseNet—softsign0.88340.87560.87570.8889
Conv2D DenseNet—softplus0.85660.84890.84780.8499
Conv2D DenseNet—selu0.85660.85690.85790.8579
Conv2D DenseNet—elu0.94790.93790.92570.9369
Conv2D DenseNet—tanh0.92350.91880.91570.9299
Conv2D DenseNet—exponential0.91230.91230.91540.9157
Table 8. Performance analysis for Conv2D ResNet model with various activation layers.
Table 8. Performance analysis for Conv2D ResNet model with various activation layers.
Conv2D ResNet ModelPrecisionRecallF-ScoreAccuracy
Conv2D ResNet—softsign0.89570.88890.88880.8949
Conv2D ResNet—softplus0.86610.87870.87880.8799
Conv2D ResNet—selu0.85860.85690.85790.8579
Conv2D ResNet—elu0.93480.93990.92570.9379
Conv2D ResNet—tanh0.91230.91660.91570.9249
Conv2D ResNet—exponential0.92120.91340.99780.9147
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, B.; Natarajan, Y.; Munisamy, S.D.; Rajendran, A.; Sri Preethaa, K.R.; Lee, D.-E.; Wadhwa, G. Deep Learning Activation Layer-Based Wall Quality Recognition Using Conv2D ResNet Exponential Transfer Learning Model. Mathematics 2022, 10, 4602. https://doi.org/10.3390/math10234602

AMA Style

Kim B, Natarajan Y, Munisamy SD, Rajendran A, Sri Preethaa KR, Lee D-E, Wadhwa G. Deep Learning Activation Layer-Based Wall Quality Recognition Using Conv2D ResNet Exponential Transfer Learning Model. Mathematics. 2022; 10(23):4602. https://doi.org/10.3390/math10234602

Chicago/Turabian Style

Kim, Bubryur, Yuvaraj Natarajan, Shyamala Devi Munisamy, Aruna Rajendran, K. R. Sri Preethaa, Dong-Eun Lee, and Gitanjali Wadhwa. 2022. "Deep Learning Activation Layer-Based Wall Quality Recognition Using Conv2D ResNet Exponential Transfer Learning Model" Mathematics 10, no. 23: 4602. https://doi.org/10.3390/math10234602

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop