Next Article in Journal
Using Drones to Estimate and Reduce the Risk of Wildfire Propagation in Wildland–Urban Interfaces
Previous Article in Journal
TC-Verifier: Trans-Compiler-Based Code Translator Verifier with Model-Checking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Learning Framework for Enhancing Bridge Damage Prediction

by
Amal Abdulbaqi Maryoosh
1,
Saeid Pashazadeh
2,* and
Pedram Salehpour
1
1
Department of Computer Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz 51666-16471, Iran
2
Department of Information Technology, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz 51666-16471, Iran
*
Author to whom correspondence should be addressed.
Appl. Syst. Innov. 2025, 8(3), 61; https://doi.org/10.3390/asi8030061
Submission received: 25 March 2025 / Revised: 25 April 2025 / Accepted: 27 April 2025 / Published: 30 April 2025

Abstract

:
Bridges are crucial structures for transportation networks, and their structural integrity is paramount. Deterioration and damage to bridges can lead to significant economic losses, traffic disruptions, and, in severe cases, loss of life. Traditional methods of bridge damage detection, often relying on visual inspections, can be challenging or impossible in critical areas such as roofing, corners, and heights. Therefore, there is a pressing need for automated and accurate techniques for bridge damage detection. This study aims to propose a novel method for bridge crack detection that leverages a hybrid supervised and unsupervised learning strategy. The proposed approach combines pixel-based feature method local binary pattern (LBP) with the mid-level feature bag of visual words (BoVW) for feature extraction, followed by the Apriori algorithm for dimensionality reduction and optimal feature selection. The selected features are then trained using the MobileNet model. The proposed model demonstrates exceptional performance, achieving accuracy rates ranging from 98.27% to 100%, with error rates between 1.73% and 0% across multiple bridge damage datasets. This study contributes a reliable hybrid learning framework for minimizing error rates in bridge damage detection, showcasing the potential of combining LBP–BoVW features with MobileNet for image-based classification tasks.

1. Introduction

Bridge infrastructure is critical to our transportation networks, and ensuring bridges’ structural integrity is primary. Deterioration and damage to bridges can lead to significant economic losses, traffic disruptions, and, in severe cases, loss of life because they are susceptible to various forms of damage, such as cracks, corrosion, and erosion, which can compromise their safety and functionality. Early detection of this damage is crucial for timely maintenance and repairs, preventing catastrophic failures, and minimizing economic losses [1,2].
In the past decade, several industries have employed artificial intelligence (AI) algorithms to acquire information and insights, thus enhancing their performance. The concept of AI is to provide machines with the ability to think and act like human beings. It contains supervised and unsupervised learning strategies. Artificial intelligence methodologies, including diagnosis, prediction, clustering, and classification, are used to develop models that distinguish between various data classes [3]. These models are employed to predict categorical class labels, such as “background” or “defects” and “crack” or “non-crack” in the classification of bridge damage. These techniques are employed to transform data into knowledge, enabling specialists to improve their work. This knowledge is represented as models. A model exhibiting high-performance metrics indicates that the artificial intelligence methodology effectively extracted knowledge and distinguished between various data patterns. However, the extracted knowledge may not always provide the correct predictions [4].
In the context of feature extraction based on a handcrafted method, Cao et al. [5] trained CNN descriptors using BoVW, while Hassan et al. [6] utilized color moments as a global feature and LBP as a local feature to construct a BoVW. Additionally, various studies [7,8,9,10] utilize BoVW for image classification and recognition. Many methods have been used to modify BoVW; for instance, an improved BoVW (iBoVW) has been proposed to generate the features from each convolutional layer of the pre-trained CNN model [11]. Also, a BoVW model presented by Olaode and Naghdy [12], whereby deep feature learning via a stacked autoencoder was used to extract image features and for video-based recognition, spatiotemporal scale-coded BoW (SC-BoW) was utilized. It entails splitting spatiotemporal characteristics into sub-groups according to the geographic scale from which they were taken and then encoding the acquired multi-scale information into BoW representations [13]. The SO-BoVW model proposed by Sultani and Dhannoon [14] combined the scale invariant feature transform (SIFT) and oriented FAST and rotated BRIEF (ORB) based on the BoVW model for image classification, where k-nearest neighbor (kNN) was utilized for image classification.
Based on our analysis of concrete damage detection models mentioned in previous literature, we have seen that some studies utilize deep learning models while others employ machine learning approaches. Some studies combine image processing techniques with either deep learning or machine learning. Bhalaji Kharthik et al. [15] employed twelve transfer learning DCNN models to categorize fractures in three publicly available datasets: CCIC, SDNET for structural cracks, and BCD for bridge fissures. The researchers applied two image enhancement techniques, LBP and contrast enhancement, to the SDNET images owing to the constrained efficacy of the transfer learning DCNN models. Furthermore, the support vector machine (SVM) was trained utilizing deep features derived from the final fully connected layer of the DCNNs. The integration of deep features with SVM resulted in an enhancement in detection accuracy across all combinations of DCNN and datasets. Yang et al. [16] proposed the utilization of the DCNN framework to classify fissures across three publicly accessible datasets: CCIC, BCD, and SDNET. The proposed methodology facilitates the transfer of three distinct forms of knowledge derived from prior scholarly achievements: sample knowledge, parameter knowledge, and model knowledge. The VGG network was augmented with additional fully connected layers, thereby establishing a novel learning framework. The researchers substantiated the credibility and efficacy of the proposed methodology. Zoubir et al. [17] employed a histogram of oriented gradients (HOG) in conjunction with uniform local binary patterns (ULBPs) to derive features from a dataset comprising over 3000 images, both uncracked and cracked, which encompassed various crack patterns and representations of concrete surfaces. Nonlinear dimensionality reduction techniques were executed through kernel principal component analysis (KPCA), and three machine learning classifiers were utilized to facilitate the classification process. The findings of the experiments indicate that the classification approach predicated on the SVM model and the feature-level fusion of HOG and ULBP features subsequent to the application of KPCA yielded optimal results, as the proposed classification framework accomplished an accuracy rate of 99.26%. Zoubir, with another team [18], introduced a publicly accessible benchmark-annotated image dataset comprising over 6900 images depicting cracked and non-cracked concrete bridges and culverts. The dataset presented encompasses a variety of challenging surface conditions and comprises concrete cracks exhibiting diverse sizes and patterns. The authors analyzed the proposed dataset employing three cutting-edge DCNNs, namely VGG16, VGG19, and InceptionV3, utilizing a transfer learning approach. The models above were employed to differentiate between cracked and non-cracked images, achieving a maximum testing accuracy of 95.89%. The outcomes of the experiments illustrate the promising applicability of this dataset for training deep learning networks aimed at the recognition of concrete cracks in bridge structures. Xu et al. [19] created a convolutional neural network (CNN)-based model for crack detection that works completely from the beginning to the end. It does this by using atrous convolution, the atrous spatial pyramid pooling (ASPP) module, and depthwise separable convolution. An implementation of atrous convolution makes it easier to obtain a wider receptive field while maintaining the input data’s resolution. When the ASPP module is used, the network can gather contextual data at different levels, and the depthwise separable convolution makes the computations easier. The model proposed by the authors achieved an impressive detection accuracy of 96.37% for the BCD dataset without the necessity of pre-training.
In this study, we aim to build a classification model that minimizes the prediction errors. At first, we built a model based on the LBP–BoVW, Apriori, and MobileNet algorithms. We achieved excellent accuracy with this model, ranging from 98.27% to 100%, and very few misclassified instances with an error rate between 1.73% and 0% for all datasets. The contributions of our work are summarized as follows:
  • Reliable hybrid supervised and unsupervised learning based on LBP–BoVW features for minimizing error rate;
  • A new algorithm for reliable learning on image datasets;
  • Employing the Apriori algorithm for selecting robust features and dimensionality reduction.

2. Materials and Methods

The proposed model contains multiple steps: preprocessing, feature extraction based on LBP–BoVW, feature selection and dimensionality reduction based on the Apriori algorithm, and classification using MobileNetV3_Large algorithms. Figure 1 explains the general structure of the proposed model.

2.1. Preprocessing

Depending on the purpose, nature, quality, and use of the data, one can employ various strategies, as data preparation is a crucial step in preparing for model training. Preprocessing the model might also boost model inference and reduce training time. Reducing the size of the input images can significantly reduce the model’s training time while still preserving its performance, especially if the images are very large. This paper employs the following preprocessing techniques:
  • Image resizing works the model correctly; images must be resized to a consistent size. A large image is difficult to process effectively. Therefore, we resized the different sizes of the used datasets to 100 × 100.
  • Grayscale image: A grayscale instead of color is used in this work to simplify the image’s data and lower the processing requirements of the algorithms.
  • Data augmentation techniques are used to produce new images from existing ones to increase the size of a dataset. It enhances the model’s generalization and decreases overfitting.
  • The data normalization technique sets pixel intensity values to a predefined range, usually between 0 and 1.

2.2. Local Binary Pattern

The LBP is a textural descriptor described by Ojala et al. [20] and utilized for classification. To calculate a binary pattern, LBP compares the intensity of a center pixel (threshold) with the intensity of the 8 neighbor pixels around it. Figure 2 shows how the LBP works. LBP may be formalized as follows for a center pixel Ic and an adjacent pixel Ii (i = 1, 2, …, p) [21]:
L B P =   i = 1 p F ( I i I c )   .   2 i 1
F I =   1 ,   0 ,   I   0 , o t h e r w i s e .

2.3. LBP–BoVW

The LBP–BoVW features are extracted by training the image patches using the k-means clustering technique, where the size of the dictionary (k) is determined, which represents the number of clusters, and the cluster centroids are used as visual words. The steps of LBP–BoVW are explained as follows:
  • Determine the number of patches needed to divide each image into patches. Here, we used 250 image patches.
  • Compute the LBP for each patch.
  • To determine the size of the dictionary (k), we used 50, 100, 200, and 300, which represent the number of K-means clusters.
  • Identify each cluster’s center. These centers are the visual words. The size of the visual word vector is equal to K.
  • Compute the histograms for each image to create the local feature vectors.
Figure 1 explains the steps above.

2.4. Feature Selection and Dimensionality Reduction

The hidden patterns and knowledge are extracted from massive amounts of data in the form of association rules, with the assistance of association rule mining. One of these association rule mining algorithms, the Apriori algorithm [22], uses frequent item set mining to generate association rules from a transactional dataset. Additionally, machine learning uses it for classification, feature selection, and dimensionality reduction. Selecting characteristic features from raw data is a fundamental stage in data mining. There are two main objectives for feature extraction. The first is to improve the classifier’s efficiency by representing each data object with a limited number of useful features, and the second is to remove features that are not representative of the data object, which will increase classification accuracy [23,24]. The main objectives of dimensionality reduction and feature selection are [25]:
  • We can reduce overfitting by eliminating redundant or unnecessary features from the model, especially when working with high-dimensional data.
  • Dimensionality and feature reduction can produce a simpler, more interpretable model that is easier to comprehend and explain.
  • Reducing execution time by shrinking the model’s size, or data pruning, assists in speeding up the training and inference processes.
  • By eliminating noisy or unnecessary features, the ability of a model to generalize to previously unobserved data can be enhanced.
  • By data pruning, the best trade-off between model size and performance is achieved, resulting in a balance between accuracy and complexity.
  • Feature reduction helps to enhance model performance and lower the chance of overfitting.
We used min-support = 0.9 and min-confidence = 0.9 as a threshold for feature selection. After applying the Apriori algorithm to the features derived from LBP–BoVW, the number of features that are selected in one image differs from the number of features that are selected in another image. So, the resulting features need to be processed to obtain an equal number of features for each image. To treat these missing values, we used the imputation mean method. When data are imputed, missing values are substituted with values that make sense given the information at hand. Data can be imputed in a way that maintains the set’s size and structure, prevents the removal of information that can be valuable, and lowers the predictability and bias of the estimations. Table 1 explains the number of features after Apriori feature selection and the size of the LBP–BoVW features for the used datasets.

2.5. MobileNet

MobileNet is a computer vision model that has been made available as open source by Google. It is specifically developed for training classifiers. The network is a CNN that uses depthwise convolutions to greatly decrease the parameter count in comparison to other networks, thereby creating a lightweight deep neural network. MobileNets are compact models designed to accommodate limited resources, such as low latency and low power consumption. They are specifically parameterized to cater to a range of use cases, making them well suited for mobile applications and other situations with constrained resources. The efficiency of MobileNet is attained by using depthwise separable convolutions, which decrease the number of multiply-accumulates and the total size of the model. Although MobileNets exhibit superior speed and compactness compared to other prominent networks, they may sacrifice some accuracy in comparison to bigger, resource-intensive models. Nevertheless, its performance remains commendable, with just a little decline in accuracy [26]. In our work, we trained the datasets using MobileNetV3-Large [27].

3. Results and Discussion

3.1. Experimental Environment

In this section, we demonstrate that the proposed model is generalizable and capable of minimizing error rate by evaluating its performance on several datasets using 10-fold cross-validation.
Our model was implemented in the TensorFlow and Keras environments. We developed it in Python 3.9 using the Anaconda Navigator IDE and utilized it with Windows 10 64-bit. An i7-8550U CPU run at 1.99 GHz and 32 GB of RAM, and the experiment’s settings were as follows: the batch size was 128, the epoch was 20, and the optimization method for loss was Adam optimization.

3.2. Datasets

To test our model, we used four bridge damage datasets and compare our work with other studies that used these datasets. The experiment datasets will be described in this section.
  • DIMEC-Crack Database: Lopez Droguett et al. [28] developed a new dataset for crack semantic segmentation. The dataset contains images extracted from video captured by an unmanned aerial vehicle (UAV) equipped with high-resolution cameras for several concrete bridges. Each image has a resolution of 1920 × 1080; in their study, they extracted non-overlapping patches of 96 × 96 pixels from each raw image, but we will use the original raw images. The dataset contains 10,092 high-resolution images separated into two classes: 7872 crack and 2220 non-crack images.
  • Bridge concrete damage (BCD): Xu et al. [19] introduced a dataset that includes 2068 images of bridge cracks and non-cracks. They captured the images using the Phantom 4 Pro’s CMOS surface array camera, boasting a resolution of 1024 × 1024. The images underwent two reductions, first to dimensions of 512 × 512 and then to a size of 224 × 224 to produce the dataset. This dataset consists of 4057 photos with cracks and 2013 images without cracks.
  • Bridge datasets: The dataset created by Zoubir et al. [17] yielded a total of 1304 cracked and 1806 non-cracked RGB bridge images at a resolution of 200 × 200. These images depict various concrete surfaces and cracks from the actual bridge examination. In order to minimize crack-like noise, the images were pre-processed using a 3 × 3 median filter after being converted to grayscale.

3.3. Evaluation Metrics

Because our models are binary classification, we utilized performance metrics that experiment with binary classification models, such as accuracy, precision, recall (also called sensitivity), f1-score, ROC–AUC score, and error rate. All these metrics are taken from the confusion matrix, which shows the classifier’s performance. Four elements comprise the confusion matrix: false positive (FP), false negative (FN), true positive (TP), and true negative (TN). Let us define:
  • y = actual labels;
  • y ^ = predicted labels;
  • N = total number of samples;
  • False positive (FP): Incorrectly predicted positives (actual negatives);
    F P =   i = 1 N ( y i = 0   a n d   y ^ i = 1 )  
  • False negative (FN): Incorrectly predicted negatives (actual positives);
    F N =   i = 1 N ( y i = 1   a n d   y ^ i = 0 )  
  • True Positive (TP): Correctly predicted positives;
    T P =   i = 1 N ( y i = 1   a n d   y ^ i = 1 )  
  • True Negative (TN): Correctly predicted negatives.
    T N =   i = 1 N ( y i = 0   a n d   y ^ i = 0 )
  • The formula for evaluation metrics is as follows [29,30]:
    Accuracy = TP + NT TP + NT + FN + FP
    Precision =   TP TP + FP
    Recall =   TP TP + FN
    F 1 score = 2 × Precision × Recall Precision + Recall
    Error   Rate =   FP + FN TP + TN + FP + FN
    R O C A U C = S P + S E 2
    S P = T N T N + F P
    where SP is the true negative rate (TNR), also called specificity, and SE is the true positive rate (TPR), also called sensitivity or recall [31].

3.4. Cross-Validation

For the experiment, we use 10-fold cross-validation, with 9 folds accounting for the training dataset, 1-fold for testing, and 10 process repetitions. Cross-validation is essential for addressing the issue of overfitting. The selection of cross-validation folds is essential. We aggregated the outcomes of all folds by means of averaging.

3.5. Discussion

In this section, we display a detailed analysis of the results of the proposed models on the utilized datasets. Many factors affect model performance, such as preprocessing operations and their vital role in classification performance, feature selection, dimensionality reduction, and selecting an appropriate classifier. Also, the number of features, image quality, and lighting conditions all impact classification performance. We evaluated the performance of the proposed algorithm using MobileNet model. Table 2 explains the performance of the proposed model using difference evaluation metrics using four sizes of features for each dataset.
As shown in Table 2, the BCD dataset with feature sizes of 100 performed the best, with an accuracy of 99.99% and a high recall of 99.99%, with a precision and F1 score of 100%, ROC–AUC of 99.99% and a 0.01 error rate. The best performance was attained by the Bridge dataset with a feature size of 300, with 99.97% for all performance metrics and an error rate of 0.03. The DIMEC-Crack dataset with feature sizes of 100 performed the best, with a high precision reaching 100%, which means there were no false positives, and 99.99% for all performance metrics, with an error rate reaching 0.01%. Figure 3 shows the plot of the average of the best accuracy and loss. Figure 4 shows the confusion matrix of the best results for each dataset.
Table 3 presents the accuracy scores for each dataset and compares them with related studies that classify the same datasets. The difference in methodologies prevents a direct comparison with these studies, so we will only compare accuracy.
Bhalaji Kharthik et al. [15] employed 12 transfer learning (TL) methods: VGG16 achieved an accuracy of 99.83%, with more than 99% for each precision, recall, and F1-score; VGG19 achieved an accuracy of 99.67%, with a precision, recall, and F1-score of >99%; Xception achieved 99.67% and more than 99% for each precision, recall, and F1-score; ResNet 50 achieved an accuracy of 99.67%, with the precision, recall, and F1-score reaching >99%; ResNet 101 achieved an accuracy of 99.5%, a precision of 100%, a recall of 99%, and an F1-score of 99.5%; ResNet 152 gained an accuracy of 99.83% and >99% for each precision, recall, and F1-score; InceptionV3 achieved an accuracy of 99.83%, a precision of 99%, a recall of 99%, and an f1-score of 99%; InceptionResNet V2 achieved an accuracy of 99.5%, with a precision of 100%, a recall of 99%, and an F1 score reaching 99.5%; MobileNet achieved an accuracy of 99.83%, with more than 99% for each precision, recall, and F1-score; MobileNetV2 achieved an accuracy of 99.83%, with a precision, recall, and F1 score of >99%; DenseNet 121 achieved an accuracy of 99.67%, with a precision, recall, and F1-score of 100%; and the EfficientNetB0 model yielded an accuracy of 99.6% and a precision of 99.83%, with a precision, recall, and F1-score of >99%.
Yang et al. [16] also used transfer learning, utilizing 13 layers of VGG16 and 2 FC layers. They trained their proposal with BCD and CCIC datasets. For the BCD dataset, they achieved an accuracy of 99.72%, a precision of 96.46%, and an AUC of 99.99%. In their study, Zoubir et al. [17] used ULBP, HOG, KPCA, and SVM to classify concrete cracks in the self-made bridge dataset. They achieved an accuracy of 99.26%, a precision of 95%, a recall of 99.23%, and an F1 score of 99.12%. Zoubir, with another team [18], utilized the Bridge dataset, supplemented it with culvert images, and applied transfer learning to the CNN models. They achieved an accuracy of 94.89% in VGG16, 95.39% in VGG19, and 95.89% in InceptionV3. The InceptionV3 model produced the most favorable outcomes. The model proposed by Xu et al. [19] is based on the atrous spatial pyramid pooling (ASPP) module and depthwise separable convolution. They trained the BCD dataset on 300 epochs, achieving an accuracy of 96.37%, with a precision of 78.11%, 100% recall, and an F1-score of 87.71%. When we compared our proposal to previous studies and with the MobileNetV3_Large model, we found that it outperformed in all feature sizes.

4. Conclusions

This paper presents a method for enhancing bridge crack prediction based on LBP–-BoVW for feature extraction, an Apriori algorithm for selecting the strongest features and dimensionality reduction, and MobileNetV3_Large as a classifier. The proposed method achieved promising results for three bridge crack image datasets, with an accuracy rate of between 98.27% and 100% and an error rate of between 1.73% and 0%. Although we trained the data using 10-fold cross-validation, which tested all dataset samples with 20 epochs, our proposed system achieved promising results compared to previous studies that used the same datasets with more training epochs and did not use cross-validation or have a lower k-fold. We suggest using our proposed method on real-time datasets in future work.

Author Contributions

Conceptualization, A.A.M. and S.P.; methodology, A.A.M. and S.P.; software, A.A.M.; validation, A.A.M., S.P. and P.S.; formal analysis, A.A.M.; investigation, A.A.M.; resources, A.A.M.; writing—original draft preparation, A.A.M. and S.P.; writing—review and editing, A.A.M.; visualization, A.A.M.; supervision, S.P. and P.S.; project administration, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used in this study is publicly available and mentioned in the references.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

References

  1. Nasr, A.; Kjellström, E.; Björnsson, I.; Honfi, D.; Ivanov, O.L.; Johansson, J. Bridges in a changing climate: A study of the potential impacts of climate change on bridges and their possible adaptations. Struct. Infrastruct. Eng. 2020, 16, 738–749. [Google Scholar] [CrossRef]
  2. Ni, Y.; Mao, J.; Fu, Y.; Wang, H.; Zong, H.; Luo, K. Damage Detection and Localization of Bridge Deck Pavement Based on Deep Learning. Sensors 2023, 23, 5138. [Google Scholar] [CrossRef] [PubMed]
  3. Bi, Q.; Goodman, K.E.; Kaminsky, J.; Lessler, J. What is Machine Learning? A Primer for the Epidemiologist. Am. J. Epidemiol. 2019, 188, 2222–2239. [Google Scholar] [CrossRef] [PubMed]
  4. Krahe, C.; Kalaidov, M.; Doellken, M.; Gwosch, T.; Kuhnle, A.; Lanza, G.; Matthiesen, S. AI-Based knowledge extraction for automatic design proposals using design-related patterns. Procedia CIRP 2021, 100, 397–402. [Google Scholar] [CrossRef]
  5. Cao, J.; Huang, Z.; Shen, H.T. Local deep descriptors in bag-of-words for image retrieval. ACM Multimed. 2017, 52–58. [Google Scholar] [CrossRef]
  6. Hassan, R.Q.; Sultani, Z.N.; Dhannoon, B.N. Content-Based Image Retrieval System using Color Moment and Bag of Visual Words with Local Binary Pattern. KIJOMS 2023, 9, 7. [Google Scholar] [CrossRef]
  7. Ngoc, V.T.N.; Agwu, A.C.; Son, L.H.; Tuan, T.M.; Nguyen Giap, C.; Thanh, M.T.G.; Duy, H.B.; Ngan, T.T. The combination of adaptive convolutional neural network and bag of visual words in automatic diagnosis of third molar complications on dental x-ray images. Diagnostics 2020, 10, 209. [Google Scholar] [CrossRef]
  8. Afonso, L.C.; Pereira, C.R.; Weber, S.A.; Hook, C.; Falcão, A.X.; Papa, J.P. Hierarchical learning using deep optimum-path forest. J. Vis. Commun. Image Represent. 2020, 71, 102823. [Google Scholar] [CrossRef]
  9. Tripathi, S.; Singh, S.K.; Kuan, L.H. Bag of Visual Words (BoVW) with Deep Features--Patch Classification Model for Limited Dataset of Breast Tumours. arXiv 2022, arXiv:2202.10701. [Google Scholar] [CrossRef]
  10. Kumar, M.D.; Babaie, M.; Zhu, S.; Kalra, S.; Tizhoosh, H.R. A comparative study of CNN, BoVW and LBP for classification of histopathological images. In Proceedings of the 2017 IEEE Symposium Series on Computational Intelligence (SSCI), Honolulu, HI, USA, 27 November–1 December 2017. [Google Scholar] [CrossRef]
  11. Huang, H.; Xu, K. Combing Triple-Part Features of Convolutional Neural Networks for Scene Classification in Remote Sensing. Remote Sens. 2019, 11, 1687. [Google Scholar] [CrossRef]
  12. Olaode, A.; Naghdy, G. Local Image Feature Extraction using Stacked-Autoencoder in the Bag-of-Visual Word modelling of Images. In Proceedings of the 2019 IEEE 5th International Conference on Computer and Communications (ICCC), Chengdu, China, 6–9 December 2019; pp. 1744–1749. [Google Scholar] [CrossRef]
  13. Govender, D.; Tapamo, J.-R. Spatio-temporal scale coded bag-of-words. Sensors 2020, 20, 6380. [Google Scholar] [CrossRef] [PubMed]
  14. Sultani, Z.N.; Dhannoon, B. Modified Bag of Visual Words Model for Image Classification. ANJS 2021, 24, 78–86. [Google Scholar] [CrossRef]
  15. Bhalaji Kharthik, K.S.; Onyema, E.M.; Mallik, S.; Siva Prasad, B.V.V.; Qin, H.; Selvi, C.; Sikha, O.K. Transfer learned deep feature based crack detection using support vector machine: A comparative study. Sci. Rep. 2024, 14, 14517. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, Q.; Shi, W.; Chen, J.; Lin, W. Deep convolution neural network-based transfer learning method for civil infrastructure crack detection. Autom. Constr. 2020, 116, 103199. [Google Scholar] [CrossRef]
  17. Zoubir, H.; Rguig, M.; El Aroussi, M.; Chehri, A.; Saadane, R. Concrete Bridge Crack Image Classification Using Histograms of Oriented Gradients, Uniform Local Binary Patterns, and Kernel Principal Component Analysis. Electronics 2022, 11, 3357. [Google Scholar] [CrossRef]
  18. Zoubir, H.; Rguig, M.; Elaroussi, M. Crack recognition automation in concrete bridges using Deep Convolutional Neural Networks. MATEC Web Conf. 2021, 349, 03014. [Google Scholar] [CrossRef]
  19. Xu, H.; Su, X.; Wang, Y.; Cai, H.; Cui, K.; Chen, X. Automatic bridge crack detection using a convolutional neural network. Appl. Sci. 2019, 9, 2867. [Google Scholar] [CrossRef]
  20. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  21. Yu, Z.; Cai, R.; Cui, Y.; Liu, X.; Hu, Y.; Kot, A.C. Rethinking vision transformer and masked autoencoder in multimodal face anti-spoofing. Int. J. Comput. Vis. 2024, 132, 5217–5238. [Google Scholar] [CrossRef]
  22. Agrawal, R.; Imielinski, T.; Swami, A. Database mining: A performance perspective. IEEE Trans. Knowl. Data Eng. 1993, 5, 914–925. [Google Scholar] [CrossRef]
  23. Zakur, Y.; Flaih, L. Apriori Algorithm and Hybrid Apriori Algorithm in the Data Mining: A Comprehensive Review. E3S Web Conf. 2023, 448, 02021. [Google Scholar] [CrossRef]
  24. Kharsa, R.; Aghbari, Z.A. Association rules based feature extraction for deep learning classification. icSoftComp2022 2022, 1788, 72–83. [Google Scholar] [CrossRef]
  25. Mamdouh Farghaly, H.; Abd El-Hafeez, T. A high-quality feature selection method based on frequent and correlated items for text classification. Soft Comput. 2023, 27, 11259–11274. [Google Scholar] [CrossRef]
  26. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  27. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Adam, H. Searching for mobilenetv3. arXiv 2019, arXiv:1905.02244. [Google Scholar] [CrossRef]
  28. Lopez Droguett, E.; Tapia, J.; Yanez, C.; Boroschek, R. Semantic segmentation model for crack images from concrete bridges for mobile devices. Proc. Inst. Mech. Eng. O 2022, 236, 570–583. [Google Scholar] [CrossRef]
  29. Brzezinski, D.; Stefanowski, J. Prequential AUC: Properties of the area under the ROC curve for data streams with concept drift. Knowl. Inf. Syst. 2017, 52, 531–562. [Google Scholar] [CrossRef]
  30. Hu, B.-G.; Dong, W.-M. A study on cost behaviors of binary classification measures in class-imbalanced problems. arXiv 2014, arXiv:1403.7100. [Google Scholar] [CrossRef]
  31. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the Areas under Two or More Correlated Receiver Operating Characteristic Curves: A Nonparametric Approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef]
Figure 1. The general structure of the proposed model.
Figure 1. The general structure of the proposed model.
Asi 08 00061 g001
Figure 2. Local binary pattern calculation.
Figure 2. Local binary pattern calculation.
Asi 08 00061 g002
Figure 3. The accuracy and loss plot of the best results for (a) the BCD dataset with a feature size of 100, (b) the Bridge dataset with a feature of size 300, and (c) the DIMEC-Crack dataset with a feature size of 100.
Figure 3. The accuracy and loss plot of the best results for (a) the BCD dataset with a feature size of 100, (b) the Bridge dataset with a feature of size 300, and (c) the DIMEC-Crack dataset with a feature size of 100.
Asi 08 00061 g003
Figure 4. The confusion matric of the best results for (a) the BCD dataset with a feature size of 100, (b) the Bridge dataset with a feature size of 300, and (c) the DIMEC-Crack dataset with a feature size of 100.
Figure 4. The confusion matric of the best results for (a) the BCD dataset with a feature size of 100, (b) the Bridge dataset with a feature size of 300, and (c) the DIMEC-Crack dataset with a feature size of 100.
Asi 08 00061 g004
Table 1. Number of features after and before Apriori feature selection.
Table 1. Number of features after and before Apriori feature selection.
DatasetsNo. LBP–BoVW Features
50100200300
No. Features After Apriori Feature SelectionDIMEC-Crack Database21191817
BCD21171414
Bridge20191818
Table 2. The results of the proposed model.
Table 2. The results of the proposed model.
DatasetsPerformance MetricsNumber of Visual Words
50100200300
BCDAccuracy99.9899.9999.9899.97
Precision99.97100100100
Recall10099.9999.9799.97
F1-score99.9910099.9899.98
ROC–AUC99.9799.9999.9899.98
Error rate0.020.010.020.03
Bridge datasetAccuracy99.8299.8499.9699.97
Precision99.7899.9099.9799.97
Recall99.8199.7599.9499.97
F1-score99.7999.8399.9599.97
ROC–AUC99.8299.8399.9699.97
Error rate0.180.160.040.03
DIMEC-CrackAccuracy99.9899.9999.9899.97
Precision10010010099.99
Recall99.9799.9999.9899.97
F1-score99.9999.9999.9999.98
ROC–AUC99.9999.9999.9999.97
Error rate0.020.010.020.03
Table 3. The classification accuracy score comparison.
Table 3. The classification accuracy score comparison.
RefrencesMethodK-Fold CVNo. of EpochsDatasets
BCDBridgeDIMEC-Crack
Bhalaji Kharthik et al. [15]VGG16--99.83--
VGG1999.67--
Xception99.67--
ResNet 5099.67--
ResNet 10199.5--
ResNet 15299.83--
InceptionV399.83--
InceptionResNet V299.5--
MobileNet99.83--
MobileNetV299.83--
DenseNet12199.67--
EfficientNetB099.83--
Yang et al. [16]TL model-2099.72--
Zoubir et al. [18]TL model5-fold10-95.89-
Zoubir et al. [17]HOG + ULBP + KPCA+ SVM5-fold--99.29-
Xu et al. [19]Atrous convolution, ASPP, and depthwise separable convolution-30096.37--
MobileNetV3_Large10-fold2082.1472.5991.51
Proposed methodLBP-BoVW + Appriori + MobileNetV3_Large10-fold2099.9999.97100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Maryoosh, A.A.; Pashazadeh, S.; Salehpour, P. A Hybrid Learning Framework for Enhancing Bridge Damage Prediction. Appl. Syst. Innov. 2025, 8, 61. https://doi.org/10.3390/asi8030061

AMA Style

Maryoosh AA, Pashazadeh S, Salehpour P. A Hybrid Learning Framework for Enhancing Bridge Damage Prediction. Applied System Innovation. 2025; 8(3):61. https://doi.org/10.3390/asi8030061

Chicago/Turabian Style

Maryoosh, Amal Abdulbaqi, Saeid Pashazadeh, and Pedram Salehpour. 2025. "A Hybrid Learning Framework for Enhancing Bridge Damage Prediction" Applied System Innovation 8, no. 3: 61. https://doi.org/10.3390/asi8030061

APA Style

Maryoosh, A. A., Pashazadeh, S., & Salehpour, P. (2025). A Hybrid Learning Framework for Enhancing Bridge Damage Prediction. Applied System Innovation, 8(3), 61. https://doi.org/10.3390/asi8030061

Article Metrics

Back to TopTop