Next Article in Journal
Modified Exp-Function Method to Find Exact Solutions of Microtubules Nonlinear Dynamics Models
Previous Article in Journal
Entropy Generation for MHD Peristaltic Transport of Non-Newtonian Fluid in a Horizontal Symmetric Divergent Channel
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Cracked Tiers Detection System Based on Adaptive Correlation Features Selection and Deep Belief Neural Networks

by
Ali Mohsin Al-juboori
1,*,
Ali Hakem Alsaeedi
1,
Riyadh Rahef Nuiaa
2,
Zaid Abdi Alkareem Alyasseri
3,4,5,*,
Nor Samsiah Sani
6,*,
Suha Mohammed Hadi
7,
Husam Jasim Mohammed
8,
Bashaer Abbuod Musawi
9 and
Maifuza Mohd Amin
6
1
College of Computer Science and Information Technology, Al-Qadisiyah University, Diwaniyah 58009, Iraq
2
Department of Computer, College of Education for Pure Sciences, Wasit University, Wasit 52001, Iraq
3
Information Technology Research and Development Center (ITRDC), University of Kufa, Najaf 54002, Iraq
4
College of Engineering, University of Warith Al-Anbiyaa, Karbala 56001, Iraq
5
Institute of Informatics and Computing in Energy, Universiti Tenaga Nasional (UNITEN), Jalan Ikram-Uniten, Kajang 43000, Selangor, Malaysia
6
Center for Artificial Intelligence Technology, Faculty of Information Science &Technology, University Kebangsaan Malaysia, Bangi 43600, Selangor Darul Ehsan, Malaysia
7
Informatics Institute for Postgraduate Studies, Iraqi Commission for Computer and Informatics, Baghdad 10001, Iraq
8
Department of Business Administration, College of Administration and Financial Sciences, Imam Ja’afar Al-Sadiq University, Baghdad 10001, Iraq
9
Department of Biology, Faculty of Education for Girls, University of Kufa, Najaf 54001, Iraq
*
Authors to whom correspondence should be addressed.
Symmetry 2023, 15(2), 358; https://doi.org/10.3390/sym15020358
Submission received: 7 December 2022 / Revised: 28 December 2022 / Accepted: 7 January 2023 / Published: 29 January 2023
(This article belongs to the Section Computer)

Abstract

:
Tire defects are crucial for safe driving. Specialized experts or expensive tools such as stereo depth cameras and depth gages are usually used to investigate these defects. In image processing, feature extraction, reduction, and classification are presented as three challenging and symmetric ways to affect the performance of machine learning models. This paper proposes a hybrid system for cracked tire detection based on the adaptive selection of correlation features and deep belief neural networks. The proposed system has three steps: feature extraction, selection, and classification. First, the oriented gradient histogram extracts features from the tire images. Second, the proposed adaptive correlation feature selection selects important features with a threshold value adapted to the nature of the images. The last step of the system is to predict the image category based on the deep belief neural networks technique. The proposed model is tested and evaluated using real images of cracked and normal tires. The experimental results show that the proposed solution performs better than the current studies in effectively classifying tire defect images. The proposed hybrid cracked tire detection system based on adaptive correlation feature selection and Deep Belief Neural Networks’ performance provided better classification accuracy (88.90%) than that of Belief Neural Networks (81.6%) and Convolution Neural Networks (85.59%).

1. Introduction

Recently, the increase of electronic devices and evolving digital technologies have increased the complexity of data [1]. For that, the framework of artificial intelligence (AI) for handling data has been improved. Image classification is considered one of the biggest challenges for machine learning and is one of the complex approaches of AI. Several intelligent applications deal directly with images and the necessity for life management, such as healthy applications [2,3,4,5], security management systems [6,7,8], and safe driving [9,10,11]. Intelligent applications that deal with safety management systems are essential for researchers, companies, and humans. Detecting wear and cracks in tires is an important application for safe driving [12]. According to the National Highway Traffic Safety Administration (AHTSA), more than 9500 people died from destroyed car tires in 2021 [12]. Despite the importance of tires, few drivers keep an eye on them. Although vehicle owners know the dangers of underinflated tires, they continue to use them. With lesser-known problems such as tire aging, oxidation, and cracking, drivers are unaware that they need to replace them, so they avoid inspection altogether.
Deep learning is increasingly used for visual condition monitoring to gain insights from complex transformations and even to transition from prognostics to diagnosis [13,14,15,16]. These developments have been used with infrared imaging to detect wear and other problems in mechanical systems. The systems check whether a vehicle’s tires are broken or normal and have to analyze the snapshot of the tires. Therefore, high domination of features is attracted from images during the analysis process [17]. Generally, feature extraction algorithms describe an image with many features [3]. Not all of these features need to be insignificant to the subject of the image, so feature selection is necessary to optimize the components extracted from the image and improve machine learning performance. The direct application of machine learning on these features resulted in a poor performance [18,19]. Feature selection can be essential in enhancing the correct detection rate of machine learning algorithms.
This paper proposes a hybrid cracked tiers detection system based on adaptive correlation features selection and deep belief neural networks (H-DBN). The proposed framework consists of feature extraction using histogram of oriented gradient (HOG), selecting optimal features using hybrid gradient correlation-based feature selection (HCS), and DBN as a predictor. The problem with detecting cracks in tires is overlapping textures. The HOG extracts the features of tires without analyzing the overlap in the features of the images. The proposed feature selection approach selects the features that achieve the necessary discrimination or distinguish the different classes as well as possible. Figure 1 illustrates the proposed scenario for implementing intelligent eyes to detect a vehicle tire defect.
The dataset in [10] was used as a benchmark to test the validity of the H-DBN system. The metrics accuracy, precision, recall, and F1 score were used to evaluate the performance of H-DBN. We compare the proposed (H-DBN) system with three machine learning algorithms: naïve Bayes (BN), random forest (RF), and decision tree (DT). In addition, the performance of H-DBN compared with recent tire crack detection models in terms of accuracy.
Several studies have enforced different classification models for text messages to develop image classification. Most methods used to classify images suffer from low recognition rates when dealing with images that are very similar in texture. The overlap of texture features leads to a high error rate in classification, so machine learning and deep learning may not achieve a high accuracy rate.
Siegel et al. [10] used convolution neural network (CNN) to determine the cracks in vehicle tires. The authors explore the need for monitoring tire health and consider the evolution of a densely connected convolutional neural network. This model achieves 78.5% accuracy on cropped sample images, outperforming human performance of 55%.
In [20], a new model for classifying tire wear was proposed. The tread feature is highlighted by an attention mechanism proposed by the authors. The overlap of natural and cracked textures makes it difficult for deep learning algorithms to distinguish between them, so without early processing of the images, the algorithm certainly achieves low accuracy in recognizing the tire category.
Ali et al. [3] used personal component analysis (PCA) and histogram-based gradient (HOG) as descriptors of X-ray images for machine learning algorithms. The features selected in the proposed model are based on the wrapper model. The limitation of the proposed model is the inconsistency of the results because the wrapper model is based on randomness in selecting features. Too et al. [21] proposed two CNN models, one for feature extraction and the second for the image prediction category. The limitation of the proposed framework is the ability of convolution to extract visual patterns from different spatial positions.
Kumar [22] proposed a new framework for a feature selection model based on deep learning and the Pearson correlation coefficient. The features extracted from the traditional deep learning architectures ResNet152 and Google Net are optimized in the feature space of the proposed model using the concepts of Pearson correlation coefficient (PCC) and variance threshold. The restriction of the proposed model to static thresholds and static methods does not fit the dynamics associated with most of the problems.

2. Materials and Methods

This section discusses the Materials and Methods used in this paper.

2.1. Deep Belief Networks (DBN)

A deep belief network (DBN) is a type of neural network that includes three techniques: a deep neural network (DNN), cascading restricted Boltzmann machines (RBMs), and a back-propagation neural network (BPNN) [23]. RBMs are typically used for dimensionality reduction, extraction, and collaborative feature filtering before feeding into BPNN. In the deep belief network algorithm, the DNN is quickly trained by DBN due to using the greedy layer-wise unsupervised training method as a training technique [24]. Like many other neural networks, the DBN is based on the fundamental principle of initializing feed-forward neural networks (FFNNs) with unsupervised pre-training on unlabeled data before fine-tuning with labeled data [25]. The first RBM is trained to use contrastive divergence (CD). The weights of all RBMs are similarly trained layer by layer until the last RBM. The RBMs in the lower layers of the RBM stack learn lower-level features, and the RBMs in the higher layers learn higher-level features of the training data. Figure 2 illustrates the architecture of a DBN.
The main constrictor of RBM is visible (n nodes), Bernoulli random value hidden units (m nodes), and the weight matrix connects between them with the size ( m × n ) [23]. Figure 3 illustrates the graphical representation of the RBM.
According to the contrastive divergence algorithm (CD) strategy, the RBM trains the sample value of (v) to compute the probability of the hidden unit (h). Equation (1) calculates the energy function E(v, h) of the joint configuration {v, h} [26]:
E ( v , h ) = i = 1 m b i v i j = 1 n c j h j j = 1 n i = 1 m v i w i j h j
where b and c are the biases value of visible and hidden units, and w is the weight matrix that connects the visible units to the hidden ones. The joint probability of pair ( v , h ) is computed by Equation (2):
p ( v , h ) = e E ( v , h ) v , h e E ( v , h )
Since the units within the visible and hidden layers are not connected, the activation probabilities of ith visible unit and jth hidden unit is given as:
p ( v i = 1 h ) = σ ( j = 1 m h j w i j + a i )
p ( h j = 1 v ) = σ ( i = 1 n v i w i j + b i )
where σ ( · ) is the sigmoid logistics function.
The weight between the visible and hidden units is updated based on the positive gradient minus the negative gradient ( v h ) times some learning rate ( α ). Equation (5) calculates the update in the weight.
Δ W = α v h v h

2.2. Histogram of Oriented Gradients (HOG)

Histogram of oriented gradients (HOG) is a feature extraction technique from 2D images. It is an efficient image descriptor of the texture image [27]. The HOG extracts feature based on the oriented gradient of colors in localized portions in each window (sub-region or small regions of the image). The number of features extracted from the image by HOG depends on the cell size, number of blocks, and number of orientations [28]. Based on gradient magnitude m   ( x , y ) and orientation θ x ,   y , the cell’s characteristics ( x , y ) are calculated based on the spectral value of a pixel in position ( x , y ) . Equations (6) and (7) use the x- and y-directional gradients d x   ( x , y ) and d y ( x , y ) to derive the magnitude and orientation of pixel ( x , y ) [3].
m x , y = d x   ( x , y ) 2 d y ( x , y ) 2 2
θ x ,   y = { t a n 1 ( d y ( x , y ) d x ( x , y ) ) π   i f   d x ( x , y )   a n d   d y ( x , y ) > 0   t a n 1 ( d y ( x , y ) d x ( x , y ) ) + π   i f   d x ( x , y )   a n d   d y ( x , y ) < 0 t a n 1 ( d y ( x , y ) d x ( x , y ) )   o t h e r w i s e  

2.3. Correlation-Based Feature Selection (CFS)

The feature selection technique is one of the essential processes for optimizing the performance of machine learning algorithms [29,30]. The filter model is distinguished from the other feature selection models by the stability of the selected features and the number of features [31].
The CFS technique is a filter method unrelated to the chosen classification model. As the name implies, correlations are the only intrinsic qualities used to evaluate feature subsets. It determines the strength of the correlation between variables. Equation (8) calculates the correlation t ( x , y ) of features x , y , The feature is selected if t ( x , y ) greater than the critical value at the 0.05 significance level [32]. Equation (9) calculates the correlation between features x and y.
r ( x , y ) = i ( x i x ¯ i ) ( y i y ¯ i ) i ( x i x ¯ i ) 2 j ( y j y ¯ j ) 2
t ( x , y ) = r n 1 1 r 2
where n is the dimension of the feature.

2.4. Proposed Hybrid Adaptive–CFS and DBN (H-DBN):

The proposed hybrid adaptive–correlation-based feature selection and deep belief neural networks (H-DBN) system trains the model via four steps before: preprocessing, extracting features, selecting optimal and high correlation features, and finally, predicting the data (image) category. Figure 4 illustrates the main steps of the proposed H-DBN model.

3. Preprocessing

At this stage, the dimensions of the images are the same so that the character extraction step can extract features of the same dimension from all images. The nearest neighbor interpolation technology is used to normalize the size of the image to ( 700 × 800 ) pixels. The method used to resize the image by replication is the nearest neighbor interpolation. Accordingly, resampling and interpolation are used to resize the images. Equation (10) generates the image g ( m , n ) by factorizing ( c ) in the ( m ) direction and ( d ) in the ( n ) direction from the image f ( m , n ) .
g ( m ,   n ) = { 1 c d , c 2   m < c 2 , d 2   n < d 2 0   o t h e r w i s e  

4. Extraction Features

In the proposed H-DBN, the HOG is used as a feature extractor. This step extracts many features, some of which may not be helpful. Therefore, selecting features with high significance that can be used as input data for the DBN algorithm is necessary. Figure 5 illustrates the overlap of features extracted from HOG.

5. Adaptive Correlation-Based and Variance Feature Selection (ACFS)

The selected features based on only the correlation coefficient do not give definitive evidence to impotent features [33,34,35]. The selected features found on only the correlation coefficient do not give definitive proof these features are essential. This paper proposes a new approach to feature selection: adaptive correlation-based and variance feature selection (ACFS). It selects features based on the correlation coefficient and the feature’s variance with each class in the dataset. The correlation coefficient shows the extent of the relationship of the trait with the rest of the traits. To support this relationship, the proposed model should show the significance of the change in the specific characteristic data in both classes. In the proposed ACFS, if the evidence for the adjective changes significantly in a particular category, then it is impossible to rely on the tagging of the adjective to distinguish the adjective category. Equation (12) calculates the correlation coefficient c of feature x .
V = j = 1 k ( ( x i , j x j ¯ ) 2 n 1 )
C x = i = 1 n t ( x , y i ) + V 1 N . K
where N is the number of features, K: the number of classes, x i , j each point in ith location includes into class j ,   j = 1 , 2 , 3 , , n and i = 1 , 2 , 3 , , k , and V is the variance between feature x and another feature of the same data class.
The selection of essential features dramatically improves the performance of the machine training algorithm; therefore, the set of a threshold defining important features must be carefully prepared. If the threshold value is in the interval [−∞, +∞], it is not easy to choose the optimal value of the threshold. This paper proposes adding ±β to the gradient search to find the optimal threshold value. The initial value ( α o ) of the threshold is zero (zero is the minor possible threshold). After each iteration (t), the model either adds β to the current threshold ( α t + 1 ) or subtracts β from the threshold ( α t + 1   ), depending on the accuracy of the selected features. The model continues by adding β until reflecting the accuracy of machine learning. Equation (13) illustrates the modification of the threshold during the search progress.
α t + 1   = { α t + β ,   A t + 1 < A t α t β ,   o t h e r w i s e
where A is the accuracy, β random value between [0, 1].
It was limiting the search value to the best threshold and adding weight gradually decreasing after each search step.
ω = 0.2 ( t t m a x ) 0.1
Equation (15) is a new form of Equation (13):
α t + 1   = { α t + β ω ,   A t + 1 < A t α t β ω ,   o t h e r w i s e
Figure 6 shows the bin’s interval of the value α with the frequency of each value.
Figure 7 shows the flow of the proposed ACFS.
Algorithm 1 illustrates the flow of the proposed ACFS.
Algorithm 1: ACFS.
X← Data //vector of features
N← max number of gradient searches
αt←0
initialization;
while i ≤ N do
//select features according to Eq(14)
F ← select features(αt, X) // t current iteration
F1 ← select features((αt + β), X)
F2 ← select features((αt − β),X)
// Evaluated features F, F1, and F2 by DBN or any machine learning algorithms
Acc ← Evaluate features(F)
Acc1 ← Evaluate features (F1)
Acc2 ← Evaluate features (F2)
If Acc < Acc1 then
i ← 0
F ← F1
Else if Acc < Acc2 then
i ← 0
F ← F2
Else
i ← i+1
Return F // Optimal Features

6. Experiment Results

This section discusses the dataset and empirical results.

6.1. Dataset

A dataset with a total of 1028 images of the tire (cracked/normal) was used in this study [9]. The images were taken at various angles, from various distances, in different lighting situations, and with multiple quantities of dirt on them. Suppose the classifier is made available to the general public as a diagnostic service. This strategy ensures that the training set is typical of the kinds of photographs an end-user would take with their cell phone. Figure 8 shows an example from the tire dataset

6.2. Empirical Results

In this section, the validity of the proposed H-DBN model is tested in two experiments. The first experiment compares the proposed H-DBN with machine learning algorithms (naïve Bayes (NB), random forest (RF), and decision tree (DT)). The second experiment compares the proposed H-DBN with current models for crack detection in tires. This experiment checked the validity of the proposed feature selection on the performance of machine learning algorithms.
The algorithm NB did not perform well compared to the other algorithms because it uses a probability factor in estimating the data type. It also does not cope with the complexity of evidence from overlapping features. The accuracy of NB improved from 64% to 68% when ACFS used for feature selection before entering the data in NB. DT and RF depend on the entropy of the features used in the training algorithms, so the high accuracy depends on the correlations of the features. The results of RF and DT without feature selection (stand-alone) reach 78.6% and 63.9% of RF and DT, respectively. With feature selection from CFS, the archive of RF achieves 78.8% accuracy and DT 63.9%. The algorithms achieved high results in selecting features from the data using the proposed ACFS method compared to previous results. The proposed feature selection model improves the accuracy, precision, recognition, and f1 score of DT and RF. RF achieves 66.1%, 64.9%, 64.8%, and 64.8% in accuracy, precision, recall, and f1 score, respectively. A large number of features entered into the algorithm ANN caused the algorithm to fail to achieve positive results, as it reached a consistent diagnostic accuracy of up to (67.10%) without using feature selection. They improve the accuracy of ANN by 1% with the new feature selection. This strongly indicates that the feature extraction algorithms improved the discrimination between the different categories. The new method improved the rate of class prediction by DBN to 9.8%. Table 1 shows the comparison of the results according to the accuracy, precision, recognition, and f1 score criteria between the algorithms (NB, RF, DT, and DBN) of Standalone, CFS, and ACFS.
Figure 9 illustrates the enhancement in the true negative rate (TNR) of machine learning when using the ACFS.
DBN achieves the best accuracy, so the analysis performance of DBN when using ACFS is necessary for constructing the new framework. The proposed framework significantly improved the performance of DBN in terms of the rate of true positives rate (TPR) and the rate of true negatives (TNR). Figure 10 shows the comparison between the stand-alone DBN and the proposed H-DBN.
The proposed model shows a significant effect by establishing a high correlation between precision and recall. See receiver operating characteristic (ROC) for a visualization of the evaluation and comparison of the performance of DBN and the proposed H-DBN. Figure 11 shows the performance evaluation of DBN and H-DBN concerning ROC.
Table 2 compares the proposed H-DBN and three works in image classification over cracked tire prediction. These works were chosen as benchmarks because they are close in principle to the proposed method. The systems proposed in [10,20] for detecting cracks in tires can be compared with the proposed H-DBN model to verify the validity of the proposed system. In [3], feature selection is proposed after feature extraction by HOG, so that an indication can be given of the extent to which the selected features match the machine learning algorithm.
The proposed system outperformed standard machine learning and deep learning in terms of correct classification accuracy rate. In other words, the proposed feature selection repeatedly improved the accuracy of classifications made by machine learning algorithms and demonstrated its value as a possible way to achieve this goal.

7. Conclusions and Future Works

One of the biggest challenges for machine learning is image classification. One important application for safe driving is detecting cracks in tires. The correct detection rate of machine learning algorithms can be increased by better feature selection. The three components of the proposed framework are DBN as a predictor, hybrid gradient correlation-based feature selection to select the best features and histogram of oriented gradient (HOG) for feature extraction. It is not enough to use feature extraction directly for machine learning to extract many features common to image classes, such as the sky region and patches of color such as blue, black, etc. Therefore, feature selection and feature optimization are necessary to achieve good discrimination between image categories. The RBM in the DBN algorithm plays a vital role in efficiently optimizing features. For future work, we propose to use the RBM before the apartment layer of the deep learning architecture.

Author Contributions

Methodology, A.M.A.-j. and A.H.A.; Software, A.M.A.-j., A.H.A. and M.M.A.; Validation, Z.A.A.A., H.J.M. and N.S.S.; Formal analysis, R.R.N. and B.A.M.; Investigation, Z.A.A.A. and B.A.M.; Resources, S.M.H. and H.J.M.; Data curation, R.R.N.; Writing—original draft, A.M.A.-j., R.R.N., S.M.H., A.H.A. and M.M.A.; Writing—review & editing, Z.A.A.A., H.J.M., N.S.S. and B.A.M.; Visualization, A.M.A.-j.; Project administration, Z.A.A.A.; Funding acquisition, N.S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Universiti Kebangsaan Malaysia (Grant code: GUP-2022-060).

Data Availability Statement

The dataset consists of 1028 images of tires in total. The data is available in link https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/Z3ZYLI.

Acknowledgments

We appreciate the University Kebangsaan Malaysia (Grant code: GUP-2022-060).

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Hadi, S.M.; Alsaeedi, A.H.; Al-Shammary, D.; Alyasseri, Z.A.A.; Mohammed, M.A.; Abdulkareem, K.H.; Nuiaa, R.R.; Jaber, M.M. Trigonometric words ranking model for spam message classification. IET Netw. 2022. [Google Scholar] [CrossRef]
  2. González-Ortiz, A.M.; Bernal-Silva, S.; Comas-García, A.; Vega-Morúa, M.; Garrocho-Rangel, M.E.; Noyola, D.E. Severe Respiratory Syncytial Virus Infection in Hospitalized Children. Arch. Med. Res. 2019, 50, 377–383. [Google Scholar] [CrossRef] [PubMed]
  3. Ali, S.S.M.; Alsaeedi, A.H.; Al-Shammary, D.; Alsaeedi, H.H.; Abid, H.W. Efficient intelligent system for diagnosis pneumonia (SARS-COVID19) in X-ray images empowered with initial clustering. Indones. J. Electr. Eng. Comput. Sci. 2021, 22, 241–251. [Google Scholar] [CrossRef]
  4. Nematzadeh, S.; Kiani, F.; Torkamanian-Afshar, M.; Aydin, N. Tuning hyperparameters of machine learning algorithms and deep neural networks using metaheuristics: A bioinformatics study on biomedical and biological cases. Comput. Biol. Chem. 2021, 97, 107619. [Google Scholar] [CrossRef] [PubMed]
  5. Bassel, A.; Abdulkareem, A.B.; Alyasseri, Z.A.A.; Sani, N.S.; Mohammed, H.J. Automatic Malignant and Benign Skin Cancer Classification Using a Hybrid Deep Learning Approach. Diagnostics 2022, 12, 2472. [Google Scholar] [CrossRef]
  6. Boutros, F.; Damer, N.; Kirchbuchner, F.; Kuijper, A. ElasticFace: Elastic Margin Loss for Deep Face Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 1577–1586. [Google Scholar] [CrossRef]
  7. Alfoudi, A. University of Al-Qadisiyah Palm Vein Identification Based on Hybrid Feature Selection Model. Int. J. Intell. Eng. Syst. 2021, 14, 469–478. [Google Scholar] [CrossRef]
  8. Malini, M.S.; Patil, M. Interpolation Techniques in Image Resampling. Int. J. Eng. Technol. 2018, 7, 567–570. [Google Scholar] [CrossRef]
  9. Munawar, H.S.; Hammad, A.W.A.; Haddad, A.; Soares, C.A.P.; Waller, S.T. Image-Based Crack Detection Methods: A Review. Infrastructures 2021, 6, 115. [Google Scholar] [CrossRef]
  10. Siegel, J.E.; Sun, Y.; Sarma, S. Automotive Diagnostics as a Service: An Artificially Intelligent Mobile Application for Tire Condition Assessment. In Proceedings of the International Conference on AI and Mobile Services, Seattle, DC, USA, 25–30 June 2018; pp. 172–184. [Google Scholar] [CrossRef]
  11. Ozturk, M.; Baran, M.; Latifoğlu, F. Beyond the colors: Enhanced deep learning on invasive ductal carcinoma. Neural Comput. Appl. 2022, 34, 18953–18973. [Google Scholar] [CrossRef]
  12. Bučko, B.; Lieskovská, E.; Zábovská, K.; Zábovský, M. Computer Vision Based Pothole Detection under Challenging Conditions. Sensors 2022, 22, 8878. [Google Scholar] [CrossRef]
  13. Bhalla, K.; Koundal, D.; Bhatia, S.; Rahmani, M.K.I.; Tahir, M. Fusion of Infrared and Visible Images Using Fuzzy Based Siamese Convolutional Network. Comput. Mater. Contin. 2022, 70, 5503–5518. [Google Scholar] [CrossRef]
  14. Das, R.; Piciucco, E.; Maiorana, E.; Campisi, P. Convolutional Neural Network for Finger-Vein-Based Biometric Identification. IEEE Trans. Inf. Forensics Secur. 2018, 14, 360–373. [Google Scholar] [CrossRef] [Green Version]
  15. Ingle, P.Y.; Kim, Y.-G. Real-Time Abnormal Object Detection for Video Surveillance in Smart Cities. Sensors 2022, 22, 3862. [Google Scholar] [CrossRef] [PubMed]
  16. Mansor, N.; Sani, N.S.; Aliff, M. Machine Learning for Predicting Employee Attrition. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 435–445. [Google Scholar] [CrossRef]
  17. Nasif, A.; Othman, Z.; Sani, N. The Deep Learning Solutions on Lossless Compression Methods for Alleviating Data Load on IoT Nodes in Smart Cities. Sensors 2021, 21, 4223. [Google Scholar] [CrossRef]
  18. Jabor, A.H.; Ali, A.H. Dual Heuristic Feature Selection Based on Genetic Algorithm and Binary Particle Swarm Optimization. J. Univ. BABYLON Pure Appl. Sci. 2019, 27, 171–183. [Google Scholar] [CrossRef] [Green Version]
  19. Suwadi, N.A.; Derbali, M.; Sani, N.S.; Lam, M.C.; Arshad, H.; Khan, I.; Kim, K.-I. An Optimized Approach for Predicting Water Quality Features Based on Machine Learning. Wirel. Commun. Mob. Comput. 2022, 2022, 3397972. [Google Scholar] [CrossRef]
  20. Park, H.-J.; Lee, Y.-W.; Kim, B.-G. Efficient Tire Wear and Defect Detection Algorithm Based on Deep Learning. J. Korea Multimed. Soc. 2021, 24, 1026–1034. [Google Scholar]
  21. Wei, T.J.; Bin Abdullah, A.R.; Saad, N.B.M.; Ali, N.B.M.; Tengku Zawawi, T.N.S. Featureless EMG pattern recognition based on convolutional neural network. Indones. J. Electr. Eng. Comput. Sci. 2019, 14, 1291–1297. [Google Scholar] [CrossRef]
  22. Kumar, R.; Arora, R.; Bansal, V.; Sahayasheela, V.J.; Buckchash, H.; Imran, J.; Narayanan, N.; Pandian, G.N.; Raman, B. Classification of COVID-19 from chest x-ray images using deep features and correlation coefficient. Multimed. Tools Appl. 2022, 81, f27631–f27655. [Google Scholar] [CrossRef]
  23. Sohn, I. Deep belief network based intrusion detection techniques: A survey. Expert Syst. Appl. 2020, 167, 114170. [Google Scholar] [CrossRef]
  24. Wang, J.; He, Z.; Huang, S.; Chen, H.; Wang, W.; Pourpanah, F. Fuzzy measure with regularization for gene selection and cancer prediction. Int. J. Mach. Learn. Cybern. 2021, 12, 2389–2405. [Google Scholar] [CrossRef]
  25. Yang, Y.; Zheng, K.; Wu, C.; Niu, X.; Yang, Y. Building an Effective Intrusion Detection System Using the Modified Density Peak Clustering Algorithm and Deep Belief Networks. Appl. Sci. 2019, 9, 238. [Google Scholar] [CrossRef] [Green Version]
  26. Maldonado-Chan, M.; Mendez-Vazquez, A.; Guardado-Medina, R.O. Multimodal Tucker Decomposition for Gated RBM Inference. Appl. Sci. 2021, 11, 7397. [Google Scholar] [CrossRef]
  27. Reza, S.; Amin, O.B.; Hashem, M. A Novel Feature Extraction and Selection Technique for Chest X-ray Image View Classification. In Proceedings of the 2019 5th International Conference on Advances in Electrical Engineering (ICAEE), Dhaka, Bangladesh, 26–28 September 2019; pp. 189–194. [Google Scholar] [CrossRef]
  28. Hussein, I.J.; Burhanuddin, M.A.; Mohammed, M.A.; Benameur, N.; Maashi, M.S.; Maashi, M.S. Fully-automatic identification of gynaecological abnormality using a new adaptive frequency filter and histogram of oriented gradients ( HOG ). Expert Syst. 2021, 39, e12789. [Google Scholar] [CrossRef]
  29. Al-Shammary, D.; Albukhnefis, A.L.; Alsaeedi, A.H.; Al-Asfoor, M. Extended particle swarm optimization for feature selection of high-dimensional biomedical data. Concurr. Comput. Pract. Exp. 2022, 34, e6776. [Google Scholar] [CrossRef]
  30. Chakraborty, C.; Kishor, A.; Rodrigues, J.J. Novel Enhanced-Grey Wolf Optimization hybrid machine learning technique for biomedical data computation. Comput. Electr. Eng. 2022, 99, 107778. [Google Scholar] [CrossRef]
  31. Alkafagi, S.S.; Almuttairi, R.M. A Proactive Model for Optimizing Swarm Search Algorithms for Intrusion Detection System. J. Phys. Conf. Ser. 2021, 1818, 012053. [Google Scholar] [CrossRef]
  32. Sharda, S.; Srivastava, M.; Gusain, H.S.; Sharma, N.K.; Bhatia, K.S.; Bajaj, M.; Kaur, H.; Zawbaa, H.M.; Kamel, S. A hybrid machine learning technique for feature optimization in object-based classification of debris-covered glaciers. Ain Shams Eng. J. 2022, 13, 101809. [Google Scholar] [CrossRef]
  33. Safaeian, M.; Fathollahi-Fard, A.M.; Kabirifar, K.; Yazdani, M.; Shapouri, M. Selecting Appropriate Risk Response Strategies Considering Utility Function and Budget Constraints: A Case Study of a Construction Company in Iran. Buildings 2022, 12, 98. [Google Scholar] [CrossRef]
  34. Zhou, H.; Wang, X.; Zhu, R. Feature selection based on mutual information with correlation coefficient. Appl. Intell. 2021, 52, 5457–5474. [Google Scholar] [CrossRef]
  35. Aliff, M.; Amry, S.; Yusof, M.I.; Zainal, A.; Rohanim, A.; Sani, N.S. Development of Smart Rescue Robot with Image Processing (iROB-IP). Int. J. Electr. Eng. Technol. 2020, 11, 8–19. [Google Scholar]
Figure 1. Proposed scenario of intelligent system for detection of cracked tires.
Figure 1. Proposed scenario of intelligent system for detection of cracked tires.
Symmetry 15 00358 g001
Figure 2. Architecture of deep belief network.
Figure 2. Architecture of deep belief network.
Symmetry 15 00358 g002
Figure 3. Graphical representation of restricted Boltzmann machines.
Figure 3. Graphical representation of restricted Boltzmann machines.
Symmetry 15 00358 g003
Figure 4. Architecture of hybrid adaptive–correlation-based feature selection and deep belief neural networks.
Figure 4. Architecture of hybrid adaptive–correlation-based feature selection and deep belief neural networks.
Symmetry 15 00358 g004
Figure 5. Graphical representation of histogram of oriented gradient features.
Figure 5. Graphical representation of histogram of oriented gradient features.
Symmetry 15 00358 g005
Figure 6. Distribution value of the threshold (α) in the proposed adaptive correlation features selection.
Figure 6. Distribution value of the threshold (α) in the proposed adaptive correlation features selection.
Symmetry 15 00358 g006
Figure 7. Flow of the proposed adaptive correlation-based and variance feature selection.
Figure 7. Flow of the proposed adaptive correlation-based and variance feature selection.
Symmetry 15 00358 g007
Figure 8. Sample of images in the tires dataset.
Figure 8. Sample of images in the tires dataset.
Symmetry 15 00358 g008
Figure 9. Comparison between naïve Bayes (NB), random forest (RF), and decision tree (DT), and deep belief neural network (DBN) in terms of true negative rate (TNR).
Figure 9. Comparison between naïve Bayes (NB), random forest (RF), and decision tree (DT), and deep belief neural network (DBN) in terms of true negative rate (TNR).
Symmetry 15 00358 g009
Figure 10. Comparison between hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN) and deep belief neural network (DBN) in terms of true positives rate (TPR).
Figure 10. Comparison between hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN) and deep belief neural network (DBN) in terms of true positives rate (TPR).
Symmetry 15 00358 g010
Figure 11. Comparison between hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN) and deep belief neural network (DBN) in terms of receiver operating characteristic (ROC).
Figure 11. Comparison between hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN) and deep belief neural network (DBN) in terms of receiver operating characteristic (ROC).
Symmetry 15 00358 g011
Table 1. Comparison between naïve Bayes (NB), random forest (RF), decision tree (DT), and deep belief neural network (DBN).
Table 1. Comparison between naïve Bayes (NB), random forest (RF), decision tree (DT), and deep belief neural network (DBN).
PredictorFeature Selection ModelAccuracyPrecisionRecallF1-Score
NBStand alone 0.6490.8240.5040.402
CFS0.6620.6210.5530.528
ACFS0.6860.6860.5780.558
RFStand alone 0.7860.7040.7340.715
CFS0.7980.7260.7400.732
ACFS0.8090.7400.7470.743
DTStand alone 0.6150.6020.6100.560
CFS0.6390.6240.6280.625
ACFS0.6610.6490.6480.648
ANNStand-alone 0.6710.7260.5410.483
CFS0.6510.6880.5560.508
ACFS0.6820.7690.5930.547
DBNStand alone 0.8160.7980.8320.805
CFS0.8590.8280.8990.842
ACFS0.8900.8720.8830.877
Table 2. Compare the proposed hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN).
Table 2. Compare the proposed hybrid adaptive–correlation-based feature selection—deep belief neural network (H-DBN).
Reference Algorithm NameAccuracy
[10]A smartphone-operable densely connected convolutional neural network for tire condition assessment78%
[20]Efficient tire wear and defect detection algorithm based on deep learning85%
[3]Efficient intelligent system for diagnosis of pneumonia in X-ray images empowered with an initial clustering82%
Proposed H-DBN88%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-juboori, A.M.; Alsaeedi, A.H.; Nuiaa, R.R.; Alyasseri, Z.A.A.; Sani, N.S.; Hadi, S.M.; Mohammed, H.J.; Musawi, B.A.; Amin, M.M. A Hybrid Cracked Tiers Detection System Based on Adaptive Correlation Features Selection and Deep Belief Neural Networks. Symmetry 2023, 15, 358. https://doi.org/10.3390/sym15020358

AMA Style

Al-juboori AM, Alsaeedi AH, Nuiaa RR, Alyasseri ZAA, Sani NS, Hadi SM, Mohammed HJ, Musawi BA, Amin MM. A Hybrid Cracked Tiers Detection System Based on Adaptive Correlation Features Selection and Deep Belief Neural Networks. Symmetry. 2023; 15(2):358. https://doi.org/10.3390/sym15020358

Chicago/Turabian Style

Al-juboori, Ali Mohsin, Ali Hakem Alsaeedi, Riyadh Rahef Nuiaa, Zaid Abdi Alkareem Alyasseri, Nor Samsiah Sani, Suha Mohammed Hadi, Husam Jasim Mohammed, Bashaer Abbuod Musawi, and Maifuza Mohd Amin. 2023. "A Hybrid Cracked Tiers Detection System Based on Adaptive Correlation Features Selection and Deep Belief Neural Networks" Symmetry 15, no. 2: 358. https://doi.org/10.3390/sym15020358

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop