Next Article in Journal
A Communication Framework for Image Transmission through LPWAN Technology
Next Article in Special Issue
Recent Advances in Harris Hawks Optimization: A Comparative Study and Applications
Previous Article in Journal
Dual-Band Circularly Polarized Hybrid Dielectric Resonator Antenna for 5G Millimeter-Wave Applications
Previous Article in Special Issue
Enhanced Credit Card Fraud Detection Model Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Glaucoma Based on Elephant-Herding Optimization Algorithm and Deep Belief Network

by
Mona A. S. Ali
1,2,*,
Kishore Balasubramanian
3,
Gayathri Devi Krishnamoorthy
4,
Suresh Muthusamy
5,
Santhiya Pandiyan
6,
Hitesh Panchal
7,
Suman Mann
8,
Kokilavani Thangaraj
9,
Noha E. El-Attar
10,
Laith Abualigah
11,* and
Diaa Salama Abd Elminaam
12,13,*
1
Computer Science Department, College of Computer Science and Information Technology, King Faisal University, Al Ahsa 400, Saudi Arabia
2
Computer Science Department, Faculty of Computers and Artificial Intelligence, Benha University, Benha 12311, Egypt
3
Department of Electrical and Electronics Engineering, Mahalingam College of Engineering and Technology (Autonomous), Coimbatore 642001, India
4
Department of Electronics and Communication Engineering, NGP Institute of Technology, Coimbatore 641001, India
5
Department of Electronics and Communication Engineering, Kongu Engineering College (Autonomous), Coimbatore 638001, India
6
Department of Computer Science and Engineering, Kongu Engineering College (Autonomous), Coimbatore 638001, India
7
Department of Mechanical Engineering, Government Engineering College, Patan 384265, India
8
Department of Information Technology, Maharaja Surajmal Institute of Technology, New Delhi 110005, India
9
Department of Electrical and Electronics Engineering, College of Engineering and Technology (Autonomous), Coimbatore 641001, India
10
Department of Information System, Faculty of Computers and Artificial Intelligence, Benha University, Benha 12311, Egypt
11
Faculty of Computer Sciences and Informatics, Amman Arab University, Amman 11953, Jordan
12
Information Systems Department, Faculty of Computers and Artificial Intelligence, Benha University, Benha 12311, Egypt
13
Computer Science Department, Faculty of Computer Science, Misr International University, Cairo 12585, Egypt
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(11), 1763; https://doi.org/10.3390/electronics11111763
Submission received: 11 April 2022 / Revised: 17 May 2022 / Accepted: 23 May 2022 / Published: 2 June 2022
(This article belongs to the Special Issue Big Data Analytics Using Artificial Intelligence)

Abstract

:
This study proposes a novel glaucoma identification system from fundus images through the deep belief network (DBN) optimized by the elephant-herding optimization (EHO) algorithm. Initially, the input image undergoes the preprocessing steps of noise removal and enhancement processes, followed by optical disc (OD) and optical cup (OC) segmentation and extraction of structural, intensity, and textural features. Most discriminative features are then selected using the ReliefF algorithm and passed to the DBN for classification into glaucomatous or normal. To enhance the classification rate of the DBN, the DBN parameters are fine-tuned by the EHO algorithm. The model has experimented on public and private datasets with 7280 images, which attained a maximum classification rate of 99.4%, 100% specificity, and 99.89% sensitivity. The 10-fold cross validation reduced the misclassification and attained 98.5% accuracy. Investigations proved the efficacy of the proposed method in avoiding bias, dataset variability, and reducing false positives compared to similar works of glaucoma classification. The proposed system can be tested on diverse datasets, aiding in the improved glaucoma diagnosis.

1. Introduction

Glaucoma is a type of ocular neuropathy which threatens vision if left untreated [1]. It is a gradual condition that develops as the ocular pressure rises [2]. Nearly 80 million people are affected by glaucoma globally [3]. Optic nerves are impaired gradually due to high ocular pressure [4,5]. Approximately 75–80% of the glaucoma cases are identified only at the developed stage, which cannot be cured. One cannot visualize any known symptoms of glaucoma except for narrowing of vision at a later stage [6]. The progression is slow and painful at certain levels [6]. It requires lifelong treatment, and it is impossible to reinstate vision loss. Hence, early detection of glaucoma and treatment stands as the best means of prevention. Concerning the fundamental issue, there is a critical need to build a system that can function well without the need for excessive equipment, qualified medical practitioners, or time. Computer-assisted techniques could help detect the disease at its early stages using advanced machine-learning and deep-learning methods. Trained deep-learning models could take advantage of minor changes, such as retinal layer thinning, that human specialists cannot notice.
Nevertheless, the optic nerve damage precipitated slowly develops, and as symptoms stem from it, the disease advances significantly [7]. Nevertheless, the most up-to-date technology can probably hinder glaucoma development in patients [8,9]. Figure 1 shows sample images of normal and glaucomatous eyes.
To scrutinize various glaucoma traits, ophthalmologists utilize confocal scanning laser ophthalmoscopy (CSLO) [10], Heidelberg retina tomography (HRT), optical coherence tomography (OCT), along with fundus images [11]. For instance, numerous retinal features, optic nerves head (ONH), peripapillary atrophy, and retinal nerve fiber layer, are perceived for glaucoma diagnosis [12]. Assessment of increased IOP, abnormal visual field (V.F.), damaged ONH, etc. are usually investigated for glaucoma [13,14,15]. The OD is split into three disparate areas: the cup (central region), the neuroretinal rim, and parapapillary atrophy [16,17]. The white cup-like structure located in the disc center is the OC. The ratio of OC size to OD size is normally an important measure analyzed in glaucoma diagnosis, denoted as the cup-disc ratio (CDR). The main contributions of this paper can be summarized as follow:
  • Developing an optimized model employing a deep belief network classifier (DBN);
  • Employing modified Wiener filter (MWF), circular Hough transform (CHT), and Otsu’s thresholding for OD and OC segmentation, respectively;
  • Generating a distinct hybrid feature set to aid in diagnosis;
  • Selecting relevant features through the ReliefF algorithm based on predictive importance weights;
  • Fine-tuning DBN by elephant-herding optimization algorithm (EHO);
  • Investigating the model’s robustness to noise such as Gaussian and salt-pepper;
  • Analyzing the isolated and combined feature set contribution in glaucoma identification.
The paper is structured as follows: Section 2 outlines various works related to the proposed method. Section 3 presents the adopted architecture of the method proposed. Experimental outcomes along with dataset preparation are explored in Section 4. The conclusion is elucidated in Section 5.

2. Related Works

A brief literature review based on feature extraction and neural network classifier for glaucoma detection is elucidated in this section. Raja et al. [18] described a statistical feature extraction method based on the hyper analytic wavelet transformation (HWT) in which statistical characteristics were extracted and passed to a support vector machine (SVM). The particle swarm algorithm is used to adjust the HWT and SVM-RB simultaneously to get the optimum fit. Issac, A. et al. [19] and Koh J.E. et al. [20] presented a similar feature extraction strategy for glaucoma classification. Haralick’s features-centered categorization of glaucoma using back propagation neural networks (BPNN) was suggested by Samanta et al. [21]. The results of the experiment revealed good accuracy, sensitivity, and specificity.
Acharya, U. et al. [22] proposed an effective technique that included preprocessing, picture convolving with filter banks, and the chosen features fed into the KNN classifier. Jun et al. [23] demonstrated a super-pixel-based OD and OC segmentation model for diagnosing glaucoma. The unreliable results exhibited scope for improvement. Gift and Nirmal [24] suggested gray wolf optimized NN produce enhanced accuracy for glaucoma detection through a sequence of steps. Preprocessing, image normalization and feature extraction were done, and the features were given to the GWO-NN classifier. Anushikha et al. [25] presented an automated diagnostic system utilizing wavelet features from the segmented OD, which was extorted for analysis in addition to classification. Experimental outcomes signified an accuracy of 0.947. Ajesh et al. [26] reported a new multi-feature extraction approach for glaucoma identification and classification by integrating discrete wavelet transform (DWT) and ML algorithms that produced an accuracy of 95% in glaucoma identification.
For fine-imaging analysis, DWT was computationally intensive. Diaz-Pinto et al. [27] provided an automatic technique using retinal structural features and Luv color space for OD and OC segmentation, obtaining 81% specificity and 87% sensitivity. Studies have reported numerous methodologies for recognizing glaucoma through CAD. They use either conventional machine learning, deep learning, or both. Several works related to the diagnosis of glaucoma using CNN have been demonstrated in the literature [28,29,30,31,32]. The CNN employed was constructed from scratch, and different datasets were used to investigate the models. Data augmentation via rotation, random flip, image translation, etc. was performed to increase the dataset size artificially. OCT images were also used to segment the retinal vasculature apart from the fundus images. The deep networks were employed to extract and learn the layer properties of the retina using a pretrained backbone network and reinforcement learning. A multiscale feature generated could then be used on modules, such as the encoder–decoder, to retrieve retinal information and capture finer retinal boundaries. Many optimization algorithms are also used to solve the various optimization problems. A blended approach is adopted in our method, where the features extracted are selected and passed to an optimized deep network for classification.

3. Proposed Method

This paper reports using the DBN classifier, a robust, novel, and efficient glaucoma detection method on retinal fundus images (RFI). Input is acquired from fundus databases and preprocessed for noise removal using MWF. Utilizing CHT, the OD is then segmented from the noise-removed image. The OC is cropped from the OD image using Otsu’s thresholding algorithm. Structural and functional features are extorted and fed to the DBN classifier. The EHO algorithm optimizes DBN’s parameters to obtain a minimum cost function and the best solution for healthy and glaucomatous classification. Figure 2 illustrates the schematic diagram of the system.

3.1. Preprocessing

The fundus image collected must be preprocessed to highlight the vessel and other morphological features. Given the benefits of grayscale images, RBG is converted to grayscale first [33], then filtered to remove background noise [34]. A modified Wiener filter is employed to prevent impulse noise while preserving the edges [35]. The noise reduction aims to protect crucial structural content for disease detection, besides determining noisy pixels [36].

3.2. OD and OC Segmentation

OD segmentation is done on the image after noise removal. First, the OD location is identified. Subsequently, edges are calculated through the Canny edge filter. As the OD in the retina is a circular object, the circle detection Hough transform (CHT) is applied to identify the area. The boundary positions along with the region of the OD are obtained. The ROI termed OD is cropped from the original image for further OC segmentation. The OC is segmented from pre-segmented OD using Otsu’s thresholding [37].

3.3. Feature Extraction

Structural, textural, and intensity-based features contributing to the disease diagnosis are extorted in this phase [38]. Table 1 lists the set of features generated through this process.
Using the procedure described above, the 111 features extracted are then selected using the relief F algorithm before being fed to DBN for classification.

3.4. Feature Selection

This stage is critical because rarely discernible characteristics are deleted, putting the modeling process under more computational strain. In this work, the ReliefF algorithm is preferred for dimensionality reduction due to its promising results, as explained in [38]. ReliefF ranks features according to their weight participation, with the most active features being placed first. As illustrated in Figure 3, other features contribute far less to the last features. As a result, we can choose the 15 most potent features based on their weight and exclude characteristics that add to the model’s computational cost.
ReliefF chooses T, some instances at random, but subsequently, k searches for the closest same-class neighbors, which are referred to as the nearest hit values of H. The k-nearest neighbors are the one–one scores among the multiple classes, also known as the M(T) nearest misses. The number of nearest neighbors is set at three in our study. The 111 features are reduced to 15 optimal features based on their participation at the top of the weighted list for the highest accuracy using the procedure described above. The remaining ones are ignored because there is not much difference in output, which increases the computing weight of a model. Figure 4 depicts cumulative accuracy vs. a number of features, and 15 contribute more than the total variations.

3.5. Classification

3.5.1. Deep Belief Networks (DBN)

The DBN class of neural network (NN) can be considered a generative model that uses a set of Boltzmann machines as basic building elements [39]. Each layer of the DBNs has a restricted Boltzmann machine (RBM). DBN extracts H.L. features from the data slated for training to improve the between-classes separation power. The training is performed on all the layers through supervised mode, and the backward propagation mode modifies the weight in the network to reduce over-fitting. This work develops a DBN model trained using greedy layer-wise learning [40] by stacking up RBMs, as shown in Figure 5.
RBM concentrates on a particular layer during its learning procedure and ignores others. We assume we have a DBN with L levels, with Wi being the RBM’s weight matrix at layer i. In addition, the hidden units at the ith layer form the layer (i + 1) input unit. In the model proposed, the set of weight matrices is assigned as W = {W1, W2, W3} and the set of hidden layers as h = {h1, h2, h3}. The weight matrix between layers ith and (i + 1), which is denoted by Wi, while the jth hidden layer is denoted by hj. The following energy function was used to compute the combined distribution of the hidden and visible layers:
p ( v , h ) = e E ( v , h ) v , h e E ( v , h )
where E ( v , h ) denotes RBM’s energy function,
E ( v , h ) = i = 1 a i v i j = 1 b j h j i , j v i h j W i j
W i j denotes the weight between the visible and the hidden layer, a i and b j describe the visible and hidden layer coefficients. This study uses the stochastic gradient descent (SGD) method following log-likelihood (L.L.) to accomplish optimal training. This is accomplished by optimizing the RBM’s parameters a, b, and wij. The derivatives of the log p ( v , h ) w.r.t W i j , a i and b j must be computed to update the weights and biases. The resulting equations are
W t + 1 = W t + η ( p ( h v ) v T p ( h v ) v T ) λ W t + α Δ W t 1
a t + 1 = a t + η ( v v ˜ ) + α Δ a t 1
b t + 1 = b t + η ( p ( h v ) p ( h ˜ v ˜ ) ) + α Δ b t 1
where p ( h j = 1 v ) = σ ( i = 1 m w i j v i + b j ) , p ( v i = 1 h ) = σ ( j = 1 n w i j h j + a i ) , and σ(·) represents the logistic sigmoid function. v ˜ and h ˜ denote the reconstructed v and h, respectively. N is the number of hidden nodes, η, the learning ratio, α, the momentum weight, and λ, the weight decay. The weight matrix and accompanying bias vectors of the visible and hidden nodes are learned using contrastive divergence (CD) and persistent contrastive divergence (PCD). This optimization process uses the BP with the conventional gradient ascent algorithm to tune the weight matrices to optimal values. The optimization algorithm considers the outcome of an extra layer built over the DBN after its previous greedy training to minimize some error metrics. Softmax, or logistic units, are frequently used in this layer.

3.5.2. Elephant Herd Optimization (EHO) Algorithm

The EHO algorithm was introduced by Wang et al. in 2015 [41]. Elephants behave socially and encompass a complex structure of calves and females. An elephant group comprises numerous clans with a matriarch as the leader and her calves or other related females. A female forms a clan. EHO concerns the succeeding assumptions.
  • The elephant group is classified into clans, and each such clan comprises specific elephants.
  • A specific number of male elephants (ME) depart their clan to live independently.
  • Each clan has a leader termed the matriarch.
The matriarch group keeps the best solution in the elephant herd. The entire elephant population is divided into j clans. Matriarch ci influences the new position of each elephant ci. The elephant j in clan ci can be calculated using
x n e w ,   c i , j = x c i j + a × ( x b e s t ,   c i x c i , j ) × r
where x n e w ,   c i , j indicated the new position and x c i j denotes the old position for elephant j in the clan ci. x b e s t ,   c i represents matriarch c i , which denotes the best elephant. a ∈ [0,1] shows a scaling factor, r   [0,1]. The best elephant is computed for each clan using
x n e w , c i , j = β × x center ,   , i
Here, β     [ 0 , 1 ] indicates the second parameter that guides the impact of the x c e n t e r , c i , d delineated in
x c e n t e r , c i , d = 1 n c i × j = 1 n c i x c i , j , d
where 1 d D , and nci represent the number of elephants in clan x c i , j , d is the dth dimension of individual elephant x c i , j , d the center of clan ci ( x c e n t e r , c i , d ) can be updated (Equation (8)). The separating process could be modelled as a separation operator when tackling optimization issues. In each clan, the worst valued elephants are moved to the next position indicated by
x w o r s t ,   d = x m i n + ( x m a x x m i n + 1 ) × r a n d
Here the lower and upper bands of the search space are indicated by x m i n and x m a x , respectively. r a n d   [0,1] signifies the random value picked from the uniform distribution.
The EHO algorithm was examined for various benchmark set functions and in medical diagnosis [42,43,44,45,46], showing better results. This study employs the EHO algorithm for DBN parameter optimization. The output of the DBN model is grounded in weights and the biases of preceding layers in the network. EHO does not employ the previous individuals in the later updating process as other optimization algorithms. EHO is a swarm-inspired algorithm that deals with global optimization missions characterized by clan updating and searching operations. EHO does not resort to relaxation and is less vulnerable to noise. They perform better in constrained, optimized environments. High convergence rate and low localization errors with less execution time are the important characteristics of EHO. The algorithm can tackle non-convex ML problems directly.

3.5.3. Fine Tuning of DBM

The learning rate, hidden units, momentum weight, and weight decay are the four basic parameters set up in most RBMs. The use of traditional methods for computing the error function is an NP-hard problem due to its complexity and differentiation. Meta-heuristics have been employed to solve this issue. The EHO algorithm is used to optimize the DBN training by fine tuning the parameters in this work. Here, the parameters set are
n   [ 5 , 100 ] ,   η [ 0.1 , 0.9 ] ,   λ [ 0.1 , 0.9 ]   a n d   α [ 0.00001 , 0.01 ]
A fitness function must be designed to steer the searching process to attain the best answers to meet the objectives. Mean squared error (MSE) is adopted as the fitness function. It measures the error between the output and the desired value and is given by
M S E = 1 T j = 1 N i = 1 T ( D j ( i ) Y j ( i ) ) 2
where T—data number, N—number of the output layers.
The EHO looks for a collection of DBN parameters that minimizes MSE. D j ( i ) denotes the value from the jth unit in the DBN’s output layer at the time ‘t’, Y j ( i ) represents the jth factor of the desired value. The process is repeated until the halting criteria is met.
The optimization steps of EHO are as follows:
  • Set the EHO parameters and initialize the population.
  • Evaluate the individual fitness value (RMSE) of the DBN, as per the learning rate and the number of batch learning. Identify the optimal individual.
  • Check if the termination condition is reached; if so, end the iteration and output the result; or else, go to the next step.
  • Update each individual position. Reinitialize the individuals beyond the lower and upper limits.
  • Start a new iteration by updating the optimal individual.

4. Results and Discussion

Evaluation of the proposed model’s performance is presented in this section. The experiment is executed in MATLAB with the following specifications: the Intel Core i7 Processor, Windows 10, 3.20 GHz CPU speed, and 4GB RAM.

4.1. Dataset Preparation

This work uses DRISHTI–GS1, ACRIMA, ORIGA-Light, and LAG datasets for evaluation. The images were captured utilizing a Canon CR-1 fundus camera at 2336 × 3504 resolution with a 45° FOV and a disparate acquisition setting. The dataset used in the methodology is from public and private datasets annotated by an ophthalmologist who has over 15 years of experience in the field. The list of databases is depicted in Table 2. A total of 7280 images obtained from various public and private databases are used for investigating the proposed system’s performance after eliminating a few irrelevant images. Images are trained and tested in the ratio of 70:30, respectively.
Images that are preprocessed are subjected to OD and OC segmentation. Sample images and results of the segmentation are given below in Figure 6.

4.2. Performance Analysis

Experiments are done to show the efficacy of the DBN–EHO on different fundus datasets. Initially, the investigations are performed to show the efficacy of EHO in optimizing the DBN. EHO is compared with meta-heuristic algorithms, such as artificial bee colony (ABC), firefly algorithm (FA), harmony search (HS), cuckoo search (CS), particle swarm optimization algorithms (PSO), and differential evolution algorithms (DE). A hold-out technique with 20 training and test sets generated at random, subjected to 10 iterations for each RBM learning procedure, and a mini-batch of 20 are performed for consistent comparisons with other works. Five agents of over 50 iterations are employed to achieve convergence using all strategies in tests. Control parameters for all the algorithms are outlined in Table 3.
In Table 4, the MSE values of each algorithm on the original LAG dataset considering DBN. A Wilcoxon signed-rank test with a 0.05 significance level is utilized to analyze the method statistically. It is observed from Table 4 that EHO outperformed all the other algorithms in terms of the lowest MSE employing fewer layers. ABC and FA algorithms have also performed well next to EHO.
Taking the performance of EHO, the performance of the proposed work is assessed by finding ‘true positive’ (TP), ‘true negative’ (TN), ‘false positive’ (FP), and ‘false negative’ (FN) values. TP denotes the instances wherein glaucoma is detected correctly. TN shows the condition wherein a person with no glaucoma is classified correctly. FP indicates the number of negative instances recognized as positive. Positive instances recognized as negative are indicated by FN. Precision, accuracy, F-score, specificity, recall, sensitivity, and MCC (Mathew’s correlation coefficient) are used in this work, as in Table 5.
The performance of the proposed work on individual datasets is provided in Table 6. It is inferred that a maximum accuracy of 99.34% on the ACRIMA dataset is attained, followed by 99.31% on the LAG dataset. The classification rate is between 96.95% and 99.34%, ensuring that the DBN–EHO performed better in all the datasets. This indicates that the images in the set are captured under different illuminations. Specificity of 100% on the LAG dataset shows that the model can reduce false positive rates.
The impact of isolated and combined feature sets on the diagnosis of glaucoma in the LAG dataset is depicted in Table 7. The extraction of isolated features yielded an enhanced result, increasing the algorithm’s efficiency. When the features are combined, an accuracy of 99.3% is obtained. When the features are combined, better accuracy of 99.34% is obtained. The feature contribution indicates that the DBN optimization also contributes to appreciable performance as the accuracy ranges from 95% to 99% on different kinds of features.
The feature contribution indicates that the DBN optimization also contributes to appreciable performance as the accuracy ranges from 95% to 99% on different kinds of features. A 10-fold cross validation done to reduce bias during testing enhances the algorithm’s robustness and reduces the classifier’s misclassification rate. From Table 8, it is seen that the cross-validation accuracy across the datasets is appreciably high, ensuring that the model is free from bias. The model applies to a wide range of datasets, compensating for any imbalance. Figure 7 highlights the proposed work against various performance metrics.
A 10-fold CV is carried out to reduce bias during testing, and the results are shown in Table 8.
From Table 8, it is seen that the cross-validation accuracy across the datasets is appreciably high, ensuring that the model is free from bias. The model applies to a wide range of datasets, compensating for any imbalance.
Table 9 reports the outcomes of the classifier when compared with similar conventional ML classifiers, such as K-nearest neighbor (KNN), random forest (RF), support vector machine (SVM), and DBN without optimization. The DBN attains a more appreciable performance than the conventional ML classifiers. The DBN, when optimized for weight by EHO, still achieves better performance across all the datasets.
From Table 9, it is seen that the DBN classifier attains a more appreciable performance than the conventional ML classifiers. The DBN, when optimized for weight by EHO, still achieves better performance across all the datasets. Furthermore, to assess the robustness of the model, salt-pepper and Gaussian noise are added to the LAG dataset (original image set). Gaussian noise is predominant if the images were captured under low illumination. Salt-pepper noise is an impulse noise occurring owing to intense and sparse disturbances. Figure 8 depicts the original and noise-added image (sample). Experimentation is performed with Gaussian noise and salt-pepper noise, with the variance (σ) and noise density (d) varying from 0.1 to 0.5, respectively, and the result is reported in Figure 9.
It is seen from Figure 9 that the accuracy remains fairly the same. This demonstrates that the suggested model is extremely robust in both original and degraded datasets.
Table 10 compares the proposed model with well-known CNN models employing transfer learning. Transfer learning is applied to the DRISHTI-GS1 dataset, as the number of images is small. It is inferred that our model can work well compared to well-known, pre-trained models using transfer learning. Table 11 illustrates a comparison of similar other techniques and our technique in diagnosing glaucoma. Many optimization algorithms are also used to solve the other various optimization problems [48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64].

5. Conclusions

Glaucoma is a class of ocular neuropathy wherein the optic nerve gets vandalized, resulting in permanent vision loss. Glaucoma detection in RFI using the DBN classifier is proposed in this work. Input images from different public databases are enhanced through the preprocessing phase. The OD and OC are then segmented using CHT and Otsu’s thresholding. The various structural, intensity, and textural features are extorted from the segmented OC and OD images and fed to the DBN classifier optimized by the EHO algorithm. To investigate the performance, the technique is compared with classifier techniques, such as R.F., KNN, SVM, pre-trained CNNs, etc. Experimental outcomes exhibit the performance of the DBN classifier in recognizing the absence or presence of glaucoma accurately compared to other approaches. The proposed work concentrates on some of the important features contributing to glaucoma disease. The ReliefF algorithm selects the features before feeding them to the classifier for classifying them into either healthy or glaucomatous. The classifier performance is improved through an optimization technique where EHO does the weight update process. Dataset imbalance is also minimized as the model showed better training accuracy when images were selected randomly from each set using 10-fold cross validation. One potential drawback of this method is that it is unclear whether the attained specificities and sensitivities will be generalizable to real-world patient populations with common comorbidities, such as cataracts and surface ocular disease, which can degrade the input image quality. The performance of the computational hardware needs to be improved, along with network structure refinement and data dimension reduction, to attain competitively better computational speed. In the future, a universal domain adaptation method for various datasets (both public and private) is needed to be developed using hybrid weighted deep adversarial learning [74] and adaptive on-line validation [75]. In the future, the detection of glaucoma systems will be enhanced using the feature selection and ranking phase with different hybrid optimization algorithms. Besides, granular computing will be embedded in deep neural networks (granulated CNN) to enhance the computation speed significantly.

Author Contributions

Conceptualization, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; methodology, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; software, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; validation K.B., G.D.K., S.M. (Suresh Muthusamy), S.P., H.P., S.M. (Suman Mann) and K.T.; formal analysis, K.B., G.D.K., S.M. (Suresh Muthusamy), S.P., H.P., S.M. (Suman Mann) and K.T.; investigation, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; resources, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; data curation, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; writing—original draft preparation, K.B., G.D.K., S.M. (Suresh Muthusamy), S.P., H.P., S.M. (Suman Mann) and K.T.; writing—review and editing, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; visualization, M.A.S.A., D.S.A.E., N.E.E.-A. and L.A.; project administration, D.S.A.E.; funding acquisition, M.A.S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This Research is funded by Deanship of Scientific Research, King Faisal University, Saudi Arabia [Project No. GRANT716].

Acknowledgments

This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia [Project No. GRANT716].

Conflicts of Interest

The authors declare no potential conflicts of interest and funding.

References

  1. Quigley, H.A.; Broman, A.T. The number of people with glaucoma worldwide in 2010 and 2020. Br. J. Ophthalmol. 2006, 90, 262–267. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Dervisevic, E.; Pavlja, S.; Dervisevic, A.; Kasumovic, S.S. Challenges in early glaucoma detection. Med. Arch. 2017, 70, 203–207. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Kassebaum, N.J.; Bertozzi-Villa, A.; Coggeshall, M.S.; A Shackelford, K.; Steiner, C.; Heuton, K.R.; Gonzalez-Medina, D.; Barber, R.; Huynh, C.; Dicker, D.; et al. Global, regional, and national levels and causes of maternal mortality during 1990–2013: A systematic analysis for the Global Burden of Disease Study 2013. Lancet 2014, 384, 980–1004. [Google Scholar] [CrossRef] [Green Version]
  4. What Is Glaucoma? Available online: https://www.glaucoma.org/glaucoma/optic-nerve-cupping.php (accessed on 1 May 2021).
  5. McMonnies, C.W. Intraocular pressure and glaucoma: Is physical exercise beneficial or a risk? J. Optom. 2016, 9, 139–147. [Google Scholar] [CrossRef] [Green Version]
  6. Types of Glaucoma. Available online: http://www.glaucoma-association.com/about-glaucoma/types-of-glaucoma/chronic-glaucoma (accessed on 1 May 2021).
  7. Blindness. Available online: https://www.who.int/health-topics/blindness-and-vision-loss (accessed on 2 May 2021).
  8. Chandrika, S.; Nirmala, K. Analysis of CDR detection for glaucoma diagnosis. Int. J. Eng. Res. Appl. 2014, 2, 23–27. [Google Scholar]
  9. Al-Bander, B.; Al-Nuaimy, W.; Al-Taee, M.; Zheng, Y. Automated glaucoma diagnosis using deep learning approach. In Proceedings of the 14th International Multi-Conference on Systems, Signals & Devices (SSD), Marrakech, Morocco, 28–31 March 2017; pp. 207–210. [Google Scholar]
  10. Mardin, C.Y.; Horn, F.K.; Jonas, J.B.; Budde, W.M. Preperimetric glaucoma diagnosis by confocal scanning laser tomography of the optic disc. Br. J. Ophthalmol. 1999, 83, 299–304. [Google Scholar] [CrossRef] [Green Version]
  11. Adhi, M.; Duker, J.S. Optical coherence tomography—Current and future applications. Curr. Opin. Ophthalmol. 2013, 24, 213–221. [Google Scholar] [CrossRef] [Green Version]
  12. Septiarini, A.; Khairina, D.M.; Kridalaksana, A.H.; Hamdani, H. Automatic Glaucoma Detection Method Applying a Statistical Approach to Fundus Images. Healthc. Inform. Res. 2018, 24, 53–60. [Google Scholar] [CrossRef] [Green Version]
  13. Thorat, S.; Jadhav, S. Optic disc and cup segmentation for glaucoma screening based on super pixel classification. Int. J. Innov. Adv. Comput. Sci. 2015, 4, 167–172. [Google Scholar]
  14. Kavitha, K.; Malathi, M. Optic disc and optic cup segmentation for glaucoma classification. Int. J. Adv. Res. Comput. Sci. Technol. 2014, 2, 87–90. [Google Scholar] [CrossRef]
  15. Manju, K.; Sabeenian, R.S. Robust CDR calculation for glaucoma identification. Biomed. Res. 2018. [Google Scholar] [CrossRef] [Green Version]
  16. Mahalakshmi, V.; Karthikeyan, S. Clustering based optic disc and optic cup segmentation for glaucoma detection. Int. J. Innov. Res. Comput. Commun. Eng. 2014, 2, 3756–3761. [Google Scholar]
  17. Almazroa, A.; Burman, R.; Raahemifar, K.; Lakshminarayanan, V. Optic Disc and Optic Cup Segmentation Methodologies for Glaucoma Image Detection: A Survey. J. Ophthalmol. 2015, 2015, 180972. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Raja, C.; Gangatharan, N. A Hybrid Swarm Algorithm for optimizing glaucoma diagnosis. Comput. Biol. Med. 2015, 63, 196–207. [Google Scholar] [CrossRef] [PubMed]
  19. Issac, A.; Sarathi, M.P.; Dutta, M.K. An adaptive threshold based image processing technique for improved glaucoma detection and classification. Comput. Methods Programs Biomed. 2015, 122, 229–244. [Google Scholar] [CrossRef]
  20. Koh, J.E.W.; Ng, E.Y.K.; Bhandary, S.V.; Laude, A.; Acharya, U.R. Automated detection of retinal health using PHOG and SURF features extracted from fundus images. Appl. Intell. 2017, 48, 1379–1393. [Google Scholar] [CrossRef]
  21. Samanta, S.; Ahmed, S.K.; Salem, M.A.; Nath, S.S.; Dey, N.; Chowdhury, S.S. Haralick features based automated glaucoma classification using back propagation neural network. In Proceedings of the 3rd International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA), Bhubaneswar, India, 14–15 November 2014; Springer: Cham, Switzerland, 2015; pp. 351–358. [Google Scholar]
  22. Cheng, J.; Liu, J.; Xu, Y.; Yin, F.; Wong, D.W.; Tan, N.M.; Tao, D.; Cheng, C.Y.; Aung, T.; Wong, T.Y. Superpixel classification based optic disc and optic cup segmentation for glaucoma screening. IEEE Trans. Med. Imaging 2013, 32, 1019–1032. [Google Scholar] [CrossRef]
  23. Singh, A.; Dutta, M.K.; ParthaSarathi, M.; Uher, V.; Burget, R. Image processing based automatic diagnosis of glaucoma using wavelet features of segmented optic disc from fundus image. Comput. Methods Programs Biomed. 2016, 124, 108–120. [Google Scholar] [CrossRef]
  24. Ajesh, F.; Ravi, R.; Rajakumar, G. Early diagnosis of glaucoma using multi-feature analysis and DBN based classification. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 4027–4036. [Google Scholar] [CrossRef]
  25. Diaz, A.; Morales, S.; Naranjo, V.; Alcocer, P.; Lanzagorta, A. Glaucoma diagnosis by means of optic cup feature analysis in color fundus images. In Proceedings of the 24th European Signal Processing Conference (EUSIPCO), Budapest, Hungary, 29 August–2 September 2016; pp. 2055–2059. [Google Scholar]
  26. Gómez-Valverde, J.J.; Antón, A.; Fatti, G.; Liefers, B.; Herranz, A.; Santos, A.; Sánchez, C.I.; Ledesma-Carbayo, M.J. Automatic glaucoma classification using color fundus images based on convolutional neural networks and transfer learning. Biomed. Opt. Express 2019, 10, 892–913. [Google Scholar] [CrossRef] [Green Version]
  27. Diaz-Pinto, A.; Morales, S.; Naranjo, V.; Köhler, T.; Mossi, J.M.; Navea, A. CNNs for automatic glaucoma assessment using fundus images: An extensive validation. Biomed. Eng. Online 2019, 18, 29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Raghavendra, U.; Fujita, H.; Bhandary, S.V.; Gudigar, A.; Tan, J.H.; Acharya, U.R. Deep convolution neural network for accurate diagnosis of glaucoma using digital fundus images. Inf. Sci. 2018, 441, 41–49. [Google Scholar] [CrossRef]
  29. Orlando, J.I.; Prokofyeva, E.; del Fresno, M.; Blaschko, M.B. Convolutional neural network transfer for automated glaucoma identification. In Proceedings of the 12th International Symposium on Medical Information Processing and Analysis, Tandil, Argentina, 5–7 December 2016. [Google Scholar]
  30. Chen, X.; Xu, Y.; Wong, D.W.; Wong, T.Y.; Liu, J. Glaucoma detection based on deep convolutional neural network. In Proceedings of the 2015 37th annual international conference of the IEEE engineering in medicine and biology society (EMBC), Milan, Italy, 25–29 August 2015; pp. 715–718. [Google Scholar]
  31. Staal, J.; Abràmoff, M.D.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef]
  32. Hani, A.F.; Soomro, T.A.; Fayee, I.; Kamel, N.; Yahya, N. Identification of noise in the fundus images. In Proceedings of the 2013 IEEE International Conference on Control System, Computing and Engineering, Penang, Malaysia, 29 November–1 December 2013; pp. 191–196. [Google Scholar]
  33. Nagu, M.; Shanker, N. Image De-Noising by Using Median Filter and Weiner Filter. Int. J. Innov. Res. Comput. Commun. Eng. 2014, 2, 5641–5649. [Google Scholar]
  34. Aliskan, A.; Çevik, U. An Efficient Noisy Pixels Detection Model for C.T. Images using Extreme Learning Machines. Teh. Vjesn. —Tech. Gaz. 2018, 25, 679–686. [Google Scholar] [CrossRef]
  35. Raj, P.A.; George, A. FCM and Otsu’s Thresholding based Glaucoma Detection and its Analysis using Fundus Images. In Proceedings of the 2nd International Conference on Intelligent Computing, Instrumentation and Control Technologies (ICICICT), Kannur, India, 5–6 July 2019; pp. 753–757. [Google Scholar]
  36. Urbanowicz, R.J.; Meeker, M.; La Cava, W.; Olson, R.S.; Moore, J.H. Relief-based feature selection: Introduction and review. J. Biomed. Inform. 2018, 85, 189–203. [Google Scholar] [CrossRef]
  37. Rosa, G.; Papa, J.; Costa, K.; Passos, L.; Pereira, C.; Yang, X. Learning Parameters in Deep Belief Networks Through Firefly Algorithm. In Proceedings of the IAPR Workshop on Artificial Neural Networks in Pattern Recognition. In Artificial Neural Networks in Pattern Recognition. ANNPR 2016, Ulm, Germany, 28–30 September 2016; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2016; pp. 138–149. [Google Scholar]
  38. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  39. Wang, G.-G.; Deb, S.; Coelho, L.d.S. Elephant Herding Optimization. In Proceedings of the 3rd International Symposium on Computational and Business Intelligence (ISCBI 2015), Bali, Indonesia, 7–8 December 2015; pp. 1–5. [Google Scholar]
  40. Li, J.; Lei, H.; Alavi, A.H.; Wang, G.-G. Elephant Herding Optimization: Variants, Hybrids, and Applications. Mathematics 2020, 8, 1415. [Google Scholar] [CrossRef]
  41. Nayak, M.; Das, S.; Bhanja, U.; Senapati, M.R. Elephant herding optimization technique based neural network for cancer prediction. Inform. Med. Unlocked 2020, 21, 100445. [Google Scholar] [CrossRef]
  42. Sivaswamy, J.; Krishnadas, S.R.; Joshi, G.D.; Jain, M.; Tabish, A.U.S. Drishti-GS: Retinal image dataset for optic nerve head (ONH) segmentation. In Proceedings of the IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April–2 May 2014; pp. 53–56. [Google Scholar]
  43. Khan, S.M.; Liu, X.; Nath, S.; Korot, E.; Faes, L.; Wagner, S.K.; Keane, P.A.; Sebire, N.J.; Burton, M.J.; Denniston, A.K. A global review of publicly available datasets for ophthalmological imaging: Barriers to access, usability, and generalisability. Lancet. Digit. Health 2021, 3, e51–e66. [Google Scholar] [CrossRef]
  44. Zhang, Z.; Yin, F.S.; Liu, J.; Wong, W.K.; Tan, N.M.; Lee, B.H.; Cheng, J.; Wong, T.Y. ORIGA(light): An on-line retinal fundus image database for glaucoma analysis and research. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, Buenos Aires, Argentina, 31 August–4 September 2020; pp. 3065–3068. [Google Scholar]
  45. Li, L.; Xu, M.; Wang, X.; Jiang, L.; Liu, H. Attention Based Glaucoma Detection: A Large-Scale Database and CNN Model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10563–10572. [Google Scholar]
  46. Karthikeyan, S.; Rengarajan, N. Performance Analysis of Gray Level Co-Occurrence Matrix Texture Features for Glaucoma Diagnosis. Am. J. Appl. Sci. 2014, 11, 248–257. [Google Scholar] [CrossRef]
  47. Mookiah, M.R.; Acharya, U.R.; Lim, C.M.; Petznick, A.; Suri, J.S. Data mining technique for automated diagnosis of glaucoma using higher order spectra and wavelet energy features. Knowl.-Based Syst. 2012, 33, 73–82. [Google Scholar] [CrossRef]
  48. Danandeh Mehr, A.; Rikhtehgar Ghiasi, A.; Yaseen, Z.M.; Sorman, A.U.; Abualigah, L. A novel intelligent deep learning predictive model for meteorological drought forecasting. J. Ambient. Intell. Humaniz. Comput. 2022, 24, 1–5. [Google Scholar] [CrossRef]
  49. Gharaibeh, M.; Almahmoud, M.; Ali, M.Z.; Al-Badarneh, A.; El-Heis, M.; Abualigah, L.; Altalhi, M.; Alaiad, A.; Gandomi, A.H. Early Diagnosis of Alzheimer’s Disease Using Cerebral Catheter Angiogram Neuroimaging: A Novel Model Based on Deep Learning Approaches. Big Data Cogn. Comput. 2021, 6, 2. [Google Scholar] [CrossRef]
  50. Gandomi, A.H.; Chen, F.; Abualigah, L. Machine Learning Technologies for Big Data Analytics. Electronics 2022, 11, 421. [Google Scholar] [CrossRef]
  51. Houssein, E.H.; Hassaballah, M.; Ibrahim, I.E.; AbdElminaam, D.S.; Wazery, Y.M. An automatic arrhythmia classification model based on improved Marine Predators Algorithm and Convolutions Neural Networks. Expert Syst. Appl. 2022, 187, 115936. [Google Scholar] [CrossRef]
  52. Houssein, E.H.; Abdelminaam, D.S.; Hassan, H.N.; Al-Sayed, M.M.; Nabil, E. A Hybrid Barnacles Mating Optimizer Algorithm with Support Vector Machines for Gene Selection of Microarray Cancer Classification. IEEE Access 2021, 9, 64895–64905. [Google Scholar] [CrossRef]
  53. Houssein, E.H.; Abdelminaam, D.S.; Ibrahim, I.E.; Hassaballah, M.; Wazery, Y.M. A Hybrid Heartbeats Classification Approach Based on Marine Predators Algorithm and Convolution Neural Networks. IEEE Access 2021, 9, 86194–86206. [Google Scholar] [CrossRef]
  54. Elminaam, D.A.; Ibrahim, S.A. Building a robust heart diseases diagnose intelligent model based on RST using lem2 and modlem2. In Proceedings of the 32nd IBIMA Conference, Seville, Spain, 15–16 November 2018; pp. 5733–5744. [Google Scholar]
  55. Abd Elminaam, D.S.; Elashmawi, W.H.; Ibraheem, S.A. HMFC: Hybrid MODLEM-Fuzzy Classifier for Liver Diseases Diagnose. Int. Arab. J. E Technol. 2019, 5, 100–109. [Google Scholar]
  56. Nadimi-Shahraki, M.H.; Taghian, S.; Mirjalili, S.; Ewees, A.A.; Abualigah, L.; Elaziz, M.A. MTV-MFO: Multi-Trial Vector-Based Moth-Flame Optimization Algorithm. Symmetry 2021, 13, 2388. [Google Scholar] [CrossRef]
  57. Nadimi-Shahraki, M.H.; Taghian, S.; Mirjalili, S.; Abualigah, L.; Elaziz, M.A.; Oliva, D. EWOA-OPF: Effective Whale Optimization Algorithm to Solve Optimal Power Flow Problem. Electronics 2021, 10, 2975. [Google Scholar] [CrossRef]
  58. Nadimi-Shahraki, M.H.; Fatahi, A.; Zamani, H.; Mirjalili, S.; Abualigah, L. An Improved Moth-Flame Optimization Algorithm with Adaptation Mechanism to Solve Numerical and Mechanical Engineering Problems. Entropy 2021, 23, 1637. [Google Scholar] [CrossRef] [PubMed]
  59. Elminaam, D.S.A.; Neggaz, N.; Ahmed, I.A.; Abouelyazed, A.E.S. Swarming Behavior of Harris Hawks Optimizer for Arabic Opinion Mining. Comput. Mater. Contin. 2021, 69, 4129–4149. [Google Scholar] [CrossRef]
  60. AbdElminaam, D.S.; Neggaz, N.; Gomaa, I.A.E.; Ismail, F.H.; Elsawy, A. AOM-MPA: Arabic Opinion Mining using Marine Predators Algorithm based Feature Selection. In Proceedings of the 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt, 26–27 May 2021; pp. 395–402. [Google Scholar] [CrossRef]
  61. Shaban, H.; Houssein, E.H.; Pérez-Cisneros, M.; Oliva, D.; Hassan, A.Y.; Ismaeel, A.A.K.; AbdElminaam, D.S.; Deb, S.; Said, M. Identification of Parameters in Photovoltaic Models through a Runge Kutta Optimizer. Mathematics 2021, 9, 2313. [Google Scholar] [CrossRef]
  62. Deb, S.; Houssein, E.H.; Said, M.; Abdelminaam, D.S. Performance of Turbulent Flow of Water Optimization on Economic Load Dispatch Problem. IEEE Access 2021, 9, 77882–77893. [Google Scholar] [CrossRef]
  63. Abdul-Minaam, D.S.; Al-Mutairi, W.M.E.S.; Awad, M.A.; El-Ashmawi, W.H. An Adaptive Fitness-Dependent Optimizer for the One-Dimensional Bin Packing Problem. IEEE Access 2020, 8, 97959–97974. [Google Scholar] [CrossRef]
  64. El-Ashmawi, W.H.; Elminaam, D.S.A.; Nabil, A.M.; Eldesouky, E. A chaotic owl search algorithm based bilateral negotiation model. Ain Shams Eng. J. 2020, 11, 1163–1178. [Google Scholar] [CrossRef]
  65. Acharya, U.R.; Dua, S.; Du, X.; Chua, C.K. Automated diagnosis of glaucoma using texture and higher order spectra features. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 449–455. [Google Scholar] [CrossRef]
  66. Acharya, U.R.; Ng, E.Y.; Eugene, L.W.; Noronha, K.P.; Min, L.C.; Nayak, K.P.; Bhandary, S.V. Decision support system for the glaucoma using Gabor transformation. Biomed. Signal Process. Control 2015, 15, 18–26. [Google Scholar] [CrossRef] [Green Version]
  67. Yadav, D.; Sarathi, M.P.; Dutta, M.K. Classification of glaucoma based on texture features using neural networks. In Proceedings of the 2014 Seventh International Conference on Contemporary Computing (IC3), Noida, India, 7–9 August 2014; pp. 109–112. [Google Scholar]
  68. Maheshwari, S.; Pachori, R.B.; Kanhangad, V.; Bhandary, S.V.; Acharya, U.R. Iterative variational mode decomposition based automated detection of glaucoma using fundus images. Comput. Biol. Med. 2017, 88, 142–149. [Google Scholar] [CrossRef]
  69. Bajwa, M.N.; Malik, M.I.; Siddiqui, S.A.; Dengel, A.; Shafait, F.; Neumeier, W.; Ahmed, S. Two-stage framework for optic disc localization and glaucoma classification in retinal fundus images using deep learning. BMC Med. Inform. Decis. Mak. 2019, 19, 1–6. [Google Scholar]
  70. Shehab, M.; Abualigah, L.; Shambour, Q.; Abu-Hashem, M.A.; Shambour, M.K.Y.; Alsalibi, A.I.; Gandomi, A.H. Machine learning in medical applications: A review of state-of-the-art methods. Comput. Biol. Med. 2022, 145, 105458. [Google Scholar] [CrossRef] [PubMed]
  71. Ezugwu, A.E.; Ikotun, A.M.; Oyelade, O.O.; Abualigah, L.; Agushaka, J.O.; Eke, C.I.; Akinyelu, A.A. A comprehensive survey of clustering algorithms: State-of-the-art machine learning applications, taxonomy, challenges, and future research prospects. Eng. Appl. Artif. Intell. 2022, 110, 104743. [Google Scholar] [CrossRef]
  72. Liu, Q.; Li, N.; Jia, H.; Qi, Q.; Abualigah, L. Modified Remora Optimization Algorithm for Global Optimization and Multilevel Thresholding Image Segmentation. Mathematics 2022, 10, 1014. [Google Scholar] [CrossRef]
  73. Gharaibeh, M.; Alzu’Bi, D.; Abdullah, M.; Hmeidi, I.; Al Nasar, M.R.; Abualigah, L.; Gandomi, A.H. Radiology Imaging Scans for Early Diagnosis of Kidney Tumors: A Review of Data Analytics-Based Machine Learning and Deep Learning Approaches. Big Data Cogn. Comput. 2022, 6, 29. [Google Scholar] [CrossRef]
  74. Zhang, W.; Li, X.; Ma, H.; Luo, Z.; Li, X. Universal Domain Adaptation in Fault Diagnostics with Hybrid Weighted Deep Adversarial Learning. IEEE Trans. Ind. Inform. 2021, 17, 7957–7967. [Google Scholar] [CrossRef]
  75. Zhang, W.; Li, X.; Li, X. Deep learning-based prognostic approach for lithium-ion batteries with adaptive time-series prediction and online validation. Measurement 2020, 164, 108052. [Google Scholar] [CrossRef]
Figure 1. Eye shape (normal and glaucomatous eye).
Figure 1. Eye shape (normal and glaucomatous eye).
Electronics 11 01763 g001
Figure 2. Flow diagram of the proposed approach.
Figure 2. Flow diagram of the proposed approach.
Electronics 11 01763 g002
Figure 3. Weight ranking of features.
Figure 3. Weight ranking of features.
Electronics 11 01763 g003
Figure 4. Cumulative accuracy vs. no. of features.
Figure 4. Cumulative accuracy vs. no. of features.
Electronics 11 01763 g004
Figure 5. Stacked up RBMs model DBN.
Figure 5. Stacked up RBMs model DBN.
Electronics 11 01763 g005
Figure 6. Results: (a) original image; (b) OD segmented image; (c) OC segmented image.
Figure 6. Results: (a) original image; (b) OD segmented image; (c) OC segmented image.
Electronics 11 01763 g006
Figure 7. Results of the proposed work on all the datasets.
Figure 7. Results of the proposed work on all the datasets.
Electronics 11 01763 g007
Figure 8. Sample fundus images with noise added.
Figure 8. Sample fundus images with noise added.
Electronics 11 01763 g008
Figure 9. Recognition rate on degraded images of the LAG dataset.
Figure 9. Recognition rate on degraded images of the LAG dataset.
Electronics 11 01763 g009
Table 1. List of features extracted.
Table 1. List of features extracted.
Structural FeaturesCup to Disc Ratio (CDR), Neuro Retinal Rim (NRR), Cup Shape
Textural featuresWavelet-based features, gray level co-occurrence matrix (GLCM)
features—energy, correlation, homogeneity, contrast, and entropy gray-level run length—low gray level run emphasis, gray level non-uniformity, segmentation-based fractal texture analysis (SFTA)
Intensity featuresBrightness, color moments, super pixels, enhanced local binary pattern (ELBP), speeded-up robust feature (SURF), pyramid histogram of oriented gradients (PHOG), local energy-based shape histogram (LESH)
Table 2. Dataset labeling.
Table 2. Dataset labeling.
Database/ImagesNormalGlaucomaTotalType
DRISHIT-GSI [44]1289101Public
ACRIMA [29,45]309396705Public
OTIHS-lihjy [46]482368650Public
LAG [47]343223925824Private
Total423530457280
Table 3. Control parameters of all algorithms.
Table 3. Control parameters of all algorithms.
AlgorithmParameters
ABCN = 30, MCN = 100, limit = 20
HS HMCR = 0.7 ,   PAR = 0.7 ,   η = 1
FA γ = 1 ,   β 0 = 1 ,   α = 0.2 ,   M C N = 100
CS α = 0.1 , pa = 0.25
PSOWmax = 0.9, Wmin = 0.2, C1 = 2, C2 = 2
DEF = 0.8, C = 0.5
EHO nClan = 5 ,   α = 0.25 ,   β = 0.05
Table 4. Average MSE over the LAG dataset.
Table 4. Average MSE over the LAG dataset.
AlgorithmLayer-1Layer-2Layer-3
CDPCDCDPCDCDPCD
ABC0.08910.89400.08810.08840.08800.0878
HS0.12590.13450.12560.11690.11580.1156
FA0.08640.08640.8640.08600.08640.0862
CS0.11460.11460.11760.11750.11640.1162
PSO0.10860.10860.09880.09920.10450.1046
DE0.12500.12540.12540.12540.11580.1156
EHO0.07560.07560.07780.07780.07760.774
Table 5. Performance Metrics.
Table 5. Performance Metrics.
ParametersExpression
Sensitivity (%) T P T P + F N × 100
Specificity (%) T N T N + F P   × 100
Accuracy (%) T P + T N T P + F N + T N + F P × 100
Precision (%) T P T P + F P × 100
Recall (%) T P T P + F N × 100
F-score (%)2 × ( P r e c i s i o n ) ( R e c a l l ) P r e c i s i o n + R e c a l l × 100
Mathew’s correlation coefficient (MCC) (%) ( T P × T N ) (   F P × F N ) ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) × 100
Table 6. Performance of the proposed method employing DBN and EHO.
Table 6. Performance of the proposed method employing DBN and EHO.
DatasetAcc (%)Sens (%)Spec (%)Prec (%)Recall (%)F-Score (%)MCC
Drishti-GS196.9598.5697.4497.6996.8697.680.749
ACRIMA99.3497.198.288.9295.393.50.772
ORIGA98.5194.7398.798.5597.9295.320.784
LAG99.3199.8910096.7394.5695.640.789
Table 7. Impact of features in detecting glaucoma from LAG dataset.
Table 7. Impact of features in detecting glaucoma from LAG dataset.
FeaturesAccuracy (%)Sensitivity (%)Specificity (%)
Structural (SF)94.8795.3293.52
Intensity (IF)95.9889.2392.41
Textural (TF)96.2197.2899.33
SF + IF95.8690.7495.21
SF + TF96.7894.2397.56
IF + TF90.8895.7994.35
Selected features99.3199.89100
Table 8. 10-fold cross validation results.
Table 8. 10-fold cross validation results.
DatasetAccuracy (%)
Drishti-GS197.1
ACRIMA98.5
ORIGA96.2
LAG97.8
Table 9. Performance analysis with dissimilar classifiers.
Table 9. Performance analysis with dissimilar classifiers.
DatasetClassifierAccuracy (%)Sensitivity (%)Specificity (%)
Drishti-GSKNN95.3490.4793.08
RF94.5091.3492.33
SVM95.8696.8796.87
DBN96.2397.5696.62
DBN–EHO96.9598.5697.44
ACRIMAKNN95.6690.8693.78
RF94.3291.2490.84
SVM97.0696.6496.12
DBN97.2698.1697.06
DBN–EHO99.3497.198.2
ORIGA-LightKNN94.2296.8697.08
RF91.3488.5689.75
SVM94.8895.5696.69
DBN96.0697.6597.81
DBN–EHO98.5194.7398.7
LAGKNN94.2495.5695.85
RF92.7890.8991.48
SVM95.6095.6896.45
DBN97.5495.6797.43
DBN–EHO99.3110099.89
Table 10. Comparison of DBN-EHO with CNN models employing transfer learning.
Table 10. Comparison of DBN-EHO with CNN models employing transfer learning.
DatasetClassifierAccuracy (%)Sensitivity (%)Specificity (%)
Drishti-GSAlexNet 93.8491.5792.88
GoogLeNet95.4690.3493.36
VGG1694.1295.7796.42
DBN-EHO96.9598.5697.44
Table 11. Comparison of proposed work.
Table 11. Comparison of proposed work.
ReferencesFeatures/MethodsClassifier UsedImagesPerformance (%)
Karthikeyan and Rengarajan [65]GLCMBPNLocal datasetAccuracy—95
Issac, A. et al. [19]CDR, NRR, blood vessel featuresSVM and ANN67Accuracy—94.11
Sensitivity—100
Mookiah et al. [66]Discrete wavelet and HOSSVM60Accuracy—95
Sensitivity—93.3
Specificity—96.67
Gifta [24]GLCM, HOG, SURFGray Wolf Optimized NNN.A.Accuracy—93.1
Sensitivity—91.6
Specificity—94.1
Acharya, U.R. et al. [22]6 features from LM filter bankKNNNAAccuracy—95.8
Koh, J.E. et al. [20] PHOG, SURF featuresKNN910Accuracy—96.21
Sensitivity—97.42
Samanta et al. [21]Haralick featuresBPN60Accuracy—96.26
Sensitivity—90.43
Specificity—99.5
Acharya et al. [67]Texture and HOFRF60Accuracy—91
Acharya et al. [68]Gabor transformation and principal component analysisSVM510Accuracy—93.10; sensitivity—89.75; specificity—96.20
Yadav et al. [69]Homogeneity, Contrast, energy, correlation, entropyN.N.20Accuracy—72
Maheshwari et al. [70]Entropy and fractal SVM488Accuracy—95.19
Bajwa M. N et al. [71]ROI, Scaling2-Stage CNNORIGAAUC—0.87
Raghavendra et al. [72] -20 layer CNN1426Accuracy—98.13
Chen et al. [73]-16 layer CNNSECS, ORIGAAUC—0.881
Proposed WorkStructural, intensity, and texture featuresDBN and EHO7280Accuracy—99.34
Sensitivity—100
Specificity—99.89
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ali, M.A.S.; Balasubramanian, K.; Krishnamoorthy, G.D.; Muthusamy, S.; Pandiyan, S.; Panchal, H.; Mann, S.; Thangaraj, K.; El-Attar, N.E.; Abualigah, L.; et al. Classification of Glaucoma Based on Elephant-Herding Optimization Algorithm and Deep Belief Network. Electronics 2022, 11, 1763. https://doi.org/10.3390/electronics11111763

AMA Style

Ali MAS, Balasubramanian K, Krishnamoorthy GD, Muthusamy S, Pandiyan S, Panchal H, Mann S, Thangaraj K, El-Attar NE, Abualigah L, et al. Classification of Glaucoma Based on Elephant-Herding Optimization Algorithm and Deep Belief Network. Electronics. 2022; 11(11):1763. https://doi.org/10.3390/electronics11111763

Chicago/Turabian Style

Ali, Mona A. S., Kishore Balasubramanian, Gayathri Devi Krishnamoorthy, Suresh Muthusamy, Santhiya Pandiyan, Hitesh Panchal, Suman Mann, Kokilavani Thangaraj, Noha E. El-Attar, Laith Abualigah, and et al. 2022. "Classification of Glaucoma Based on Elephant-Herding Optimization Algorithm and Deep Belief Network" Electronics 11, no. 11: 1763. https://doi.org/10.3390/electronics11111763

APA Style

Ali, M. A. S., Balasubramanian, K., Krishnamoorthy, G. D., Muthusamy, S., Pandiyan, S., Panchal, H., Mann, S., Thangaraj, K., El-Attar, N. E., Abualigah, L., & Abd Elminaam, D. S. (2022). Classification of Glaucoma Based on Elephant-Herding Optimization Algorithm and Deep Belief Network. Electronics, 11(11), 1763. https://doi.org/10.3390/electronics11111763

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop