Next Article in Journal
A Note on Hermite–Hadamard–Fejer Type Inequalities for Functions Whose n-th Derivatives Are m-Convex or (α,m)-Convex Functions
Previous Article in Journal
Approximation of Fixed Points for Enriched Suzuki Nonexpansive Operators with an Application in Hilbert Spaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Detection and Classification of Meningioma Tumor from MR Images Using Sea Lion Optimization and Deep Learning Models

by
Aswathy Sukumaran
1,2,* and
Ajith Abraham
1,3
1
Machine Intelligence Research Labs (MIR Labs), Scientific Network for Innovation and Research Excellence, P.O. Box 2259, Auburn, WA 98071-2259, USA
2
Department of Computer Science and Engineering, Jyothi Engineering College, Thrissur 679531, Kerala, India
3
Centre for Artificial Intelligence, Innopolis University, 420500 Innopolis, Russia
*
Author to whom correspondence should be addressed.
Axioms 2022, 11(1), 15; https://doi.org/10.3390/axioms11010015
Submission received: 19 September 2021 / Revised: 17 December 2021 / Accepted: 20 December 2021 / Published: 30 December 2021

Abstract

:
Meningiomas are the most prevalent benign intracranial life-threatening brain tumors, with a life expectancy of a few months in the later stages, so this type of tumor in the brain image should be recognized and detected efficiently. The source of meningiomas is unknown. Radiation exposure, particularly during childhood, is the sole recognized environmental risk factor for meningiomas. The imaging technique of magnetic resonance imaging (MRI) is commonly used to detect most tumor forms as it is a non-invasive and painless method. This study introduces a CNN-HHO integrated automated identification model, which makes use of SeaLion optimization methods for improving overall network optimization. In addition to these techniques, various CNN models such as Resnet, VGG, and DenseNet have been utilized to give an overall influence of CNN with SeaLion in each methodology. Each model is tested on our benchmark dataset for accuracy, specificity, dice coefficient, MCC, and sensitivity, with DenseNet outperforming the other models with a precision of 98%. The proposed methods outperform existing alternatives in the detection of brain tumors, according to the existing experimental findings.

1. Introduction

The brain tumor and its analysis are of tremendous interest because of the growing innovations in medical image processing. An abnormal tissue development within the brain causes a tumor. According to the National Brain Tumor Foundation’s (NBTF) global review, the improvement of brain tumor diagnoses among patients and the death rate due to brain tumors are outpacing earlier years’ findings clinical experts can give patients more effective e-health care services, thanks to developments in medical imaging for this enhancement [1]. E-health care systems have a wide range of applications in medicine [2]. Due to their high accuracy and efficient results as presented by the radiologist, computer-vision-based biomedical imaging systems have gained appeal among clinical specialists, allowing them to handle treatment-related concerns more efficiently. Positron emission tomography (PET), magnetic resonance imaging (MRI), and computed tomography (CT) are the most popular modalities used to examine a brain tumor [3]. Magnetic resonance imaging (MRI) is well-known medical equipment that can be used to diagnose and study a variety of disorders, including brain tumors, neurological ailments, and epilepsy, among others. Typically, a system that is entirely processed by hardware/computer aids in the automation of this process, in order to produce correct and timely results. Machine learning (particularly deep learning) improvements have made it easier to discover, classify, and measure trends in medical images.
The major focus of this research is identifying tumor growth in the brain through the use of MRI scans. Moosa et al. discovered that these cancerous cells influence the health and functioning of the patient’s brain. The most frequent benign intracranial tumor is meningioma [4]. Arachnoid cap cells, which are cells within the thin spiderweb-like membrane that protects the brain and spinal cord, give rise to these types of tumors. The arachnoid is one of three protective layers that surround the brain and spinal cord, commonly known as the meninges. These tumors normally grow inward, forcing on the brain or spinal cord, but they can also grow outward, thickening the skull. The majority of meningiomas are benign tumors that grow slowly. Certain blood vessels have cysts (fluid sacs), calcifications (mineral deposits), and tightly packed groupings of blood veins. Meningiomas are responsible for 30% of all types of brain tumors. They can occur in a variety of areas across the brain [5]. Many image processing technologies and methodologies have been used in the diagnosis and treatment of brain tumors. Artificial intelligence and machine learning, in particular, have played an essential role in the medical industry, as well as providing a key tool for supporting various medical professions. Several machine learning algorithms are employed to construct completely automated CAD systems, particularly for image segmentation and classification, which are used to aid radiologists as a second view [6].

1.1. Research Challenges and Related Problems

Normally, it is found that segmenting and classifying images in a database with a huge number of images needs more calculation time. The MRI images are recorded in numerous sagittal planes so that the database can be enlarged to include all possible planes. Preprocessing is the first step in creating a CAD system, and it is used to avoid overfitting by deleting undesirable areas and improving classification accuracy. It also aids in the removal of noise, which can cause visual blurring. In some circumstances, pixel correlation is ignored for greater speed, which can have an impact on the final result of the CAD system. The goal of this study is to use CNN architecture to solve the difficulty indicated above, as well as to categorize three different tumor types from an uneven database [7]. The efficiency of a system is determined by how effectively it performs in a complex context, as most existing systems are constrained to making minimal use of available resources.

1.2. Key Contribution

This paper focuses on detecting and efficiently classifying meningioma tumors using reformed DL models, with the following highlights:
  • Image features are improved using a unique enhancing technique, and objective parameters such as PSNR, MSE, and SSIM are calculated.
  • Clustering is a powerful, exploratory data analysis tool for gaining an intuition of the data’s structure. The K-means algorithm is an iterative technique that attempts to split a dataset into K distinct non-overlapping subgroups (clusters), each of which has a data point that belongs to a single group. The proposed work made an attempt to make intra-cluster data points as comparable as possible while maintaining clusters as distinct one. The segmentation is performed by K-means and with modified K-means clustering methods. Performance measures from both techniques are calculated, and the best one is selected for further processing.
  • After the optimized feature extraction technique, Sealion optimization (SLA12), the brain image is classified using Harris hawks optimization (HHO) [8]. The features are obtained with high accuracy and precision, and with much less computational time.
The paper is organized as follows: Section 1.3 provides a related study on the tumor from other researchers. Section 2 describes the details of the suggested approach and also includes experimental information and findings. Section 3 presents the conclusion and future work.

1.3. Related Works

Numerous steps have been taken to establish a highly accurate and reliable system for brain tumor identification. However, due to high inter and intra shape, texture, and contrast variations, it remains a challenging problem. On the basis of traditional ML and deep learning algorithms, a number of techniques for automatic brain tumor classification from MR images have been presented. Preprocessing is the first step in this process. Despite the numerous technologies available, extracting characteristics and identifying brain tumors remains a bit problematic. Furthermore, a universal approach for categorizing and detecting brain cancers is difficult to come by. A few alternative approaches have been explored using a variety of datasets and based on the methodology used for segmentation and classification. The number of images in the database is the largest issue with using neural networks to classify and segment MRI images. Furthermore, because MRI images are taken in many planes, the possibility of using all accessible planes could expand the database. Preprocessing is essential before inputting the images into the neural network, as this could impact the classification output by overfitting. The main focus of this research is to develop a system that properly recognizes and classifies tumors with a narrow percentage of interpretation error. For the classification of brain tumors in MRI scans, a variety of approaches and methodologies have been proposed. FCM, ANN, SVM, PCA, CNN, MLP, DWT, RF, KNN, K-means, EM, and others are some of the most commonly used algorithms in the classification and segmentation processes to extract vital information from medical imaging modalities. The findings of some of the most significant and well-known studies are summarized in Table 1.

2. Methodology

The goal of this research is to use deep learning techniques to implement various stages of detection and segmentation techniques in MRI brain imaging. To determine abnormality in the brain image, the proposed method employs an augmenting approach, as well as segmentation and classification techniques. The system’s suggested framework is depicted in Figure 1. BRATS 2015 is the database used by the system. These input images are first preprocessed to eliminate noise and adjust the image’s brightness and contrast for preprocessing, an adaptive diffusion filter and a boosted anisotropic diffusion filter were used; this outperforms the existing anisotropic diffusion filter and is compared to other filters to validate the performance. The ROI is retrieved from the preprocessed image using the modified K-means technique. Swallow swarm optimization (SSO) and adaptive swallow swarm optimization (ASSO) are used in identifying and extracting features from the segmented image, and SLA12 is used to optimize the extracted features. This improved feature is fed to the classifier, which divides images into normal and abnormal categories. CNN is the classifier utilized in this case.

2.1. Image Acquisition

The BRATS 2015 data set was employed in this analysis, and it contains 325 high-grade (HGG) and 54 low-grade (LGG) images scanned for meningioma patients. In the 2015 BRATS multimodal MRI dataset, there were also four scanning sequences available for each patient, using weighted T1, gadolinium-enhancing contrast (T1C), weighted T2, and FLAIR. Each subject was scanned with T1. The HGG picture from BRATS 2015 is used in this report. Table 2 shows the most commonly used dataset for brain tumors. Figure 2 gives the most frequently used dataset details, and Table 2 gives the link for the online data repository.

2.2. Preprocessing: Image Enhancement

Preprocessing of the image is done to enhance the image quality. The dynamic range of the selected characteristics is increased so that they may be easily spotted. To extract the relevant information, the MRI images of the brain are first obtained and preprocessed. The quality of photographs might decline during the image production process due to a variety of factors, such as being out of focus, noise, the optical system, distortion, relative motion between the camera and the scene, and so on. A variety of preprocessing approaches is used. The goal of image enhancement is to make information in images more interpretable or to give better input for other automated image processing techniques. Film artefacts such as labels and marks on the MRI image, as well as the high-frequency components, are removed during the enhancing stage. The MRI picture is turned into a standard image without noise, film artefacts, or labels after enhancing techniques are applied. MRI images are frequently contaminated by noise, such as Gaussian and Poisson noise [25]. The great majority of de-noising methods assume white Gaussian noise is additive. Edge-preserving bilateral filter, total variation, and non-local means are some of the strategies meant to remove Gaussian noise. Various image enhancing techniques are used in this study, including bilateral filtering (BF), curvelet transform (CT), adaptive histogram equalization (AHE), and boosted anisotropic diffusion filter (BADF).
● Bilateral filtering
Bilateral filtering is a technique for smoothing images while keeping the edges. The use of bilateral filtering has rapidly advanced, and it is now employed in image processing applications such as image de-noise, picture enhancement, and so on [26]. The pixel of each point in conventional low-pass filtering is assumed to be equal to that of the nearby points:
k d 1 ( x ) | | f ( δ ) c ( δ , x ) d δ
where ( c ( δ , x ) ) is the arithmetic contact between the region centre x and a nearby point δ . It is possible that both input (f) and output (ℎ) images are multiband. Furthermore,
                                        k d ( x ) = | | c ( δ , x ) d δ
Nevertheless, the pixels of points at borders are essentially identical to the closing points. Hence, the lines are blurred. This filter blends gray rates based on both numerical proximity and multispectral similarities, thus allowing values that are identical in both domains. Similarly, the filtering range is specified as follows:
                                                                                          h ( x ) = k r 1 ( x ) | | f ( δ ) s ( f ( δ ) , f ( x ) d δ
where s ( f ( δ ) , f ( x ) computes the photographic similarity between the region center x and a nearby point δ , and the kernel assesses pixel similarity in this situation. In this instance, the ongoing standardization is as follows:
          k r 1 ( x ) = | | s ( f ( δ ) , f ( x ) d δ
The bilateral filtering is described as given below:
h ( x ) = k 1 ( x ) | | f ( δ ) c ( δ , x ) s ( f ( δ ) , f ( x ) d δ
where k ( x ) = c ( δ , x ) f ( f ( δ ) , f ( x ) d δ . Each region combines filtration area, and their distance will be defined as bilateral filtration. It substitutes the pixel value at a related and near-average pixel value. In flat regions, pixel values in a small neighborhood are related to each other, and the bilateral filter effectively acts as a normal domain filter, averaging the weak, inadequately associated variations between the pixel values produced by the noise. Figure 3a depicts the clinical database input, while Figure 3b depicts the bilateral filtering image for all of those inputs.
● Curvelet Transform
The curvelet transform is a multi-scale transform that is based on the wavelet transform. The basic elements of the curvelet transform are dimension, location, and orientation. It has excellent orienting characteristics [27]. In comparison to the wavelet transform, the curvelet transform is the best method for removing noise. It accurately depicts curved objects, allowing them to be used in graphical applications [28]. The main applications are edge detection and picture de-noising. The bilateral filtering image for all of those inputs is shown in Figure 3c.
A curvelet coefficient t (a,b,c) is given by the following:
t   (   a , b , c ) : f , φ a , b , c
where a = 0, 1…—scale limit; b = 0, 1…—orientation limit; c = (c1, c2), c1, c2 Z-translation limit.
A curvelet coefficient is the internal multiplication among elements f K2 (M2) and curvelet φa,b,c, as follows:
t ( φ a , b , c ) = f , φ a , b , c M 2 f ( X ) φ A , B , C ( X )   d X ¯ = 2 ( 2 π ) 2 F ^ ( y ) U a ( M θ , y ) e i X C ( A , B ) , y dy
The curvelets with a scale of 2−a, orientation θ 1 , and position XC(A,B) = R θ 1 1 (c1.2a,c2.2−a/2) can be represented as:
              φ a , b , c = φ a , ( R θ 1 ( X X c A , B ) ) )
where θ 1 = 2 π .2 a / 2 .1 , with l = 0, 1… 0 < θ 1 < 2 π , and R θ being rotation by θ 1 radians.
The following is a description of the curvelet transform algorithm. The photo was taken as I.
  • Using additive wavelet transform, the image ‘I’ is separated into three sub-bands: s1, s2, and s3.
  • The tiling procedure is carried out in sub-bands s1 and s2.
  • After that, the ridgelet transform is applied to each of the tiling sub bands s1 and s2.
It is mostly concerned with the sharpening of curves. Curvelet transforms are based on partitioning a picture into minute overlapping tiles and then applying a ridgelet transform. This is primarily utilized in the medical field. The curvelet has a number of important qualities, one of which is:
L = W 2
where L = length = 2−A, W = width = 2−(2*A), and A = scale.
In a tiling procedure, the altered image’s sub-bands s1 and s2 are separated into overlapping tiles, resulting in a reduced dimension, and then the curving line is changed into a straight line, avoiding edge effects. To approximate the curving line into overlapping tiles, segmentation is used. Discrete segmented tiles are subjected to the ridgelet transform. This ridgelet transform is used to detect shapes. The ridgelet transform calculation is highly complex since it involves many stages, but nowadays it is simple because it does not employ the ridgelet transform, therefore the time complexity is substantially decreased [29].
The waveform φ a ( x ) is defined as FT φ a ( y ) U a ( y ) .
Let the frequency window U a be defined in the Fourier domain by:
            U   ( p , θ ) = 2 2 a 4 M ( 2   p ) N ( 2 a 2 θ ) 2 π
where M is the radial window, N is the angular window, and a / 2 is the truncated integer of a/2.
● Adaptive Histogram Equalization
The main purpose of image enhancement is to improve certain characteristics of an image to improve its visual quality. This is a variation on the classic histogram equalization method. It improves visual contrast by altering the image’s intensity values. The bilateral filtering image for all of those inputs is shown in Figure 3d. Every bar on the equalized histogram is of the same height, according to the physical meaning of the histogram. That is:
                p s ( s ) d s = p r ( r ) d r
Assume that   s = T ( r ) is a monotonically increasing function in the interval, and that its inverse, r = T 1   ( s ) , is similarly monotonic. We can conclude (12) from (11):
                    p s ( s ) = [ p r ( r ) 1 d s d r ] r = T 1   ( s ) = p r ( r ) 1 p r ( r ) = 1
The traditional histogram equalization algorithms mapping relationship is as follows: the relationship between I (the grey value of the pixel in the original image) and f i (the grey value of the pixel in the enhanced image) is discrete in discrete conditions.
    f i = ( m 1 ) T ( r ) = ( m 1 ) k = 0 i q k Q
where m is the number of gray levels in the original image, q   k denotes the number of pixels in the image with kth gray level, and Q denotes the total number of pixels in the image. If an image comprises n different gray levels, and each grey level’s occurrence probability of ith gray level is p i , the grey level’s entropy can be described as follows:
                                                          e ( i ) = p i l o g p i
The entropy of the whole image is as follows:
E = i = 0 n 1 e ( i ) = i = 0 n 1 p i l o g p i
● Adaptive Boosted Anisotropic Diffusion filter (BADF)
The original image undergoes anisotropic diffusion, which is a non-linear and space-variant modification. This filter was first introduced in 1990 (Perona and Malik, 1990). Its main aim is to smooth an image without blurring it while keeping its important details. This filter starts by setting the required iterations number. Figure 3e shows the bilateral filtering image for all those inputs [30]. After creating the diffused image, the proposed BADF adds a partial differential equation (PDE) that enhances the existing anisotropic diffusion filter diffusion, which is missing at the borders and boundaries, and can also be used to smooth the surface. It is a powerful picture enhancing tool that uses unsupervised machine learning. It not only smooths out the image, but it also keeps certain key details like edges and textures [31]. Depending on intensive tests, good results were attained when the iterations number was set to 20. Next, the size of the image is determined so that it can be used when calculating the four nearest-neighbor differences. At this point, the iterative process starts by computing the four nearest-neighbor differences using Equations (16)–(19).
        N I i , j = I i 1 , j I i , j
        S I i , j = I i + 1 , j I i , j
        S I i , j = I i + 1 , j I i , j
                W I i , j = I i , j 1 I i , j                    
where I i , j is the smoothed image at each iteration, in which at the first iteration, it should fulfil I i , j = f i , j , where f i , j represents the input image. i , j   are image coordinates; N , S , E , W represent the detected differences at the east, west, north, and south, respectively. The image’s high-frequency components are held by the identified differences. The high-frequency components in each direction are then attenuated using four conduction operators computed using Equations (20) and (21).
            g N = 1 1 + ( N I i , j k ) 2
              g s = 1 1 + ( S I i , j k ) 2
      g E = 1 1 + ( E I i , j k ) 2
  g W = 1 1 + ( W I i , j k ) 2
where K is a scalar that controls the smoothness degree, in that K should satisfy (K > 1), as increasing the value of K produces a smoother output. The value of K is set to 7 in the standard anisotropic diffusion filter. In this study, parameter K is computed automatically based on local statistics using Equation (24)
k = 2 | m e a n ( f i , j ) ( 0.75 σ ( f i , j ) ) |
where σ represents the standard deviation. Next, the entire image is smoothed using the divergence, which can be calculated using Equation (10).
I i , j = I i , j + 0.25 [ ( g N N I i , j ) + ( g S S I i , j ) + ( g E E I i , j ) + ( g W W I i , j ) ]
where I i , j is a smoothened image.
Outputs for the different stages of the preprocessing method, such as bilateral filtering, curvelet transform, adaptive histogram equalization, and adaptive boosted anisotropic diffusion filter are clearly depicted in Figure 3. Statistical analysis for both normal and abnormal brain images for the preprocessing method in terms of signal-to-noise ratio, mean square error, and structural similarity index measure for all the methods, such as bilateral filtering, curvelet transform, adaptive histogram equalization, and adaptive boosted anisotropic diffusion filter, is given in Table 3 and Table 4.

2.3. Segmentation

The objective of the segmentation of the brain tumor image is to identify the tumor location and spread, that is the active tumorous tissue, edema, and necrotic tissue. In this study, modified K-means and HHO algorithms are used for segmentation [32,33]. The modified K-means algorithm helps to enhance the boundary region, thereby being able to remove noise in that area, and it also helps to speed up the K-means clustering method by avoiding the empty clusters in the iterative process. Moreover, a combination of modified K-means with HHO helps to deal with false detection problems at the time of segmentation by increasing the efficiency and accuracy of the system. Segmented results of MRI images using hybrid structure of modified K-means and HHO is given in Figure 4.
● Modified K-means clustering with HHO
The key advantage of employing the method of the modified K-means is that it selects all the scenarios by considering every difference in the data set, instead of arbitrarily choosing a certain group of initial centroids, as is usually done. Modified K-means will spread the instances of the data set to appropriate clusters with the highest precision. In contrast to previous K-means, it is not necessary to determine the number of K clusters, because this is done dynamically in the modified K-means methodology (see Algorithm 1).
Algorithm 1. Modified K-Mean clustering.
Step 1: Calculate distance matrix distance d i ( x i , x j ) ,
where   i = 0 ,   1 ,   2 ,   . N 1   and j = 0 ,   1 ,   2 ,   N 1 .
Step 2: Calculate the threshold value ‘Th’ using (26).
i = 0 N 1     j = 0 N 1     d i s t ( x i , x j ) N N (26)
where ‘D’ ( d 1 , d 2 . d n ) is the data set.
‘K’ is the cluster. ‘n’ is the number of data points.
‘X’ ( X 1 , X 2 . X n ) is an instance of the data point. T h is the threshold. ‘c’ is the cluster center.
Step 3: Find the minimum mean from X i   t o   X j using (27).
m i n ( i = 0 , j = 0 N 1     d i s t ( x i , x j ) ) (27)
Step 4: Determine the minimal mean value index   X i .
Choose the data point X i as the initial centroid.
Step 5: Until the data points change, the group repeat steps 6 and 7.
Otherwise, go to step 8.
Step 6: Determine the distance between each of the K cluster centers c j and each data point xi.
if ( T h > = d i )
Set nearest cluster with data point xi.
else
K = K + 1;
Step 7: Recalculate the centroid of every cluster.
Step 8: End.
● Harris Hawks Optimization (HHO)
Mathematically, Harris hawks behaviors are proposing a new metaheuristic stochastic method, the HHO algorithm. The defining characteristic of Harris hawks’ activity is that via good coordination, they trace, encircle, and approach the possible prey (rabbits in most cases) and then attack them. In hunting, skillful escape behavior called “surprise pounce” is efficiently performed. The HHO method includes exploration and exploitation phases, similar to other meta-heuristic algorithms.
Harris hawks will be randomly searching for prey during the exploration phase according to the following equation:
X ( t + 1 ) = {   X rand   ( t ) r 1 | X rand   ( t ) 2 r 2 X ( t ) | q 0.5 ( X rabbit   ( t ) X m ( t ) ) r 3 ( L B + r 4 ( U B L B ) ) q < 0.5   }
In iteration (t + 1), the hawks are located at X(t + 1), the rabbit (victim) is located at Xrabbit (t), r1 to r4 and q are randomly labelled from 0 to 1, Xrand (t) represents a selected hawk at a random position, and Xm is the average location of the current hawk population, as determined by Equation (29):
X m ( t ) = 1 N i = 1 N X i ( t )
where Xi(t) denotes each hawk’s location for iteration t, and N indicates the overall hawk number. There is a transitional phase before the exploitation stage once the discovery stage is completed. In this transitional period, the energy of the rabbit should be shaped according to Equation (30):
E = 2 E 0 ( 1 t T )
where E is the rabbit’s escape energy, E0 is the initial rabbit energy state, and the maximum number of iterations is T. The E0 value can vary from −1 to 1 dependent on the victim’s physical fitness. This indicates that the victim loses energy while E0 is heading towards −1, and vice versa. During these last stages of the processing of the algorithm, the Harris hawks abruptly approach their prey. Four tactics of attack are available. Here, r is seen as an escape probability. When E ≥ 0.5 and r ≥ 0.5, Harris’s hawks slowly utilize a soft besiege technique to surround the prey. The following is the mathematical model:
                    X i t + 1 = Δ X i t E | J X p r e y X i t | ,   Δ X i t = X p r e y X i t
where Xi(t + i) is the distance between the current individual and the prey, and J indicates the strength of the prey’s jumping during the escape, and it takes a random value in the range [0, 2]. When E < 0.5, r ≥ 0.5, because of inadequate escape energy, the prey can’t escape, and the location of the Harris hawks is written as follows:
X i t + 1 = X prey   E | Δ X i t |
When the Harris hawks do soft besiege with increasing fast dive tactics to confuse prey when the prey have sufficient energies to effectively escape, that is E ≥ 0.5, r < 0.5. It can be represented as follows:
X i t + 1 = { Y = X prey   E | J X prey   X i t |   ,   if   f ( Y ) < f ( X i t ) Z = Y + S × Levy ( d ) ,   if   ( Z ) < f ( X i t ) }
The problem dimension is denoted by d, and S is a 1 × d random vector. The Lévy Flight function is Lévy, When E < 0.5, r < 0.5, the prey has insufficient escape energy. The Harris hawks will attack the prey in the following ways:
X i t + 1 = { X prey   E | J X prey   X m t | ,   if   f ( Y ) < f ( X i t ) Z = Y + S × Levy ( d ) ,   if   f ( Z ) < f ( X i t ) }

2.4. Feature Extraction and Optimized Feature Selection

During the feature extraction step, the necessary high-quality data or characteristics of the image are obtained in the form of statistical features, form, color, and texture. The extraction of functions is a key step to reduce the classifier complexity used to categorize the characteristics of an image. In addition, the extraction function is efficiently used to enhance the accuracy of the diagnosis process by selecting major features. This paper covers two new features, the area of the tumor and the dice coefficient index, which are retrieved from the segmented image. There are a variety of existing strategies for image feature extraction; however, in this research, the adaptive swallow swarm optimization (ASSO) is utilized, and SLA12 (sea lion) is used to improve feature selection, ensuring that only important features are used by the classifier to categorize tumor type.
In the ASSO algorithm, the number of swallows is regarded to be the test case characteristic. Basically, with a test suite SW = (T 1, …. T) and C = (c 1, …c m). The operating procedure of the SSO algorithm is just like PSO, though it has a few more characteristics, which include the usage of three kinds of particles: (a) aimless particles (oi), (b) explorer particles (ei), and (c) leader particles (li).
Each one of them has specific duties in the group. The ei particles are accountable for finding the best optimal hybrid criteria function space. They carry out this searching for an optimal hybrid criterion based on the impact of different parameters (V. Keerthika):
  • Position of the global leader.
  • Position of the local leader.
  • The best individual experience along the path.
  • The earlier path.
The particles utilize the equations below for carrying out a search for an optimal hybrid criterion and proceeding with the test suite process:
V   H L   i + 1 = V   H L   i + α   H L   r a n d (   ) ( e   b e s t e   i   )   + β H L   r a n d ( ) ( H L   i e   i   )  
The feature selection technique chooses the best features from the features available, reducing the classifier’s complexities. The accuracy, specificity, and sensitivity of the classifiers will all improve as a result of this improved feature extraction. The formula for some of the useful aspects in statistics is provided in Table 5.
In this work, Table 6 gives the computational analysis for different feature extraction methods. Table 7 show the comparison statistics of their performance measure of the existing feature extraction method with the proposed ASSO and SLA12. From the table, it is evident that SLA12 exceeds the other optimization approach with the greatest outcomes in Table 6, Table 7 and Table 8. Table 7 and Table 8 show some of the most prominent aspects of the statistical analysis of first- and second-order, respectively.

2.5. Classification

The classification method is designed to classify all pixels into one of several classes within digital images. Multi-spectral data are usually utilized to classify, and, in effect, a numerical foundation for categorization is employed as the spectral pattern in the data for every pixel. The image grading aims to detect and show the characteristics in the image about the item which these characteristics represent truly at the ground as a distinctive grey level (or color). The grading of images refers to image labelling in one of several established classifications. In this study, the selected feature is supplied to the classifier for training purposes following optimum function selection. We use CNN classifiers together with many models, such as ResNet50, VGG16, DenseNet and CNN, in our research with optimal sea lion optimization [34].
● Convolutional Neural Network (CNN)
The proposed system uses CNN-based artificial segmentation methods to explore smaller 3 × 3 kernels. The tiny kernels help the architecture become deeper by employing fewer network weights. To achieve functionality maps, a coevolutionary layer is used to convert a sign or an image using kernels. The weight of a kernel thereby ties a unit to the previous layer in a feature map [35,36]. The kernel weights are reversed during the training phase to enhance particular input features. Convolution layers weigh less than dense FC layers because kernels are shared across all units with the same features, making CNN training easier and lowering the chance of overriding. Moreover, because the same kernel across the whole image is combined, the same feature is identified regardless of the location, resulting in invariance translation [4]. The neighborhood information, which is a helpful source of context information, is taken into consideration by employing kernels. The output of every brain unit is normally non-linear. Table 9 shows the complete CNN network. The framework on multigrade tumor segmentation and classification using the CNN network is shown in Figure 5 and Figure 6 gives the diagrammatic representation on cascaded CNN architecture on brain tumor segmentation.
The retrieved characteristics become abstract with increasing depth if we stack numerous coevolutionary levels. The initial layers improve characteristics such as edges grouped as motifs, components, or objects in the succeeding layers [13,37].
In the context of CNN, the following ideas are essential:
  • Initialization: Convergence must be achieved. The initialization of Xavier is used [38]. This maintains activation and gradient-regulated levels, otherwise gradients that are back-propagated can burst or disappear.
  • Activation Function: Non-linear modification of the data is done through activation functions. Rectifier linear units are defined as follows:
    f ( x ) = m a x ( 0 , x )
Better outcomes than traditional sigmoid and hyperbolic tangent functions were discovered to be achieved, and training accelerated. Imposing a constant 0 can nevertheless interfere with the gradient flow and consequent weight adjustment. These constraints are addressed using a version dubbed the leaky rectified linear unit (LReLU); the negative half of the function has a modest slope. This function is defined as follows:
f ( x ) = ( 0 , x ) + α m i n ( 0 , x )  
where α is the parameter indicating leakiness. SoftMax is used in the final FC layer.
  • Pooling and Regularization: This mixes the feature maps with spatially close-by features. This arrangement of potentially redundant parts makes the image concise and invariant for slight changes such as minor details. Max-pooling or medium swimming is more common to link features.
  • The overfitting is used to minimize. The levels of FC here utilize Dropout. It eliminates nodes with probability from the network in every training stage. This drives all FC layer nodes to learn a better representation of the data, thereby avoiding co-adaptation of nodes. Every node is used during testing. Dropout may be regarded as an ensemble of many networks and bags, since a piece of the training data is training each network.
  • Data Augmentation: This may be used to develop training sets and eliminate overfitting. It may be utilized to increase the exercise size and to minimize overfitting. Since the center voxel is used for the patch class, the increase in data was limited to rotational operations. Some writers additionally look at translations of images, although this might lead to a misclass being attributed to the patch for segmentation [39]. We extended the data set while training by rotating the original patch and producing new patches. We have used numerous angles of 90 in our proposal to explored other alternatives.
  • Loss Function: During the training, this is the function that will be lowered. The cross-entropy categorical approach was employed.
    H = j v o x e l s       k c l a s s e s       c j , k l o g ( c ^ j , k )
    where the probabilistic predictions (after the SoftMax) are represented by c ^ , while the target is represented by c.
The architecture and training for our CNN are discussed in the following parts.
  • Architecture: In intra-tumoral structures, brain tumors show great diversity, which is a segmentation challenge. We developed the CNN and optimized the transformation of intensity normalization for each grade of tumor to reduce such complexity.
The loss function must be reduced to train the CNN, although it is not particularly linear. Harris hawks is used as an improvement approach that takes steps in line with the negative gradient toward the local minimum level. In low-curvature locations, though, it can be lenient. We also use the accelerated momentum from Nesterov to accelerate the process in these regions. The velocity remained constant, and after each epoch, the learning rate was reduced linearly. We see an era as a whole pass of the training samples.
● Hybrid CNN With Sea Lion Optimization algorithm
By estimating the likelihood of each image voxel belonging to the target, the tumor or lesion part is isolated from the background. There are four input parts and convolution parts in the CNN network. The layer processes the incoming image so that the desired image patches are produced. The section convolution processes the designed image patches in which multifaceted filters are operated and feature maps are issued. The completely linked layer grouping all the maps is also available. A prediction value for classifying each voxel image is assessed, and a segmentation map is shown. The classification section helps to give each voxel a probability. The SoftMax function is used to convert maps into category probabilities in most convolutional neural networks [40]. The model is developed with Kera’s. Selected 33 × 33 patches of MRI scans were used to grade the center-pixel of the four-layer sequential model [41]. There are four channels for each input, one of them being a channel for each image sequence, and the network may learn the relative pixel intensities of each class. Here the classification architectures Resnet, VGG, and DenseNet have been used in this proposed system [42,43,44].
● VGG
The VGG16 and VGG19 architectures are built on and characterized by the same models. They have employed 3 × 3 coevolutionary filters layered in greater depth on each other and reduce the volume by maximum pooling. At last, the SoftMax classification system, which was built based on their 2014 ImageNet submission and which achieved first and second places in the localization and classification track, is complete with the three fully connected (FC) layers, two of 4096 and one of 1000 neurons (the one displays class scores).
● DenseNet
DenseNet’s dense block consists of convolution and non-linear layers. Some optimizing methods such as drop-out and batch normalization are also applied. In addition, outputs from the preceding layers are combined, rather than the summation in the compact block given by DenseNet. Assume that the form of an image (28, 28, 3) in the input is the RGB color space in three images. First, we spread the image across the first N channels and get the image (28, 28, N). Every successive overlapping layer generates k features, with the same height and width. Different optimization strategies are applied in deep neural networks by modifying parameters like weights and learning rates to decrease loss and improve precision. The sea lion optimization (SLnO) was used in this section [8].
● ResNet
A bottleneck design is utilized for deeper networks such as ResNet50, ResNet152, and so on. Input images of any height or width can be taken by the network as numerous channel widths of 32 and 3. The input size is considered 224 × 224 × 3 for explanatory purposes. The initial convolution and max-pooling of all ResNet architectures are achieved using 7 × 7 and 3 × 3 kernel sizes, respectively. Three layers are layered over each residual function F. The three layers are 1 × 1, 3 × 3, and 1 × 1 convolution. The 1 × 1 convolution layer reduces and restores dimensions afterward. The 3-to-33×3 layer is left as a minor input/output bottleneck. The network also contains an average pooling layer and a fully linked layer of 1000 neurons (ImageNet class output).
● Sea Lion Optimization
Sea lions are clever animals [45]; the key characteristic of sea lions is their quick reaction to the movement of fish. They also have amazing senses that allow them to detect fish prey even in the dark depths. Their eyes point to the prey; they can focus swiftly on the prey. Specifically, they may extensively enlarge their pupils to give their eyes a clear underwater vision with enough light. However, in a cloudy atmosphere, occasionally the vision is not clear enough. Therefore, the most important feature of sea lions is their high-sensitivity whiskers. These whiskers aid them in spotting the location of prey. If the prey swims, a wake or waves are left behind. Sea lions are able to track fish using their whiskers. The following are the primary steps in the hunting process of sea lions: (i) chasing prey using their whiskers, (ii) calling for other subgroup members, (iii) encircling and trapping the prey, (iv) attacking the prey. The SLnO method assumes that either the optimum or closest answer is the target prey. This conduct is characterized mathematically by Equation (39).
D i s t _ = | 2 B _ . P ( t ) _ S L ( t ) _ |
Distance and SL(t) are the position vectors between the sea lion and the target lion P(t). The distance between the target lion is represented by Dist. The present iteration is known as t, and B _ as [0, 1] is a random vector that multiplies by 2 to increase the scope for finding an optimal or almost ideal solution for search agents.
S L ( t + 1 ) _ = P ( t ) _ D i s t _ . c _
where (t + 1) refers to the succeeding iteration, and c has fallen linearly between 2 and 0 throughout all iterations, as the leader of the sea lion must travel around the target prey. When a sea lion spots a victim, the rest of the group forms a ring around it and attacks [29,42]. Equations (41)–(43) are used to represent this behavior:
S P l e a d e r _ = | ( V 1 _ ( 1 + V 2 _ ) ) / V 2 _ ) |
V 1 _ = s i n θ
V 2 _ = s i n
where S P l e a d e r _ denotes the sea lion leader’s sound speed, and V 1 _ and V 2 _ denote sound speed in water and air, respectively.
The s i n θ   and   s i n represent the speed of sound reflected in a medium that is same as of air. The speed of sound refracted into the sea for calling members in underwater is given as.
S L _ ( t + 1 ) = | P ( t ) _ S L ( t ) _ | . c o s ( 2 π m ) + P ( t ) _
where SL denotes the distance between the one searching (sea lion) and the best answer (target prey), | | denotes an actual absolute value, and m denotes a random count between −1 and 1. Every sea lion goes in a circular path around the prey (bait ball) to begin looking for food in a bait ball’s periphery. As a result, Cos (2 m) is used mathematically to describe this occurrence. If c _   is more than one, the global search agent will run the SLnO algorithm to find the best global solution. Equations (45) and (46) propose this.
D i s t _ = | 2 B . _ S L _ r n d ( t ) S L ( t ) _ |
S L ( t + 1 ) _ = S L _ r n d ( t ) D i s t _ . c _
where S L _ r n d is a random sea lion taken from the present population. Random solutions are the first element of the proposed SLnO algorithm. Every search agent moves to a new site at random or based on the best solution. For both the exploration as well as exploitation stages, parameter (C) is minimized between 2 and 0 throughout iterations. If the | C _ | value is more than one, the search agent is chosen at random. If | C _ |   is less than one, search agents are likely to be upgrading their sites. Finally, the SLnO algorithm is completed with the fulfillment of a final requirement (see Algorithm 2).
Table 10 shows the time to calculate different optimization strategies, and Table 10 gives statistical analysis of different optimized procedures with regard to different classifiers. Table 11 gives the evaluation measures used in the classification frameworks and Table 13 gives the Performance value of classifiers. From f1–f15 in Table 13, the advantages and shortcomings of the algorithm tested in these benchmarks to acquire a deeper knowledge and insight can be seen. F1–f15 has multimodal functions.
Algorithm 2. Sea Lion Optimization.
Initialize the population
Choose S L _ r n d
Output from the segmentation result is taken as input (i)
Apply fitness function to every search agent.
If the search agent does not belong to any S   _ l e a d e r
If (i < max iter)
Compute S   _ l e a d e r as per the Equation (43)
If ( S   P _ l e a d e r < 0.25)
If (H < 1)
Using Equation (41), update the search agent’s position
Else
Randomly   choose   a   search   agent   S _ r n d .
Using Equation (46), update the current position of the search agent
End if
Else
Using Equation (45), update the current position of the search agent
End if
Go to first if the condition is met
Else
For each search agent, compute the fitness function.
S   _ should be updated to reflect the optimal solution.
Return S   _ , the best solution
End if
End if
Stop
Best   solution   ( S   P _ l e a d e r ) is obtained as the output for training the classifiers.
Figure 7 gives the analytical results for EFA for sea lion, are mainly focused on the parameters such as LA, LT, OS, OP, EI, and MC, and the observed results are obtained based on path estimates, Chi-square, DF, P, CMIN/DF, GFI, SRMR, NFI, RFI, IFI, TLI, CFI, RMSEA, and RMR.
On multiple MRI images, the suggested technique was checked and simulated in Python using Kera’s and TensorFlow. For this experiment, the BRATS dataset was employed. Preprocessing methods are used to eliminate noise and improve image quality in the sample images (Figure 3). Following this process, the tumor area is segmented from images and the segmented findings obtained are compared to the tumor area reported and segmentation limitations in the data set, and are summarized in Table 10.
The performance of existing and proposed techniques is evaluated in this section based on the segmentation and classification of the tumors in MRI images. Table 11 shows the suggested technique compared with current approaches in terms of tumor accuracy, precision, sensitivity, segmentation F measurement, Jaccard index, Dice overlapping index, accuracy, and MCC (Matthews’ correlation coefficient). The comparison is based on a set of 50 MRI scans, which are divided into two groups: normal and abnormal. Of these 50 MRIs, 25 random MRIs were selected for use in the training process, with 15 normal and 10 anomalous images; the remaining 15 MRI images included 9 ordinary ones and 16 anomalous MRIs. In addition, using SLA12 as an optimization technique potentially boosted the efficiency of every network that we used for analysis. Table 12 and Table 13 provide the resulting performance values.
Finally, a classification procedure is applied to obtain the severity stages and better accuracy. By analyzing Table 4, the performance measures such as PSNR, MSE, and SSIM are analyzed for preprocessed brain cancer MRI images, and the values are obtained. As can be seen from Table 12, the experimental result proves that by using the optimized technique, the overall accuracy is high. From our experiments on different images, it is observed that the proposed method works well in both cases when the objects in the image are indistinct and distinct from the background. Each image MCC is high for CNN with the HHO model. This denotes that the proposed study has achieved a better result. Figure 8 gives the experimental output image for the input image for different stages such as preprocessing, segmentation and for sealion optimization.
The experimental findings highlight the efficiency of our suggested technique to address more segmentation challenges by boosting segmentation quality and precision within a minimum period. Table 13 shows a greater precision, sensitivity, and specificity of the sea-lion-based CNN classifiers; the error value is less than the suggested classifier. The results show that the suggested method performs quite well. The implementation of the proposed tumor detection approach has been shown to enhance the effectiveness and precision of clinical practice. The time analyses of several deep models are shown in Table 10. Table 13 represents the overall model analysis with regard to computational time. The precision of different models is illustrated in Figure 9a,b which exhibits the error of different models in which DenseNet has less error and more accuracy. Figure 10a illustrates the sensitivity and specificity of different models, and Figure 10b displays the F1 score with regard to different models in which DenseNet offers greater performance than the alternatives with sea lion optimization. Figure 11a shows several models without sea lion, and Figure 11b shows different models with sea lion, in which this calculation is reduced on a particular level when optimized.

2.6. Comparative Analysis with the State-of-the-Art Methods

We compare our findings to those of other researchers, including the existing system to demonstrate the validity of our method [45,46,47]. To classify and segment, the CNN algorithm with deep learning method is employed in our proposed system. Table 14 shows the comparative study outcome. The Table 14 provides the comparison findings with the current outcomes of the proposal. The analysis achieves the highest accuracy in our suggested technique.

3. Conclusions

For the diagnosis of brain tumors, an accurate diagnosis with no nonconformities is essential. Any misinterpretation has irreversible consequences. Every year, brain tumor cases are increasing, and this has also brought about a certain increase in strain on medical staff in this area. The segmentation of brain tumor images requires a precise and practical strategy to swiftly resolve growing demand. This document provides optimal models based on this background to increase segmentation and cluster accuracy and get automatic identification without operator intervention. It consists of CNN and a serially connected integrated sea lion optimization technique. In comparison to typical classifiers, this classification provides excellent precision. Different architectures and detailed hyper-parameter tweaking may be designed to further increase the model’s efficiency. In brain tumor patients, the system presented can play an important predictive role in tumor identification. Furthermore, in our proposed method, preprocessing is operated on a small region rather than the entire image, which in turn increases the computational cost and increases the complexity of the circuit. The complexity in the circuit increases the time consumption, so further concentration has been given to rectify this drawback. Moreover, in other clinical fields relating to health imaging, in particular with regards to lung cancer and breast cancer, whose worldwide mortality is quite high, our presented method can play a beneficial role in the early diagnosis of high-risk diseases.

Author Contributions

All authors contributed for the preparation of manuscript. A.S. developed the concept and Manuscript, A.A. supervised and reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Acknowledgments

This research has been supported by The Analytical Center for the Government of the Russian Federation (Agreement No. 70-2021-00143 dd. 01.11.2021, IGK 000000D730321P5Q0002).

Conflicts of Interest

The authors declare no conflict of interest.

Futuristic Way to Multidisciplinary Research

As this paper mentions the story of brain tumors over the past decade, this will be potentially very helpful for the healthcare community (neurosurgeons) and research community for better integration and development of models for even more accurate results in terms of theoretical and practical aspects.

Bibliometric Inferences

This paper gives much a brief account of various publications put forwarded by many active researchers. In addition, it provides a brief review of various researchers who bring great innovations to the diagnosis of brain tumors. Therefore, for this, an essential tool is required for getting a clear view of this area.

Available Datasets

This paper brings over two different datasets that are available for training and testing, such as the DL model, and will provide support for the research community for better integrated models.

Limitations and Future Enhancements

The most advanced technology (i.e., DL) has certain limitation for the healthcare community that need to be addressed and taken care of. The limitations involve the enormous volume of data that needs to be fed in, the quality of data for better processing, dealing with natural data, and the perpetuality of the domain, and these can be addressed by using other features, encrypting every piece of data, ensuring the privacy of the system that is about to develop, and keeping DL to the right path by incorporating the knowledge of experts. Creating DL systems that are time sensitive will lead to a better DL model that could effectively support for the healthcare system.

References

  1. El-Dahshan, E.S.A.; Mohsen, H.M.; Revett, K.; Salem, A.B.M. Computer-aided diagnosis of human brain tumor through MRI: A survey and a new algorithm. Exp. Syst. Appl. 2014, 41, 5526–5545. [Google Scholar] [CrossRef]
  2. Saba, T.; Mohamed, A.S.; El-Affendi, M.; Amin, J.; Sharif, M. Brain tumor detection using fusion of hand crafted and deep learning features. Cogn. Syst. Res. 2020, 59, 221–230. [Google Scholar] [CrossRef]
  3. Logeswari, T.; Karnan, M. An improved implementation of brain tumor detection using segmentation based on the hierarchicaself-organizing map. Int. J. Comput. Theory Eng. 2010, 2, 591. [Google Scholar] [CrossRef] [Green Version]
  4. Kabir-Anaraki, A.; Ayati, M.; Kazemi, F. Magnetic resonance imaging-based brain tumor grades classification and grading via convolutional neural networks and genetic algorithms. Biocybern. Biomed. Eng. 2019, 39, 63–74. [Google Scholar] [CrossRef]
  5. Jana, G.C.; Swetapadma, A.; Pattnaik, P.K. Enhancing the performance of motor imagery classification to design a robust brain computer interface using feed forward back-propagation neural network. Ain Shams Eng. J. 2018, 9, 2871–2878. [Google Scholar] [CrossRef]
  6. Sidhu, H.S.; Benigno, S.; Ganeshan, B.; Dikaios, N.; Johnston, E.W.; Allen, C.; Kirkham, A.; Groves, A.M.; Ahmed, H.U.; Emberton, M.; et al. Textural analysis of multiparametric MRI detects transition zone prostate cancer. Eur. Radiol. 2017, 27, 2348–2358. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Afshar, P.; Mohammadi, A.; Plataniotis, K.N. Brain tumor type classification via capsule networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Athens, Greece, 2018; pp. 3129–3133. [Google Scholar]
  8. El-Omari, N.K.T. Sea Lion Optimization Algorithm for Solving the Maximum Flow Problem. IJCSNS 2020, 20, 30. [Google Scholar]
  9. Kabir, M.A. Automatic brain tumor detection and feature extraction from MRI image. Sci. World J. 2020, 8, 695–711. [Google Scholar]
  10. Deepak, S.; Ameer, P.M. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef] [PubMed]
  11. Banerjee, S.; Mitra, S.; Shankar, B.U. Automated 3D segmentation of brain tumor using visual saliency. Inf. Sci. 2018, 424, 337–353. [Google Scholar] [CrossRef]
  12. Arasi, P.R.E.; Suganthi, M. A Clinical Support System for Brain Tumor Classification Using Soft Computing Techniques. J. Med. Syst. 2019, 43, 1–11. [Google Scholar] [CrossRef] [PubMed]
  13. Sathi, K.A.; Islam, M.S. Hybrid Feature Extraction Based Brain Tumor Classification using an Artificial Neural Network. In Proceedings of the 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA), Greater Noida, India, 30–31 October 2020; pp. 155–160. [Google Scholar] [CrossRef]
  14. Jiachi, Z.; Shen, X.; Zhuo, T.; Zhou, H. Brain tumor segmentation based on refined fully convolutional neural networks with a hierarchical dice loss. arXiv 2017, arXiv:1712.09093. [Google Scholar]
  15. Gumaei, A.; Hassan, M.M.; Hassan, M.R.; Alelaiwi, A.; Fortino, G. A hybrid feature extraction method with regularized extreme learning machine for brain tumor classification. IEEE Access 2019, 7, 36266–36273. [Google Scholar] [CrossRef]
  16. Nyoman, A.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain tumor classification using convolutional neural network. In Proceedings of the World Congress on Medical Physics and Biomedical Engineering 2018, Prague, Czech Republic, 3–8 June 2018; Springer: Singapore, 2019; pp. 183–189. [Google Scholar]
  17. Mohan, N. Tumor Detection From Brain MRI Using Modified Sea Lion Optimization Based Kernel Extreme Learning Algorithm. Int. J. Eng. Trends Technol. 2020, 68, 84–100. [Google Scholar] [CrossRef]
  18. Roy, S.; Bandyopadhyay, S.K. Detection and Quantification of Brain Tumor from MRI of Brain and it’s Symmetric Analysis. J. Inf. Commun. Technol. 2012, 2, 477–483. [Google Scholar]
  19. Mohsen, H.; El-Dahshan, E.-S.A.; El-Horbaty, E.-S.M.; Salem, A.-B.M. Classification using deep learning neural networks for brain tumors. Future Comput. Inform. J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  20. Özyurt, F.; Sert, E.; Avci, E.; Dogantekin, E. Brain tumor detection based on Convolutional Neural Network with neutrosophic expert maximum fuzzy sure entropy. Measurement 2019, 147, 106830. [Google Scholar] [CrossRef]
  21. Praveen, G.B.; Agrawal, A. Hybrid approach for brain tumor detection and classification in magnetic resonance images. Commun. Control Intell. Syst. CCIS 2015, 2015, 162–166. [Google Scholar] [CrossRef]
  22. Meenakshi, R.; Anandhakumar, P. Brain Tumor Identification in MRI with BPN Classifier and Orthonormal Operators. Eur. J. Sci. Res. 2012, 85, 559–569. [Google Scholar]
  23. Aswathy, S.U.; Devadhas, G.G.; Kumar, S.S. Brain tumor detection and segmentation using a wrapper based genetic algorithm for optimized feature set. Clust. Comput. 2018, 22, 13369–13380. [Google Scholar] [CrossRef]
  24. Sharma, M.; Purohit, G.N.; Mukherjee, S. Information Retrieves from Brain MRI Images for Tumor Detection Using Hybrid Technique K-means and Artificial Neural Network (KMANN). In Networking Communication and Data Knowledge Engineering; Springer: Singapore, 2018. [Google Scholar] [CrossRef]
  25. Asokan, A.; Anitha, J. Adaptive Cuckoo Search based optimal bilateral filtering for denoising of satellite images. ISA Trans. 2020, 100, 308–321. [Google Scholar] [CrossRef] [PubMed]
  26. Rachmad, A.; Chamidah, N.; Rulaningtyas, R. Image Enhancement Sputum Containing Mycobacterium Tuberculosis Using A Spatial Domain Filter. IOP Conf. Ser. Mater. Sci. Eng. 2019, 546, 052061. [Google Scholar] [CrossRef]
  27. Karthik, R.; Menaka, R.; Chellamuthu, C. A comprehensive framework for classification of brain tumour images using SVM and curvelet transform. Int. J. Biomed. Eng. Technol. 2015, 17, 168. [Google Scholar] [CrossRef]
  28. Nayak, D.R.; Dash, R.; Majhi, B.; Prasad, V. Automated pathological brain detection system: A fast discrete curvelet transform and probabilistic neural network based approach. Expert Syst. Appl. 2017, 88, 152–164. [Google Scholar] [CrossRef]
  29. Bhadauria, H.; Dewal, M. Medical image denoising using adaptive fusion of curvelet transform and total variation. Comput. Electr. Eng. 2013, 39, 1451–1460. [Google Scholar] [CrossRef]
  30. Kumar, R.R.; Kumar, A.; Srivastava, S. Anisotropic Diffusion Based Unsharp Masking and Crispening for Denoising and Enhancement of MRI Images. In Proceedings of the 2020 International Conference on Emerging Frontiers in Electrical and Electronic Technologies (ICEFEET), Patna, India, 10–11 July 2020; IEEE: Patna, India, 2020; pp. 1–6. [Google Scholar]
  31. Anoop, B.N.; Joseph, J.; Williams, J.; Jayaraman, S.; Sebastian, A.M.; Sihota, P. A prospective case study of high boost, high frequency emphasis and two-way diffusion filters on MR images of glioblastoma multiforme. Australas. Phys. Eng. Sci. Med. 2018, 41, 415–427. [Google Scholar] [CrossRef] [PubMed]
  32. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  33. Ahmed, M.; Seraj, R.; Islam, S.M.S. The k-means Algorithm: A Comprehensive Survey and Performance Evaluation. Electronics 2020, 9, 1295. [Google Scholar] [CrossRef]
  34. Zhou, X.; Tang, Z.; Xu, W.; Meng, F.; Chu, X.; Xin, K.; Fu, G. Deep learning identifies accurate burst locations in water distribution networks. Water Res. 2019, 166, 115058. [Google Scholar] [CrossRef]
  35. Yanda, M.; Wei, M.; Gao, D.; Zhao, Y.; Yang, X.; Huang, X.; Zheng, Y. CNN-GCN aggregation enabled boundary regression for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Cham, Switzerland, 2020; pp. 352–362. [Google Scholar]
  36. Toğaçar, M.; Ergen, B.; Cömert, Z. BrainMRNet: Brain tumor detection using magnetic resonance images with a novel convolutional neural network model. Med. Hypotheses 2020, 134, 109531. [Google Scholar] [CrossRef]
  37. Lodh, C.C.; Mahanty, C.; Kumar, R.; Mishra, B.K. Brain Tumor Detection and Classification Using Convolutional Neural Network and Deep Neural Network. In Proceedings of the 2020 International Conference on Computer Science, Engineering and Applications (ICCSEA), Gunupur, India, 13–14 March 2020; IEEE: Gunupur, India, 2020; pp. 1–4. [Google Scholar]
  38. Xavier, G.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  39. Sharif, M.; Amin, J.; Nisar, M.W.; Anjum, M.A.; Muhammad, N.; Shad, S. A unified patch based method for brain tumor detection using features fusion. Cogn. Syst. Res. 2020, 59, 273–286. [Google Scholar] [CrossRef]
  40. Badža, M.M.; Barjaktarović, M. Segmentation of Brain Tumors from MRI Images Using Convolutional Autoencoder. Appl. Sci. 2021, 11, 4317. [Google Scholar] [CrossRef]
  41. Xing, Y.; Yu, J.; Zhang, F.; Gong, Y. Image Denoising Algorithm Based on Local Adaptive Nonlinear Response Diffusion. Mater. Sci. Eng. 2020, 790, 012103. [Google Scholar] [CrossRef]
  42. Wu, W.; Li, D.; Du, J.; Gao, X.; Gu, W.; Zhao, F.; Feng, X.; Yan, H. An Intelligent Diagnosis Method of Brain MRI Tumor Segmentation Using Deep Convolutional Neural Network and SVM Algorithm. Comput. Math. Methods Med. 2020, 2020, 4916497. [Google Scholar] [CrossRef]
  43. Rao, B.S. Accurate leukocoria predictor based on deep VGG-net CNN technique. IET Image Process. 2020, 14, 2241–2248. [Google Scholar]
  44. Naoya, T.; Mitsufuji, Y. D3Net: Densely connected multidilatedDenseNet for music source separation. arXiv 2020, arXiv:2010.01733. [Google Scholar]
  45. Rao, I.V.; Rao, V.M. Massive MIMO perspective: Improved sea lion for optimal antenna selection. Evol. Intell. 2020, 14, 1831–1845. [Google Scholar] [CrossRef]
  46. Badža, M.M.; Barjaktarović, M. Classification of Brain Tumors from MRI Images Using a Convolutional Neural Network. Appl. Sci. 2020, 10, 1999. [Google Scholar] [CrossRef] [Green Version]
  47. Veeramuthu, A.; Meenakshi, S.; Darsini, V.P. Brain Image Classification using Learning Machine Approach and Brain Structure Analysis. Procedia Comput. Sci. 2015, 50, 388–394. [Google Scholar] [CrossRef] [Green Version]
  48. Seetha, J.; Raja, S.S. Brain Tumor Classification Using Convolutional Neural Networks. Biomed. Pharmacol. J. 2018, 11, 1457–1461. [Google Scholar] [CrossRef]
  49. Krishnakumar, S.; Manivannan, K. Effective segmentation and classification of brain tumor using rough K means algorithm and multi kernel SVM in MR images. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 6751–6760. [Google Scholar] [CrossRef]
  50. Jun, W.; Zhang, Q.; Liu, M.; Xiao, Z.; Zhang, F.; Geng, L.; Liu, Y.; Wang, W. Diabetic macular edema grading based on improved Faster R-CNN and MD-ResNet. Signal Image Video Process. 2020, 15, 743–751. [Google Scholar]
Figure 1. Schematic representation of the proposed research study.
Figure 1. Schematic representation of the proposed research study.
Axioms 11 00015 g001
Figure 2. Pictorial representation of online data sources most frequently used by world-wide researchers.
Figure 2. Pictorial representation of online data sources most frequently used by world-wide researchers.
Axioms 11 00015 g002
Figure 3. (a) Depiction of three different input image grades. (b) Image of bilateral filtering. The curvelet transform image is shown in (c). (d) Image with an adaptive histogram equalized histogram. (e) Adaptive boosted anisotropic diffusion filter.
Figure 3. (a) Depiction of three different input image grades. (b) Image of bilateral filtering. The curvelet transform image is shown in (c). (d) Image with an adaptive histogram equalized histogram. (e) Adaptive boosted anisotropic diffusion filter.
Axioms 11 00015 g003
Figure 4. Segmented results of MRI images using hybrid structure of modified K-means and HHO. (a) Input image, (b) preprocessed image, (c) adaptive diffusion filter, (d) segment output after applying hybrid modified K-means and HHO.
Figure 4. Segmented results of MRI images using hybrid structure of modified K-means and HHO. (a) Input image, (b) preprocessed image, (c) adaptive diffusion filter, (d) segment output after applying hybrid modified K-means and HHO.
Axioms 11 00015 g004
Figure 5. The framework on multigrade tumor segmentation and classification using the CNN network.
Figure 5. The framework on multigrade tumor segmentation and classification using the CNN network.
Axioms 11 00015 g005
Figure 6. Diagrammatic representation on cascaded CNN architecture on brain tumor segmentation.
Figure 6. Diagrammatic representation on cascaded CNN architecture on brain tumor segmentation.
Axioms 11 00015 g006
Figure 7. Exploratory factor analysis for sea lion optimization algorithm.
Figure 7. Exploratory factor analysis for sea lion optimization algorithm.
Axioms 11 00015 g007
Figure 8. Experimental output for the given input brain image (a) gives the input meningioma brain image (b)preprocessed brain image after applying CLAHE and anisotropic diffusion filter (c) gives the segmented output of the image (d) parameter space output after applying the sealion optimization and figure (e) gives the object space output.
Figure 8. Experimental output for the given input brain image (a) gives the input meningioma brain image (b)preprocessed brain image after applying CLAHE and anisotropic diffusion filter (c) gives the segmented output of the image (d) parameter space output after applying the sealion optimization and figure (e) gives the object space output.
Axioms 11 00015 g008
Figure 9. (a) Confusion matrix without optimization. (b) Confusion matrix with optimization.
Figure 9. (a) Confusion matrix without optimization. (b) Confusion matrix with optimization.
Axioms 11 00015 g009
Figure 10. (a) Sensitivity, specificity vs. models. (b) F1-score vs. models.
Figure 10. (a) Sensitivity, specificity vs. models. (b) F1-score vs. models.
Axioms 11 00015 g010
Figure 11. (a) Computation time without sea lion. (b) Computation time with sea lion.
Figure 11. (a) Computation time without sea lion. (b) Computation time with sea lion.
Axioms 11 00015 g011
Table 1. Overview of recent work using CNN method used for tumors detection.
Table 1. Overview of recent work using CNN method used for tumors detection.
AuthorMethodologyProposed Solution and ApproachSoftware/Tools/Languages/Simulation/Implementation LibrariesEvaluationChallenges
Md Ahasan Kabir (2020)
[9]
GLCM for feature extraction and ANN for classificationGA for feature selection, ANN classification modelNot mentionedAccuracy: 99.5% for the BRATS, 98.3% for augmented BRATS
Specificity: 97%
High dimensionality of matrix
S. Deepak
(2019)
[10]
Deep convolutional
neural networks using transfer learning
GoogLeNet,
deep CNN-SVM
MATLAB 2018bAccuracy: 97.1%Poor performances of the transfer learning model, misclassification of samples from the meningioma class.
Banerjee et al., (2018)
[11]
Deep convolutional neural networks (ConvNets) trained on multi-sequence MR images.DenseNet, VGG16Terser flow and PythonAccuracy: 97%Computational complexity is high
Arasi,1 P et al., (2019)
[12]
GLCM feature extraction method. Lion-optimized boosting SVMGenetic optimized median, hierarchical fuzzy clustering algorithm, SVMMATLAB softwareAccuracy: 97.69%
Specificity: 96.7%
The correlation of pixels is not considered
Khaleda A Sathi et al., (2020)
[13]
GLCM for feature extraction and ANN for classificationGabor, DWTNot mentionedAccuracy: 97.99
Specificity: 98.34
Takes more time in image analysis
Zhou et al., (2018)
[14]
Convolutional neural
networks
DenseNet-RNN,
DenseNet-LSTM,
DenseNet-DenseNET
Tensor Flow, Nvidia Titan
Xp GPU
Accuracy: 92.13%Time consumption is high
Abdu, Hassan (2019)
[15]
PCA-NGISTPCA-NGIST for feature extraction and RELM for classificationMATLAB Accuracy: 94.23
Specificity: 96.56
Information loss
Overfitting problem
Nyoman et al.,
(2019)
[16]
Convolutional neural
network
AlexNet, VGG16, ResNetMATLABAccuracy: 84.19%Not satisfactory for noisy, nonuniform, and high-intensity images.
Narendra Mohan, (2020)
[17]
Modified sea-lion-optimization-based KELASGLDM and LESH based feature extraction and MSLO-based KELM for classificationNot mentionedAccuracy: 94.67%
Specificity: 97.78%
Slow process
Roy, (2012)
[18]
Modular approach to
solve MRI
segmentation
Symmetry analysisNot mentionedAccuracy: 91%Time
Consuming
Mohsen et al., (2018)
[19]
Deep learning neural networks Feature extraction using
DWT and reduction using PCA technique
and classification using DNN
MATLAB R2015aAccuracy: 98
Specificity: 97.45
Time
Consuming
Fatih Özyurt et al. (2019)
[20]
Convolutional neural
network
NS-EMFSE-CNN, SVM and KNNNot mentionedAccuracy: 95.62%Long calculation time
Praveen G.B., Anita Agrawal, 2015
[21]
Machine learningGLCM for feature extraction & SVM for classificationMATLAB 2014 resolution 256*256 using a 2.2 GHz, I3 windows OS machine.Accuracy: 96.63
Specificity: 98.63
Takes more time
Mohsen et al., (2018)
[19]
Deep neural networkPrincipal Components
Analysis (PCA), Discrete Wavelet
Transform (DWT)
Weka 3.9 and MATLAB R2015aAccuracy: 96.97%Long calculation time
Meenakshi, 2012
[22]
BPN classifier and orthonormal
operators
K-means clustering,
BPN classifier
MATLABAccuracy: 92%Accuracy
can be improved
in less time
S. U. Aswathy et al.,
(2018)
[23]
Wrapper-based genetic algorithmTexture-based feature extraction, wrapper-based GA for feature selection, and SVM as classifier.Not mentionedAccuracy: 98.2
Specificity: 97.90
High dimensionality of matrix
Afshar et al., (2018)
[7]
Convolutional neural
networks (CNNs)
Capsule networks
(CapsNets)
Python 2.7 and Keras
library
Accuracy: 86.56%Applicable for tumor core only
Manorama Sharma, G.N. Purohit and Saurabh Mukherjee, 2018
[24]
K-means and artificial neural network (KMANN)GLCM is used for feature extraction and tumor classification by (KMANN)Not mentionedAccuracy: 98.65
Specificity: 97.89
Information loss
S. Deepak
(2019)
[10]
Deep convolutional
neural networks using transfer learning
GoogLeNet,
deep CNN-SVM
MATLAB 2018bAccuracy: 97.1%Poor performances of the transfer learning model, misclassification of samples from the meningioma class
Table 2. Different sources of data and techniques of acquisition.
Table 2. Different sources of data and techniques of acquisition.
Sl.NoDataset Sources Link (Access on: 6 April 2020)
1Medical images from BRAINIXT1w, T2whttps://www.medicalimages.com/search/brain.html
2BraTS 2018T1 w, T1 contrast-enhanced MRI, T2 MRI, and T2 FLAIR MRIhttps://www.med.upenn.edu/sbia/brats2017/data.html
3Harvard Medical School websiteT1w, T2w, FLAIR, ASL, SWI, time of flight, resting state BOLD, and DTI sequences.http://med.harvard.edu/AANLIB/
4TCGA-GBM, TCGA-LGGT1w, T2w, FLAIRhttps://wiki.cancerimagingarchive.net/display/Public/TCGA-LGG
5BraTS-2018 ISLES-2018T1w, T2w, FLAIR, ASL, SWI, time of flight, resting state BOLD, and DTI sequences.http://www.isles-challenge.org/
6IBSR dataset CyprusT1w, T2w, FLAIRhttp://www.medinfo.cs.ucy.ac.cy/
7BraTS 2015 and BraTS 2017T1 w, T1 contrast-enhanced MRI, T2 MRI, and T2 FLAIR MRIhttps://www.med.upenn.edu/sbia/brats2017/data.html
8Brain web (simulated brain database)T1w, T2w, FLAIRhttp://brainweb.bic.mni.mcgill.ca/brainweb/
9Brain MRIT1-w, FLAIRhttps://figshare.com/articles/dataset/brain_tumor_dataset/1512427
10Harvard Medical School T1w, T2whttp://www.med.harvard.edu/aanlib/home.html
Table 3. Parameters for measuring preprocessed images. The peak signal-to-noise ratios for various filters in five abnormal and normal images are tabulated.
Table 3. Parameters for measuring preprocessed images. The peak signal-to-noise ratios for various filters in five abnormal and normal images are tabulated.
Sl.no Peak Signal-to-Noise Ratio
BFCTAHEBADF
Abnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal Images
132.973532.6536632.1670733.3729934.3723333.2887040.6528941.84186
232.7005632.8095530.6937334.1284534.1805733.8090639.1746842.63007
333.3656032.7103030.9786933.3744834.3147132.9553239.4679941.82132
433.1259732.6925731.5790934.0328534.3822033.6294140.0633942.52042
532.4492232.6756528.0570434.0167334.0517333.5837236.5481742.55479
Table 4. (a) Gives the mean square error for various filters in five abnormal and normal images. (b) Gives the structural similarity index measure for various filters in five images of abnormal and normal. (c) The signal-to-noise ratios for various filters in five abnormal and normal images are tabulated as follows.
Table 4. (a) Gives the mean square error for various filters in five abnormal and normal images. (b) Gives the structural similarity index measure for various filters in five images of abnormal and normal. (c) The signal-to-noise ratios for various filters in five abnormal and normal images are tabulated as follows.
(a)
Sl.no Mean Square Error
BFCTAHEBADF
Abnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal Images
12.82103.52945.32722.65812.84173.56882.61621.3523
23.24082.77867.59741.95992.85662.81533.75250.9324
32.15033.41029.47812.54662.87344.33404.64761.4266
42.34363.50457.71072.20842.83643.03313.91811.0669
54.43403.366011894911.67222.07113.08675.98220.9888
(b)
Sl.no Structural Similarity Index Measure
BFCTAHEBADF
Abnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal Images
10.9426380.9408070.9126080.8936600.97060.95770.9757810.987964
20.9374720.9434800.874530.9124530.95970.93880.9540540.957021
30.9537440.9446870.9325350.9043030.98090.97800.9975240.949472
40.9522830.9407350.8989870.9033510.94150.95800.9629980.997604
50.9284840.9392510.8544470.8636880.99920.92800.9322560.929219
(c)
Sl.no Signal-to-Noise Ratio
BFCTAHEBADF
Abnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal ImagesAbnormal ImagesNormal Images
132.652231.768628.958131.764734.818328.677032.970435.7544
231.967832.559027.371733.435933.677331.515131.344937.4646
333.795632.045526.426231.467534.466326.943330.471035.4479
433.423531.929827.283932.871434.875530.496731.157536.8703
530.692531.980725.337533.141432.918030.284929.371937.1594
Table 5. Gives the statistical parameters for calculating feature extraction.
Table 5. Gives the statistical parameters for calculating feature extraction.
Extracted FeaturesEquation
Mean M = ( 1 m × n ) x = 0 m = 1 y = 0 n = 1 f ( x , y )
Standard Deviation S D ( σ ) = ( 1 m × n   ) x = 0 m = 1 y = 0 n = 1 ( f ( x , y ) M ) 2
Entropy E = x = 0 m = 1 y = 0 n = 1 f ( x , y )   log 2 f ( x , y )  
Skewness S k ( X ) = ( 1 m n ) ( f ( x , y ) M ) 3 S D 3
Kurtosis K urt   ( x ) = ( 1 m n ) ( f ( x , y ) M ) 4 S D 4
Contrast C o n = k = 0 m 1 y = 0 n 1 ( x y ) 2 f ( x + y )
Correlation C orr = x = 0 m 1 y = 0 n 1 ( x , y ) f ( x , y ) M x M y σ x σ y
Coarseness C ness = 1 2 m + n k = 0 m 1 y = 0 n 1 f ( x , y )
Table 6. Computational time analysis for the different feature extraction methods.
Table 6. Computational time analysis for the different feature extraction methods.
MethodsSSOASSOPSOGAGSAFFPropSLA12
Computation time (s)4.51003.77604.37974.43889.52493.92373.81772.3456
Table 7. First-order and second-order statistical analysis based on performance measure.
Table 7. First-order and second-order statistical analysis based on performance measure.
Statistical Analysis Based on Performance Measures for First-Order Statistics
Performance MeasureASSOSSOGAPSOFFGSAPropSLA12
Mean3.23522.64442.46712.43461.51871.52752.39222.3462
Median2.80122.29702.39762.46192.45031.49062.35922.3309
Std Deviation1.60380.59800.99270.93390.91860.25220.93392.3307
Best2.89012.66093.61693.76843.75031.76633.65553.6294
Worst1.00251.21101.34481.17261.00721.35581.02221.0238
Statistical Analysis Based on Performance Measures for Second-Order Statistics
Performance MeasureASSOSSOPSOGAGSAFFPropSLA12
Entropy0.94563.0330.46572.0991.1222.7860.87650.6534
Skewness0.00550.09870.09840.01050.00510.02000.07890.0011
Kurtosis2.34 × 1061.34 × 1061.88 × 1062.44 × 1063.34 × 1061.33 × 1052.04 × 1052.74 × 105
Contrast0.26500.65740.98450.45630.78340.34520.93420.3452
Correlation0.98640.98350.98230.96740.83450.83450.98660.9875
Coarseness8.87611.8979.86510.87611.976412.76511.99813.856
Table 8. Comparison of different optimization techniques in feature extraction.
Table 8. Comparison of different optimization techniques in feature extraction.
FunctionsASSOPSOSSOGAGSAFFPropSLA12
Avg/Std.Avg/Std.Avg/Std.Avg/Std.Avg/Std.Avg/Std.Avg/Std.Avg/Std.
F128.3976/323.72938.9854/372.9322−12,389.05/382.57026329.3456/321.821−6981.15/848.8447−14,219.09/472.6603737.9524/371.954225.23801/295.86792
F247.6299/23.423812.3789/42.42313.78 × 1015/1.44 × 101446.0247/24.240832.30133/8.735742.66 × 1015/2.43 × 101412.3578/42.32144.12021/1.90716
F35.119 × 105/3.244 × 1054.78 × 105/2.04 × 1054.32 × 1015/2.37 × 10155.012 × 105/3.024 × 1052.85 × 105/1.81 × 1055.00 × 1015/2.94 × 10154.89 × 105/2.02 × 1056.42 × 1017/6.31 × 1015
F41.9688/2.85321.2776/1.98210.00285/0.007611.9574/2.35410.00992/0.011160.00496/0.008761.2783/1.98210.07942/1.99659
F51.6734/2.52371.2731/1.6231−1.00040/0.001901.9872/2.34573.81 × 1011/3.86 × 10110.00067/0.002991.2701/1.62480.98765/0.10078
F62.1748/2.89451.3472/1.48370.00037/0.000462.1348/2.73210.00366/0.020050.00287/0.003961.3472/1.42150.98937/0.42215
F71.3337/0.98771.0367/11.98011.00641/0.136221.0227/0.98671.03113/0.181471.02163/0.145841.0367/12.3181.06157/15.13597
F80.3199/0.81990.9812/1.47830.00052/0.000230.3156/0.81790.00056/0.000240.00061/0.000440.9043/1.37820.28946/0.71108
Table 9. Overall network of CNN.
Table 9. Overall network of CNN.
Column 1Column 2Column 3Column 4Column 5
StructureCurrent layerTypeOutput sizePrevious layer
Conv_lConv_l_1Conv2dH * W * 64Input
Conv_1_2Conv2dH’W * 64Conv_1_1
Conv_2Conv_2_1Maxpool+Conv2d(1/2) * (W/2) * 128Conv_1_2
Conv_2_2Conv2d(H/2) * (W/2) * 128Conv_2_1
Conv_3Conv_3_1Maxpool+Conv2d(H/4) * (W)4) * 256Conv_2_2
Conv_3_2Conv2d(H/4) * (W/4) * 256Conv_3_1
Conv_3_3Conv2d(H/4) * (W)4) * 256Conv_3_2
Conv_4Conv_4_1Maxpool+Conv2d(H/8) * (W/8) * 512Conv_3_3
Conv_4_2Conv2d(H/8) * (W/8) * 512Conv_4_1
Conv_4_3Conv2d(H/8) * (W/8) * 512Conv_4_2
Conv_5Conv_5_1Multi-scale fusion(H/8) * (W/8) * 256Conv_4_3
Upsmaple1Upsmaple1Upsample(H/4) * (W/4) * 256Conv_5_1
Conv_6Conv6_1Multi-scale fusion(H/4) * (W)4) * 128Upsmaple1
Upsmaple2Upsmaple2Upsample(H/2) * (W/2) * 128Conv_6_1
Conv_7Conv_7_1Multi-scale fusion(H/2) * (W/2) * 2UpsmapleZ
AvgpoolAvgpoolGlobalAveragePool2Conv_7_1
OutputOutputPixel-wise product(H/2) * (W/2) * 1Conv_7_1,Avgpool
Table 10. Analysis on computational time for different classifiers.
Table 10. Analysis on computational time for different classifiers.
MethodsANNSVMPNNCNNSea Lion Optimization with CNN
ResNetVGGDenseNet
Computational Time (min)3.4584.6324.2374.1122.87012.27002.2345
Table 11. Evaluation measures used in the classification frameworks.
Table 11. Evaluation measures used in the classification frameworks.
Performance MeasureFormulae
Sensitivity T P T P + F N × 100%
Accuracy T P + T N T P + F P + F N + T N × 100%
Precision T P T P + F P × 100%
F-measure 2 P r e c i s i o n R e c a l l   P r e c i s i o n + R e c a l l  
Jaccard Index J ( P , Q ) = S ( P Q ) S ( P Q )
Dice Overlap Index (DOI) D ( P , Q ) = 2 X P Q P + Q
Accuracy T P + T N T P + F P + F N + T N × 100%
MCC T P × T N F P × F N ( T P + F P ) ( F N + T N ) ( T N + F P ) ( T N + F N )
Table 12. Performance evaluation of proposed segmentation models.
Table 12. Performance evaluation of proposed segmentation models.
METRICSK-MeansModified K-MeansHHOCNN with HHO
Image 1Dice Coefficient0.98180.98630.96580.9868
Jaccard Coefficient0.96420.97290.92560.9744
MCC0.98140.98590.91030.9863
Accuracy0.99910.99920.99800.9995
Sensitivity0.94560.92470.89340.9127
Specificity0.93460.89230.90120.9983
Image 2Dice Coefficient0.78460.91330.92540.9442
Jaccard Coefficient0.64560.84050.96750.8945
MCC0.76470.91630.93240.9447
Accuracy0.99890.99910.89120.9995
Sensitivity0.98230.92140.98230.9847
Specificity0.92340.78650.91670.9882
Image 3Dice Coefficient0.79640.87140.92170.9327
Jaccard Coefficient0.66170.77210.93570.8732
MCC0.81170.87750.98120.9346
Accuracy(%)0.99570.99720.98430.9973
Sensitivity0.89670.98760.96730.9657
Specificity0.89120.86790.91240.9878
Image 4Dice Coefficient0.90880.96160.94520.9882
Jaccard Coefficient0.83280.92600.956320.9774
MCC0.91140.96170.943520.9884
accuracy0.99740.99860.92140.9987
Sensitivity0.89670.92310.95630.9972
Specificity0.89990.92130.89410.9657
Image 5Dice Coefficient0.98180.98630.89120.9867
Jaccard Coefficient0.96420.97290.98120.9745
MCC0.98140.98590.91230.9862
accuracy0.99910.99920.88920.9997
Sensitivity0.91240.98230.88240.9878
Specificity0.93420.99120.87230.9342
Image 6Dice Coefficient0.78460.91330.91260.9443
Jaccard Coefficient0.64560.84050.89090.8941
MCC0.76470.91630.90120.9443
accuracy0.99890.99910.91430.9995
Sensitivity0.91120.99820.88720.9126
Specificity0.89230.93650.88910.9457
Table 13. Performance value of classifiers.
Table 13. Performance value of classifiers.
MetricsCNNSea Lion Optimization with CNN
ResNetVGGDenseNet
Accuracy0.96780.96510.97560.9867
Error0.04450.04530.03670.0289
Sensitivity0.96670.96540.97730.9830
Specificity0.98330.965350.97330.9834
Precision0.96020.960260.97440.97612
F1-Score0.96980.960020.97220.978
Table 14. Comparison with state-of-the-art methods for our proposed system.
Table 14. Comparison with state-of-the-art methods for our proposed system.
MethodsAccuracy (%)Sensitivity (%)Precision (%)Specificity (%)
Wu et al. (2020)
[42]
96.195.496.9196.7
Lodh et al. (2020)
[37]
96.296.3996.39196.6
Badza et al. (2020)
[46]
9595.0594.2494.2
Veeramuthu et al. (2015)
[47]
96.59696.7895
Seetha et al. (2018)
[48]
96.0897.019796.23
Krishnakumar et al. (2021)
[49]
8585.03485.8984.3
Jun et al. (2020)
[50]
89.5990.7785.9788.4
Proposed Method
SL+resnet 5096.596.5496.0296.54
SL+VGG9797.7396.02697.33
SL+DenseNet9898.3097.4498.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sukumaran, A.; Abraham, A. Automated Detection and Classification of Meningioma Tumor from MR Images Using Sea Lion Optimization and Deep Learning Models. Axioms 2022, 11, 15. https://doi.org/10.3390/axioms11010015

AMA Style

Sukumaran A, Abraham A. Automated Detection and Classification of Meningioma Tumor from MR Images Using Sea Lion Optimization and Deep Learning Models. Axioms. 2022; 11(1):15. https://doi.org/10.3390/axioms11010015

Chicago/Turabian Style

Sukumaran, Aswathy, and Ajith Abraham. 2022. "Automated Detection and Classification of Meningioma Tumor from MR Images Using Sea Lion Optimization and Deep Learning Models" Axioms 11, no. 1: 15. https://doi.org/10.3390/axioms11010015

APA Style

Sukumaran, A., & Abraham, A. (2022). Automated Detection and Classification of Meningioma Tumor from MR Images Using Sea Lion Optimization and Deep Learning Models. Axioms, 11(1), 15. https://doi.org/10.3390/axioms11010015

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop