Next Article in Journal
Sepsis-Induced Coagulopathy: An Update on Pathophysiology, Biomarkers, and Current Guidelines
Previous Article in Journal
Characterization and microRNA Expression Analysis of Serum-Derived Extracellular Vesicles in Severe Liver Injury from Chronic HBV Infection
Previous Article in Special Issue
A Prospective Approach to Integration of AI Fracture Detection Software in Radiographs into Clinical Workflow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images

1
Department of Information Systems Engineering, Mugla Sitki Kocman University, Mugla 48000, Turkey
2
Department of Artificial Intelligence, Mugla Sitki Kocman University, Mugla 48000, Turkey
3
Department of Computer Engineering, Gachon University, Seongnam-si 13120, Republic of Korea
4
Daeyang AI Center, Department of Data Science, College of Software & Convergence Technology, Sejong University, Seoul 05006, Republic of Korea
5
Department of Robotics, Hanyang University, Ansan 15588, Republic of Korea
6
Department of Computer Engineering, Sir Syed University of Engineering & Technology, Karachi 75300, Pakistan
*
Authors to whom correspondence should be addressed.
Life 2023, 13(2), 349; https://doi.org/10.3390/life13020349
Submission received: 6 December 2022 / Revised: 17 January 2023 / Accepted: 22 January 2023 / Published: 28 January 2023

Abstract

:
Brain MR images are the most suitable method for detecting chronic nerve diseases such as brain tumors, strokes, dementia, and multiple sclerosis. They are also used as the most sensitive method in evaluating diseases of the pituitary gland, brain vessels, eye, and inner ear organs. Many medical image analysis methods based on deep learning techniques have been proposed for health monitoring and diagnosis from brain MRI images. CNNs (Convolutional Neural Networks) are a sub-branch of deep learning and are often used to analyze visual information. Common uses include image and video recognition, suggestive systems, image classification, medical image analysis, and natural language processing. In this study, a new modular deep learning model was created to retain the existing advantages of known transfer learning methods (DenseNet, VGG16, and basic CNN architectures) in the classification process of MR images and eliminate their disadvantages. Open-source brain tumor images taken from the Kaggle database were used. For the training of the model, two types of splitting were utilized. First, 80% of the MRI image dataset was used in the training phase and 20% in the testing phase. Secondly, 10-fold cross-validation was used. When the proposed deep learning model and other known transfer learning methods were tested on the same MRI dataset, an improvement in classification performance was obtained, but an increase in processing time was observed.

1. Introduction

The brain, an organ that searches for meaning and self-inquiry, serves as the center of the entire nervous system to control the body’s other organs. Therefore, any abnormality in the brain can endanger human health. Among such abnormalities, brain tumors, hydrocephalus, and cerebral hemorrhage are the most severe. Brain tumors are divided into primary (primary) and secondary (secondary). Primary brain tumors consisting of the brain’s own cells can be benign (benign) or malignant (malignant). Secondary brain tumors occur when cancerous cells that appear in another part of the body spread to the brain. Primary tumors are located in brain tissue, while secondary tumors expand into brain tissue from other parts of the human body through the bloodstream [1].
According to the World Health Organization (WHO), brain tumors can be divided into four grades. Grade 1 and grade 2 tumors describe lower-grade tumors (e.g., meningiomas), while grade 3 and grade 4 tumors consist of more severe ones (e.g., glioma). In clinical practice, the incidence rates of meningioma, pituitary, and glioma tumors are approximately 15%, 15%, and 45%, respectively [2].
A series of physical and neurological examinations are performed to diagnose brain tumors. Diagnosis is made with MR (Magnetic Resonance) and CT (Computerized Tomography), and biopsy and pathological evaluation are performed to confirm the diagnosis. Among all these imaging modalities, MR is considered the most preferred because it is the only non-invasive and non-ionizing modality [3]. As a result, the type and stage of the cancer are learned precisely, and a treatment plan is prepared. Manual review of medical images in this process is time-consuming, hectic, and even error-prone due to patient flow [4]. To solve this problem in the performed study, it is proposed to develop an automated computer-aided diagnostic (CAD) system to ease the workload of classification and diagnosis of brain MRI and act as a tool to assist radiologists and doctors.
The application was carried out using the Python programming language. Based on free and open-source code logic, Python’s standard library, development tools, and many other libraries can be used free of charge as open-source code without needing a license. Another reason the Python language is preferred is that deep learning libraries used in diagnosing brain tumors exist in Python in the background. The study’s preferred deep learning method for image processing is a subset of machine learning where artificial neural networks and algorithms inspired by the human brain learn from data. When we look at the literature, there are many studies on diagnosing and classifying brain tumors. Different data are collected from different data bases i.e through cloud and big data [5,6,7,8] abnd these data center can be accessed through wired and wirelessly. However, the feature vector must be extracted first to establish a model definition or machine learning system in classical machine learning algorithms. Experts in the field are needed to extract this feature vector. These processes take a large amount of time and keep the expert busy.
For this reason, these techniques cannot process raw data without preprocessing and expert assistance. On the other hand, deep learning has made significant progress by eliminating this problem that those working in machine learning have been dealing with for many years. This was achievable because, unlike traditional machine learning and image processing techniques, deep networks perform the learning process on raw data.
As a result, the literature and experimental research concluded that the models created using transfer learning methods could not achieve the expected success in health image data classification, and there were some areas open to improvement. For this reason, the applied models were examined, and their strengths and weaknesses were reported. Then, a new model structure was created on the CNN’s architectural structure in which the strong points in the models were kept and the weak points were removed. This model architecture uses a convolutional network structure with both dense layers. The proposed new deep learning model improves classification performance and other known transfer learning methods, but an increase in processing time was observed.This study aims to develop a binary classification method for brain tumors (meningiomas and gliomas) and multiclass classification using SoftMax and KNN, non-deep methods with handcrafted features, as well as CNN deep learning methods with transfer learning. The main contributions of the article include:
  • In the study, a modified CNN-based system was proposed to increase the accuracy in the classification of brain tumors over previously labeled data.
  • Comparison of the proposed model with three different models (simple CNN, VGG16, and ResNet) realized using these labeled data.
  • After the models were run, reports were made for each model, and their matrices were created. In this report, the shortcomings of each model were determined.
  • In order to eliminate the deficiencies found as a result of the examinations on the models, another CNN model was created, and the errors in the previously examined systems were tried to be eliminated.
  • The accuracy rate obtained from studies conducted with labeled data from the Internet has been further increased.
  • Medical professionals can use a clinical decision support system to detect brain tumors, which has been developed over the model with increased accuracy.
  • Data found on the Internet, previously labeled and made suitable for classification, were used.

2. Materials and Methods

The human brain is one of the most complicated organs in our body, with new features being discovered every day. For this reason, many studies on this organ exist in the literature. These can be classified as medical and engineering studies of technology on brain data. For this reason, screening studies related to brain tumors with a wide literature perspectives were limited to the following topics. The following headings were scanned:
  • Machine and deep learning algorithms;
  • Morphological-based segmentation methods;
  • CNN-based classification.
In the study carried out by Gwak et al., a model using deep feature and machine learning classifiers from ensemble learning models was proposed. The study extracted deep features from brain MR images using the transfer learning method in a deep convolutional neural network. Various machine learning classifiers then evaluated the extracted deep features. The top three deep features that perform well in the machine learning classifier are selected and combined into a deep feature collection. Experimental results emphasize that the performance of the ensemble obtained from deep features can help to significantly improve the success of the model [2].
In the study by Asaf Raza et al., a hybrid deep learning model named DeepTumorNet was developed for three types of brain tumors (CT) (glioma, meningioma, and pituitary tumor classification) over a basic convolutional neural network (CNN) architecture. On the GoogLeNet architecture of the CNN model, the last 5 layers were removed, and 15 new layers were added. A Leaky ReLU activation function has been added to the feature map to increase the significance of the model. The proposed model was tested on a public research dataset for evaluation purposes and achieved 99.67% accuracy, 99.6% precision, 100% recall, and 99.66% F1 score [9].
Traditional approaches to image processing are performed by the process of extracting features from the lower layers. However, this algorithm is not particularly suitable for medical images. In their study, researchers named Lakshmi and Rao aimed to detect brain tumors early using the deep learning approach and hyperparameters by using the Inception-v3 convolutional neural network model. In the study, it is seen that the accuracy of the Inception-v3 algorithm was recorded as 99.34% in the training data and 89% in the validation data [10].
In the study by Chenjie Ge et al., a model using a graph-based semi-supervised learning method was proposed to benefit more from unlabeled data. The proposed model was tested on two glioma datasets, the TCGA dataset for IDH mutation prediction and the MICCAI dataset for glioma grading. It has been reported that an 86.53% test accuracy was obtained in the TCGA dataset and 90.70% in the MICCAI dataset [11].
As can be seen from the literature, many studies have been carried out on brain tumors in recent years. It can be seen that there are many medical imaging techniques for the diagnosis of these tumors [12]. It has been concluded that most of the previous studies were on the segmentation of areas of brain tumors, but recent studies have focused on classifying these areas into different types of brain tumors. Based on this, the study was carried out to increase the success rate of the binary classification of brain tumors.
During the literature research phase of the study, many recent studies were examined. The most striking of these studies are presented in Table 1. These studies draw attention because the issues to be considered in the model to be created are included. The table below discusses the models used in the studies, the success rates, and the results obtained from the study.

3. Methodology

3.1. Overview

Convolutional neural networks are a sub-branch of deep learning and are often used to analyze visual information. CNN or ConvNet is short for convolutional neural network. Convolutional neural networks consist of many layers that can be trained. These are the input, convolution, jointing, and full link layers. The convolution layer and the jointing layer can be fine-tuned with hyperparameters. Different CNN architectures combined with different transfer learning techniques have achieved great success thanks to their improved performance in image classification. In this way, they have surpassed traditional machine-learning models in the last few years. CNN algorithms can automatically recognize much more important visual patterns with less preprocessing on the raw pixels of images. Serving as a test environment, ImageNet has played an important role in advancing deep image recognition architectures. CNN algorithms started to be used rapidly with the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) competition, which was first held in 2012 [22].
Another CNN architecture, VGG16Net, was ranked second in ILSCRC for accuracy in 2014 [18]. The largest created VGG16 architecture consists of 16 layers, 3 of which are fully connected and contain an average of 144 million parameters. The layers contain five pooling layers 2 × 2 in size in each convolution layer. There is also a softmax linear layer at the output. The ReLU activation function is applied in all fully connected layers, and at the same time, the Dropout Layer is used in fully connected layers. Compared to popular methods, it is a CNN model that is considered to have a high computational load due to its large parameters.
DenseNet is a recently proposed custom convolutional neural network model where the current layer connects with all previous layers [23]. The structure has some advantages over existing structures, such as mitigating the disappearing gradient problem, enhancing feature propagation, promoting feature reuse, and reducing the number of parameters. A deep DenseNet is defined as a series of DenseNets (called dense blocks) sequentially connected by additional convolution and pooling operations between consecutive dense blocks. We can create a deep neural network flexible enough to represent complex transformations with such a structure. An example of deep DenseNet is shown in Figure 1.

3.2. Methods and Materials

Some changes were made in the modified CNN model, taking into account the shortcomings of other models. As a result of the examination of the VGG16 Net architectural structure, it was observed that the VGG16 Net convolutional layers positively affected the success rate in areas where health data are not used. For this reason, it was decided to use convolutional layers in the model created. Afterward, the DenseNet architecture was examined. Thanks to the dense layers of this architectural structure, it was determined that a more detailed search was carried out in the MRI images. For this reason, convolutional networks were built with a more intense working principle in the model created. Finally, a decrease in the success rate was detected due to the low number of layers in the analysis of the CNN’s simple architectural structure. For this reason, the number of layers was increased. In this way, searches of the image were made more detailed. The adjustments made in the created model and the increase in the number of layers affected the classification rates and increased the success rate. The Figure 2 explain the components of modified CNN networks.
After the model was created, two different paths were chosen for training and testing. First, the data set was trained in a classical way and separated as a test. Then, the training data were used during deep learning. Afterward, the test data and the success rates of the model were examined. The K-fold cross-validation method was chosen as a second method, and K was determined as 3. The reason for choosing this method was to verify the model’s reliability by creating independent data sets and to prove that the model will show the same success in all kinds of data sets.
During the material selection stage, research was carried out from many sources. As a result of the research, more than 7000 images shared as open source on Kaggle and intended for use in brain tumor image processing studies were obtained. These images consist of 4 different classes in total, classified as three different types of brain tumors and images of healthy individuals. Later, these data were divided to be used in the training and testing phases. The purpose of this was to not use the data used in the training phase in the testing phase. Such a distinction has been made because reclassifying the data the model has seen before during machine learning would affect the accuracy rate. Afterward, studies using the same data set on Kaggle were examined. The reason for these studies was to identify the deficiencies in other studies and to eliminate these deficiencies in our study. Three popular studies were examined. Two of them are models that have been trained by machines before and are expected to have a high success rate. Finally, a simple CNN architecture was examined. This review aims to measure the success of the CNN architecture, which is generally used in image processing in the health field, and then to determine how contributions should be made to the modified CNN architecture.

3.2.1. K-Fold Cross Validation

An extra k-fold cross-validation method was added to test the model’s accuracy and examine the results. This method was added to retrain and test the model with independent data sets other than manually separated data sets.
Cross-validation is a model validation technique that tests how a statistical analysis will yield an independent data set. Its primary use is to predict with what accuracy a prediction system will work in practice. In a prediction problem, the model is usually trained on a set of “known data” (“training set”) and tested against a set of “unknown data” (“validation set” or “test set”). This test aims to measure the trained model’s generalization ability to new data and to detect overfitting or selection bias problems. Simple approach: Set aside 75% for training and 25% for testing. However, while the data are fragmented, some biases and errors may occur in the training and testing of the model, depending on the distribution of the data. Here, k-fold cross-validation divides the data into equal parts according to a determined k number, ensuring that each part is used for both training and testing, thus minimizing deviations and errors caused by dispersion and fragmentation. The K value is usually chosen as 3 or 5. This value can be selected as 10 or 15, but this will cause a costly calculation and time loss.

3.2.2. Dataset

An open-source brain tumor dataset was used to analyze and evaluate our model, which was developed using different CNN architectures, as shown in Figure 3. This dataset was obtained by combining three datasets (figshare, SARTAJ dataset, Br35H) [24]. There are four classes in total in the data set. These are: brain MRI images from glioma, meningioma, oituitary, and healthy individuals. There are 1623 images for glioma, 1627 images for meningioma, 1769 images for pituitary, and 2002 images for healthy individuals. A total of 7021 MRI images were used. The dataset is open-sourced in the Kaggle application. Each file is a 512 × 512 JPEG file with a label indicating the type of brain tumor. This data set was used as input data for each model.
For classifying and defining the brain tumor using the MRI images of the brain tumor dataset, three different models were first processed; then, the model created with these models was compared. Examined models included: standard CNN architecture, VGG16Net architecture, and DenseNet architecture. These models were compared with the later modified model.

4. Results and Discussion and Evaluation of Parameters

The effectiveness of the proposed brain tumor classification and detection system is evaluated by calculating the four main outcome-based evaluation metrics used to test the classifier: true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). The performance of the proposed system is evaluated using the following parameters, as can be seen in Table 2.
Accuracy determines the ability to accurately distinguish brain tumor types. To estimate the accuracy of a test, we calculate the ratio of true positive and true negative for all evaluated cases calculated by the following relationships:
A c c u r a c y = T P + F N T P + T N + F P + F N
On the other hand, recall is a metric that shows how many of the operations we need to estimate as positive are positive.
R e c a l l = T P T P + F N
Precision, on the other hand, shows how many of the values we estimated as positive are actually positive:
P r e c i s i o n = T P T P + F P
The F1 Score value shows us the harmonic mean of Precision and Recall values. The reason why it is a harmonic mean instead of a simple mean is that we should not ignore the extreme cases. If it were a simple average calculation, a model with a Precision value of 1 and a Recall value of 0 would have an F1 Score of 0.5, which would mislead us:
F 1 = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l

4.1. Training Parameters

The training options and hyperparameter networks assigned for training are listed in Table 3. Here, the same training options are used to compare the performance of different architectures. Networks are trained at a learning rate of 1010 for 100 epochs. The data set was divided into two training and testing. A total of 5650 data points were used in the training phase, and 1371 data points were used in the testing phase. The model was trained for 100 epochs. The horizontal axis refers to the number of epochs, and the vertical axis refers to the error rate.

4.2. Training Phase

As a result of the research carried out on Kaggle, three different networks were selected to be examined. These networks are Basic CNN, VGG16Net, and DenseNet. Then, the data percentage to be used in the training phase was determined on the chosen data set: 80% of the data is reserved for use during the training phase and 20% for the testing phase. During the training phase, the accuracy rates of architectural structures, validation accuracy rates, and changes in loss functions were examined and tabulated at each step. These changes were examined separately for each model. It was learned that the same errors occurred in certain classes during the classification in the previously trained models. In the simple CNN network, classification errors arose due to deficiencies in the layers. These errors were reduced to minimum levels in the modified CNN network, and high accuracy rates were obtained during the training phase.

4.2.1. Basic CNN Architecture

During the literature research phase of the study, the models in other studies were examined. As a result of this research, it was concluded that the popularities of the classical CNN model, VGG16Net, and DenseNet models in the studies were high, so it was decided to examine studies of these models. In the classical CNN architecture, firstly, the structure of the layers was examined. There is only one convolutional layer and a pooling layer in this structure. Then, classification is performed by passing information to the classical neural network. In the proposed approach, convolutional and pooling layers have been developed and inserted into the classical neural network after multiple preprocessing steps.
As a result of the research carried out on Kaggle, it was decided that the first structure to be examined would be a simple CNN architecture. The purpose of choosing this architectural structure is to observe how successful the CNN architecture will be without changing the layers. As a result of these observations, it was decided how changes should be made in the layers. The system was operated with a total of 10 epochs, and the resulting values were examined. These values are shown in Table 4.
As can be seen from the graphics, the accuracy of the system is up to 92%. However, to test the reliability of the system, an application was made using the test data set. The reason for this is that even if the success rates of the created models are high during the training phase, their inadequacy in the testing phase has been observed before. As a result of this application, and as can be seen in Figure 4, an error matrix was created, and Precision, Sensitivity, and F1 Scores were examined over this error matrix.
As can be seen from the error matrix in Figure 4, the simple CNN model works well for two classes (Healthy Individual and Pituitary). However, the success rate in other classes was less than expected. Even if the success rate in other classes is sufficient as a result of a normal examination, a success rate of 89–90% is not sufficient in a study on cancer in the field of health. For this reason, this model is not a model that can achieve the necessary success in real life. The weak points of this model were examined, and it was determined which parts were missing. Even though the success rate in MRI images of healthy individuals and patients with the pituitary disease was at the expected rate, the success rates in MRI images in the glioma and meninglioma tumor classes were below the expected levels.
In order to examine the reliability of the model in detail, the values of each tumor class at the test stage were examined separately. In this way, it has become easier to detect the model’s weak points. As we have previously examined in the error matrix, although the success rate in healthy individuals and patients with a pituitary brain tumor was 97–98%, the success rate in meninglioma and glioma tumor types remained below 90%, as shown in Table 5. This ratio is not sufficient for application in the field of health. The reason for this low success rate is the equation we use to calculate the F1 score. While calculating the F1 score, the value that affects the ratio the most is the “False Positive” value. This value occurs when people’s test results are negative, but they have a disease, and it is the most important value in medical applications. This is the weakest point of this system since this value is too high in the case of meninglioma.

4.2.2. VGG16Net Architecture

The VGG16Net architectural structure uses convolutional layers in groups of two or three, and it is different from CNN model structures. At the same time, the VGG16Net architecture functions as a previously trained model. The reason for making a comparison between VGG16Net and the proposed model is that the complexity in the former’s network structure is higher than the proposed model. We also want to see how much a pre-trained model will affect the success rate.
As the second model, it was decided to examine the VGG16Net architectural structure. A new model was created using the VGG16 transfer learning method. The purpose of this model is to increase its reliability by applying training with large datasets. This model was run at the same epoch number with the simple CNN architecture, and its values during the training phase are given in Table 6.
As seen in Table 6, the accuracy rate of the model regularly increases throughout the training phase. However, during the validation stage, that is, the pre-test phase, the accuracy values could not meet the expected values and did not show a regular increase. The validation accuracy decreased again in some epochs. Afterward, the trained model was tested, and an error matrix was created. A table was created to examine this matrix in detail and to see the “Sensitivity,” “Precision,” and “F1 Score” values. As seen on the confusion matrix, the VGG 16 Model worked with a lower success rate than the simple CNN model, and the success rate decreased in the same classes. The model that made many mistakes in the glioma brain tumor class was determined as a model that is unlikely to be used in the health field as a success rate.
Later, when we examined the performance of Table 7 according to their classes, it made too many mistakes and examined the deficiencies in meninglioma and glioma tumor types similar to the previous model. It did not achieve the expected success in distinguishing tumor classes at the classification stage, as shown in Figure 5.

4.2.3. DenseNet Architectural Structure

The reason for examining the DenseNet architecture is that in DenseNet, each layer receives inputs from the previous layers and transfers all feature maps to the next layers. The aim here is to compare the CNN architectural structure and the DenseNet architectural structure, which transfers collective information. This way, the operation of feature maps will be compared, and the differences between the models will be examined.
Finally, a different transfer learning model was examined based on the VGG16 transfer learning model. The reason for choosing the DenseNet architecture is that it can better learn the details in the inputs and outputs by using dense layers. The DenseNet model is also a model that has been previously trained with different data sets. It was run on the same data set, with the same number of epochs as the other models. The values at the training stage are given in Table 8.
Like the VGG16 model, the DenseNet model showed a regular increase in accuracy during the training phase. However, also like the VGG16 model, there is a tide in the accuracy rates in the pre-test phase, causing a loss of confidence. After some epochs, there were decreases. In order to examine the success rates in the test phase, a table showing the “Sensitivity”, “Precision,” and “F1 Score” values was created.
As can be seen in Table 9, unlike the VGG16 architecture, DenseNet’s architecture failed to detect meninglioma and glioma tumors, although the Sensitivity ratio was high in healthy individuals and pituitary tumor patients. Likewise, the apparent insufficiency in F1 scores is due to the high number of “False Positives.”

4.2.4. Modified CNN Architecture

As a result of the examination of the three models, deficiencies in each model were determined. Afterward, a CNN model was created in order to eliminate these deficiencies. In this model, unlike the simple CNN model, a more advanced network is created, and images are analyzed in more detail in this network. The values of the model operated in the form of 20 epochs during the training phase are given in Table 10.
As can be seen in Table 10, a 99% success rate was achieved in the training phase, unlike other models. In the pre-test phase, this rate increased to 95%. In order to test this model and examine in which parts it makes mistakes, an error matrix, followed by a “Sensitivity”, “Precision”, and “F1 Score” table, was created.
As a result of the examinations on the error matrix, all MR images were correctly classified, except for 23 images in total. A “False Positive”, which is the most important for us, was only made in three images. In the developed model, many of the deficiencies of the other models were eliminated, and the F1 score was examined to check the model’s reliability, as shown in Figure 6. In Table 11, additional evaluation metrics are provided for each class. These metrics convey useful information regarding the modified model’ predictive power for each class.
In order to prove this increase and to ensure the reliability of the model, a K-fold cross-validation method was also added, and the results of this method were examined. As a result of the cross-validation process, it has been proven that the model’s work with independent data sets does not reduce its success rate, and it works better than other models.
In the K-fold cross-validation, K was determined as 10. In the first stage, K was determined as five, and the results were examined. After the K-fold application, an increase in the success rate was observed. The reason why five-fold cross-validation was chosen first is to determine how it would affect the operation of the system. Later, due to the increase in the success rate, K was determined as ten and run again. The reason for this was to observe whether there would be an increase in the success rate as a result of 10-fold cross-validation. However, success rates continued to rise between 97 and 95. It was decided to use Google Colab during the operations. This is because the model that is running on Jupyter has been shown to have a very long processing time. The fast transaction service provided via Google Colab was utilized. The comparative results for the brain tumor databases are depicted in Table 12.
The developed CNN architectural structure showed a higher success than other architectural structures. The classical CNN architectural structure could not reach the desired success rate due to the low number of layers. In the examination of the model, it was observed that there were some deficiencies in feature extraction due to the number of layers, and for this reason, the success rate decreased in some classes.
In the VGG16Net architectural structure, it has been observed how to measure the success rate of pre-trained models and how increasing the number of convolutional layers will affect the success rate. It was concluded that the previously trained models had a low overshoot rate on health data. For this reason, the margin of error is high in the classification stage of such architectural structures.
Finally, the reason for examining the DenseNet architectural structure is to observe how the use of denser layers in the transfer learning method and the collective knowledge transfer will yield results. Although dense layers are more effective than the VGG16Net architectural structure, the error rate was high in some classes due to the transfer learning method.
The conclusion that is understood as a result of these examinations is that the layers should be used intensively and that the training phase should be carried out without using transfer learning methods. For this reason, an intensive CNN model was created in the proposed method, in which 80% of the data was used in the training phase and the model was self-trained without using transfer learning methods.
Table 13 was created to compare and examine the average values of precision, sensitivity, and F1 Scores of all models
Table 14 gives average evaluation metrics (F1 Score, Precision, and Recall) for a comparison of the results having the same dataset as the state-of-the-art studies.
When the table is examined, it can be observed that our research is in close proximity with the studies that reach high accuracy among modern studies. There is also some consistency between all three scales. Detailed interventions that can be carried out in the data preparation and cleaning stages can raise the values a little higher.

5. Conclusions

In our study, researchers determined weak points by examining many image processing methods. These studies were conducted to examine the missing parts of previous studies and eliminate these deficiencies. First, the simple structure of the CNN architecture was examined, and some tests were carried out using our study’s dataset. As a result of these tests, it has been determined that there are some deficiencies in classification due to the small number of layers. Afterward, the VGG16Net and DenseNet architectural structures were examined. These models were examined to determine how the transfer learning method would affect their success rates and to observe how the use of dense layers will contribute to them. It has been observed that the transfer learning method does not affect the health field’s success rate, even if the success rate is high in other areas. For this reason, it was decided not to use the transfer learning method in the model created in this study. Afterward, DenseNet analysis was carried out. Although it affected the success rate positively in dense layers, the success rate did not increase to the expected level due to transfer learning. For this reason, it was decided that a structure with dense layers should be created, and the training phase should be completed in person. A CNN architecture was used in the created model, and the density of the layers in the preprocessing stage was kept at a high rate. In this way, the rate of obtaining information from the data has increased. Since it was decided not to use the transfer learning method, it was concluded that more data were needed. For this reason, as a result of the research, a dataset published as open source on Kaggle was found. In this dataset, there are four classes, namely, Glioma, Meninglioma, Pituitary, and No Tumor, and these classes contain more than 7000 images in total. The dataset is divided into 80% training and 20% testing phases. The expected success rate was achieved due to the large amount of data in the training phase and the fact that the proposed model was created with dense layers. This way, a successful training phase was achieved thanks to the dense layers without using transfer learning methods, and a success rate of 94–97% was achieved. In future studies, it is aimed to improve the weak points of our model. The most important weakness of the proposed model is the long processing time. The reason for the late response is that the model was created using both dense and convolutional networks. Afterward, we would like to add segmentation areas with tumors. This way, after the tumor classification is complete, it will be possible to calculate in which part of the brain the tumor is located and its size. If these goals are completed, a system will be created that will help people working in the field of health to lighten their burdens.

Author Contributions

Conceptualization, O.İ.B. and O.Ö.; methodology, O.İ.B., O.Ö., H.G. and F.K.; software, O.İ.B.; validation, F.K., J.K. and J.H.; formal analysis, O.İ.B., O.Ö. and H.G.; investigation, O.İ.B., O.Ö. and H.G.; resources, O.İ.B.; data curation, O.İ.B. and O.Ö.; writing—original draft preparation, O.İ.B., O.Ö., H.G. and F.K.; writing—review and editing, O.İ.B., O.Ö., H.G., F.K., J.K., U.e.L. and J.H.; visualization, O.İ.B.; supervision, F.K.; funding acquisition, F.K.and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Louis, D.N.; Perry, A.; Reifenberger, G.; Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization Classification of Tumors of the Central Nervous System: A summary. Acta Neuropathol. 2016, 131, 813–820. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Kang, J.; Ullah, Z.; Gwak, J. MRI-Based Brain Tumor Classification Using Ensemble of Deep Features and Machine Learning Classifiers. Sensors 2021, 21, 2222. [Google Scholar] [CrossRef] [PubMed]
  3. Gao, H.; Jiang, X. Progress on the diagnosis and evaluation of brain tumors. Cancer Imaging 2013, 13, 466–481. [Google Scholar] [CrossRef] [Green Version]
  4. Alzubaidi, L.; Santamaría, J.; Manoufali, M.; Mohammed, B.; Fadhel, M.A.; Zhang, J.; Al-Timemy, A.H.; Al-Shamma, O.; Duan, Y. MedNet: Pre-trained Convolutional Neural Network Model for the Medical Imaging Tasks. arXiv 2021, arXiv:2110.06512. [Google Scholar]
  5. Hussain, S.M.; Wahid, A.; Shah, M.A.; Akhunzada, A.; Khan, F.; Amin, N.U.; Arshad, S.; Ali, I. Seven pillars to achieve energy efficiency in high-performance computing data centers. Recent Trends Adv. Wirel. IoT-Enabled Netw. 2019, 93–105. [Google Scholar]
  6. Farooqi, M.M.; Shah, M.A.; Wahid, A.; Akhunzada, A.; Khan, F.; ul Amin, N.; Ali, I. Big data in healthcare: A survey. Applications of intelligent technologies in healthcare. Appl. Intell. Technol. Healthc. 2019, 143–152. [Google Scholar]
  7. Khan, F.; Khan, A.W.; Shah, K.; Qasim, I.; Habib, A. An algorithmic approach for core election in mobile ad-hoc network. Journal of Internet Technology. J. Internet Technol. 2019, 20, 1099–1111. [Google Scholar]
  8. Khan, F.; Tarimer, I.; Alwageed, H.S.; Karadağ, B.C.; Fayaz, M.; Abdusalomov, A.B.; Cho, Y.I. Effect of Feature Selection on the Accuracy of Music Popularity Classification Using Machine Learning Algorithms. Electronics 2022, 11, 3518. [Google Scholar] [CrossRef]
  9. Raza, A.; Ayub, H.; Khan, J.A.; Ahmad, I.; Salama, A.S.; Daradkeh, Y.I.; Javeed, D.; Rehman, A.U.; Hamam, H. A Hybrid Deep Learning-Based Approach for Brain Tumor Classification. Electronics 2022, 11, 1146. [Google Scholar] [CrossRef]
  10. Lakshmi, M.J.; Rao, S.N. Brain tumor magnetic resonance image classification: A deep learning approach. Soft Comput. 2022, 26, 6245–6253. [Google Scholar] [CrossRef]
  11. Ge, C.; Gu, I.Y.-H.; Jakola, A.S.; Ya, J. Enlarged Training Dataset by Pairwise GANs for Molecular-Based Brain Tumor Classification. IEEE Access 2020, 8, 22560–22570. [Google Scholar] [CrossRef]
  12. Işın, A.; Direkoğlu, C.; Şah, M. Review of MRI-based Brain Tumor Image Segmentation Using Deep Learning Methods. Procedia Comput. Sci. 2016, 102, 317–324. [Google Scholar] [CrossRef] [Green Version]
  13. Kurup, R.V.; Vishvanathan, S.; Kp, S. Effect of Data Pre-processing on Brain Tumor Classification Using Capsulenet. In Proceedings of the International Conference on Intelligent Computing and Communication Technologies, Hyderabad, India, 9–11 January 2019; pp. 110–119. [Google Scholar]
  14. Siar, M.; Teshnehlab, M. Brain Tumor Detection Using Deep Neural Network and Machine Learning Algorithm. In Proceedings of the 9th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 24–25 October 2019. [Google Scholar]
  15. Maharjan, S.; Alsadoon, A.; Prasad, P.W.C.; Al-Dalain, T.; Alsadoon, O.H. A novel enhanced softmax loss function for brain tumour detection using deep learning. J Neurosci Methods 2020, 330, 108520. [Google Scholar] [CrossRef] [PubMed]
  16. Arı, A.; Hanbay, D. Deep learning-based brain tumor classification and detection system. Turk. J. Electr. Eng. Comput. Sci. 2018, 26, 5. [Google Scholar]
  17. Boustani, A.E.; Aatila, M.; Bachari, E.E.; Oirrak, A.E. MRI Brain Images Classification Using Convolutional Neural Networks. In Proceedings of the Advanced Intelligent Systems for Sustainable Development (AI2SD’2019), Marrakech, Morocco, 8–11 July 2019; pp. 308–320. [Google Scholar]
  18. Arif, M.; Ajesh, F.; Shamsudheen, S.; Geman, O.; Izdrui, D.; Vicoveanu, D. Brain Tumor Detection and Classification by MRI Using Biologically Inspired Orthogonal Wavelet Transform and Deep Learning Techniques. J. Healthc. Eng. 2022, 2022, 2693621. [Google Scholar] [CrossRef]
  19. Budati, A.K.; Katta, R.B. An automated brain tumor detection and classification from MRI images using machine learning techniques with IoT. Environ. Dev. Sustain. 2022, 24, 10570–10584. [Google Scholar] [CrossRef]
  20. Dehkordi, A.A.; Hashemi, M.; Neshat, M.; Mirjalili, S.; Sadiq, A.S. Brain Tumor Detection and Classification Using a New Evolutionary Convolutional Neural Network. arXiv 2022, arXiv:2204.12297. [Google Scholar] [CrossRef]
  21. Biratu, E.S.; Schwenker, F.; Debelee, T.G.; Kebede, S.R.; Negera, W.G.; Molla, H.T. Enhanced Region Growing for Brain Tumor MR Image Segmentation. J. Imaging 2021, 7, 22. [Google Scholar] [CrossRef]
  22. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  23. Abirami, M.; Vennila, B.; Suganthi, K.; Kawatra, S.; Vaishnava, A. Detection of Choroidal Neovascularization (CNV) in Retina OCT Images Using VGG16 and DenseNet CNN. Wirel. Pers. Commun. 2022, 127, 2569–2583. [Google Scholar] [CrossRef]
  24. Nickparvar, M. Brain Tumor MRI Dataset. 2021. Available online: https://doi.org/10.34740/KAGGLE/DSV/2645886 (accessed on 24 April 2022).
  25. Deepak, S.; Ameer, P.M. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef] [PubMed]
  26. Alqudah, A.M.; Alquraan, H.; Qasmieh, I.A.; Alqudah, A.; Al-Sharu, W. Brain tumor classification using deep learning technique--a comparison between cropped, uncropped, and segmented lesion images with different sizes. arXiv 2020, arXiv:2001.08844. [Google Scholar]
  27. Saleh, A.; Sukaik, R.; Abu-Naser, S.S. Brain tumor classification using deep learning. In Proceedings of the 2020 International Conference on Assistive and Rehabilitation Technologies (iCareTech), Gaza, Palestinec, 28–29 August 2020; pp. 131–136. [Google Scholar]
  28. Ghassemi, N.; Shoeibi, A.; Rouhani, M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed. Signal Process. Control 2020, 57, 101678. [Google Scholar] [CrossRef]
Figure 1. Three-Block DenseNet Architecture.
Figure 1. Three-Block DenseNet Architecture.
Life 13 00349 g001
Figure 2. Modified CNN Architecture.
Figure 2. Modified CNN Architecture.
Life 13 00349 g002
Figure 3. Classes of the Dataset.
Figure 3. Classes of the Dataset.
Life 13 00349 g003
Figure 4. Basic CNN Confusion Matrix.
Figure 4. Basic CNN Confusion Matrix.
Life 13 00349 g004
Figure 5. VGG16 Architectural Structure Confusion Matrix.
Figure 5. VGG16 Architectural Structure Confusion Matrix.
Life 13 00349 g005
Figure 6. Modified Model Confusion Matrix.
Figure 6. Modified Model Confusion Matrix.
Life 13 00349 g006
Table 1. Comparative Analysis.
Table 1. Comparative Analysis.
Ref.YearMethodologyDatasetResultDrawback
[13]2020PNN
Classification CNN
KaggleTCIA%90 AccuracyLack of comparative analysis
[14]2019Inception Pre-Trained CNNBRATS 13,14,17,18%99.12
Accuracy
Complex Approach
[15]2019CNN with a Modified Softmax Loss FunctionBRATS, ISLES, FLAIR, DWI%98.9 AccuracyIt can be developed and can be tested with large datasets.
[16]2018ELM-LRFMNI Brain Website%97 Classification SuccessMore efficient than related work because training time is shorter than others
[17]2029CNNTCIA, Kaggle%90
Accuracy
Long Process Time
[18]2022CNN Harvard Medical School website%98.5 AccuracyComputational time, system complexity, and memory space requirements are too much
[19]2022SVMBRATS 2017%92.3 AccuracySVM cannot handle a larger data set.
[20]2022Evolutionary CNNBRATS 2015 data set and brain image datasets from Harvard Medical School,%97.4 AccuracyComplex System
[21]2021HOG + LBP + deep features BRATS2015%96 DSCThe problem in thresholding point
Table 2. Confusion matrix definition.
Table 2. Confusion matrix definition.
ClassStatementFormula
True PositiveImages that belong to a patient who is sick and correctly known by the modelTP
True NegativeImages that belong to a patient who is healthy and correctly known by the modelTN
False PositiveImages that belong to a patient who is healthy but diagnosed as sick by the modelFP
False NegativeImages that belong to a patient who is sick but diagnosed as healthy by the modelFN
Table 3. Hyperparameters.
Table 3. Hyperparameters.
ParametersValue
Momentum0.9
Initial learning rate10−4
Maximum number of epochs100
Validation Split0.1
Table 4. Simple CNN Architecture Training Stage Accuracy Values.
Table 4. Simple CNN Architecture Training Stage Accuracy Values.
Number of EpochsLoss FunctionAccuracy RateValidation Loss FunctionValidation Accuracy Rate
11.20980.64351.82100.6485
20.59770.75820.55310.7680
30.48060.79590.65290.7680
40.48260.81540.47410.8629
50.38040.85020.58920.7856
60.30720.86850.49320.8084
70.22910.89690.33900.9086
80.21020.90990.33100.9174
90.19530.91660.35050.9033
100.17320.92320.32900.9192
Table 5. Performance of the CNN model by class.
Table 5. Performance of the CNN model by class.
Brain Tumor ClassesPrecisionRecallF1 Score
00.890.900.89
10.970.980.97
20.880.850. 86
30.950970.96
Table 6. VGG16 Architecture Training Phase Accuracy Values.
Table 6. VGG16 Architecture Training Phase Accuracy Values.
Number of EpochsLoss FunctionAccuracyValidation Loss FunctionValidation Accuracy Rate
18.30110.57014.13410.7255
22.83750.77312.80200.7824
31.91710.82862.12380.8023
41.45980.85271.74710.8222
51.13120.86981.69810.8293
60.95220.87851.60330.8222
70.78340.89051.25390.8478
80.67010.90131.17610.8592
90.56150.90510.97380.8720
100.49400.91030.96410.8663
Table 7. VGG16 Model Performance Table by Classes.
Table 7. VGG16 Model Performance Table by Classes.
Brain Tumor ClassesPrecisionRecallF1 Score
00.800.890.84
10.930.960.94
20.850.670.75
30.830.940.90
Table 8. DenseNet Architectural Structure Education Phase Values.
Table 8. DenseNet Architectural Structure Education Phase Values.
Number of EpochsLoss FunctionAccuracy RateValidation Loss FunctionValidation Accuracy Rate
10.57600.77870.32210.8766
20.40700.84270.38870.8574
30.34600.86340.41530.8574
40.30950.88550.33640.8775
50.30300.88790.33550.8836
60.28940.88280.37570.8688
70.27830.89930.31160.8889
80.27470.89580.36820.8329
90.25070.90680.25590.9064
100.25040.90300.37870.8608
Table 9. DenseNet Architectural Structure Performance Values by Class.
Table 9. DenseNet Architectural Structure Performance Values by Class.
Brain Tumor ClassesPrecisionRecallF1 Score
00.750.990.85
10.910.990.95
20.930.830.88
30.920.580.71
Table 10. Modified Model Training Stage Values.
Table 10. Modified Model Training Stage Values.
Number of EpochsLoss FunctionAccuracy RateValidation Loss FunctionValidation Accuracy Rate
10.76990.72711.11230.6538
20.42670.84290.65230.7522
30.32690.87730.76340.7135
40.21420.92280.31450.8981
50.18670.93104.23100.6520
60.14220.94960.55620.8418
70.07500.97420.23940.9262
80.05990.97970.54380.8453
90.04930.98500.19740.9455
100.04720.98550.19360.9438
110.04040.98770.20020.9438
120.03830.98770.19820.9525
130.03170.99220.20540.9438
140.03420.98870.20390.9455
150.03420.99140.20400.9473
160.03360.99100.20560.9438
170.03120.99100.20500.9438
180.03230.99100.20450.9455
190.03900.98950.20480.9455
200.03040.99140.20460.9455
Table 11. Performance Values by Modified Model Classes.
Table 11. Performance Values by Modified Model Classes.
Brain Tumor ClassesPrecisionRecallF1 Score
00.960.970.97
10.970.980.98
20.960.910.94
30.950.990.97
Table 12. Ten-fold cross-validation performance rates (%).
Table 12. Ten-fold cross-validation performance rates (%).
Fold Number
N
Training AccuracyTest Accuracy
10.960.94
20.950.93
30.960.94
40.950.9256
50.970.94
60.960.93
70.950.94
80.970.94
90.970.95
100.960.94
Table 13. Comparison Table of Models.
Table 13. Comparison Table of Models.
Model NameNumber of EpochsAvg. PrecisionAvg. RecallAvg. F1 Score
Basic CNN100.92250.90.92
VGG16 Net100.86250.86250.8575
DenseNet100.87750.87750.8475
Modified CNN100.960.960.9650
Table 14. Comparison Table of Models with other papers.
Table 14. Comparison Table of Models with other papers.
Raza et al. [9]Deepak & Ameer [25] Alqudah et al. [26]Saleh et al. [27]Ghassemi et al. [28]Our study
F1 Score99.6695.3498.9199.7595.0996.00
Precision99.6094.7098.98-95.2896.00
Recall10096.0098.85-94.9196.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Özkaraca, O.; Bağrıaçık, O.İ.; Gürüler, H.; Khan, F.; Hussain, J.; Khan, J.; Laila, U.e. Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images. Life 2023, 13, 349. https://doi.org/10.3390/life13020349

AMA Style

Özkaraca O, Bağrıaçık Oİ, Gürüler H, Khan F, Hussain J, Khan J, Laila Ue. Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images. Life. 2023; 13(2):349. https://doi.org/10.3390/life13020349

Chicago/Turabian Style

Özkaraca, Osman, Okan İhsan Bağrıaçık, Hüseyin Gürüler, Faheem Khan, Jamil Hussain, Jawad Khan, and Umm e Laila. 2023. "Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images" Life 13, no. 2: 349. https://doi.org/10.3390/life13020349

APA Style

Özkaraca, O., Bağrıaçık, O. İ., Gürüler, H., Khan, F., Hussain, J., Khan, J., & Laila, U. e. (2023). Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images. Life, 13(2), 349. https://doi.org/10.3390/life13020349

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop