Next Article in Journal
Production of a Monoclonal Antibody for the Detection of Forchlorfenuron: Application in an Indirect Enzyme-Linked Immunosorbent Assay and Immunochromatographic Strip
Next Article in Special Issue
Brain Tumor Segmentation and Classification from Sensor-Based Portable Microwave Brain Imaging System Using Lightweight Deep Learning Models
Previous Article in Journal
Nonenzymatic Electrochemical Glutamate Sensor Using Copper Oxide Nanomaterials and Multiwall Carbon Nanotubes
Previous Article in Special Issue
Utilization of Personalized Machine-Learning to Screen for Dysglycemia from Ambulatory ECG, toward Noninvasive Blood Glucose Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Deep Learning Based Microwave Brain Image Network Model for Brain Tumor Classification Using Reconstructed Microwave Brain (RMB) Images

by
Amran Hossain
1,2,*,
Mohammad Tariqul Islam
1,*,
Sharul Kamal Abdul Rahim
3,
Md Atiqur Rahman
1,
Tawsifur Rahman
4,
Haslina Arshad
5,
Amit Khandakar
4,
Mohamed Arslane Ayari
6 and
Muhammad E. H. Chowdhury
4,*
1
Centre for Advanced Electronic and Communication Engineering, Department of Electrical, Electronic and Systems Engineering, Faculty of Engineering and Built Environment, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia
2
Department of Computer Science and Engineering, Dhaka University of Engineering and Technology, Gazipur, Gazipur 1707, Bangladesh
3
Wireless Communication Centre, University Teknologi Malaysia, Skudai 81310, Malaysia
4
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
5
Institute of IR4.0, Universiti Kebangsaan Malaysia (UKM), Bangi 43600, Malaysia
6
Department of Civil and Architectural Engineering, Qatar University, Doha 2713, Qatar
*
Authors to whom correspondence should be addressed.
Biosensors 2023, 13(2), 238; https://doi.org/10.3390/bios13020238
Submission received: 21 November 2022 / Revised: 23 January 2023 / Accepted: 30 January 2023 / Published: 7 February 2023
(This article belongs to the Special Issue Biomedical Signal Processing in Healthcare and Disease Diagnosis)

Abstract

:
Computerized brain tumor classification from the reconstructed microwave brain (RMB) images is important for the examination and observation of the development of brain disease. In this paper, an eight-layered lightweight classifier model called microwave brain image network (MBINet) using a self-organized operational neural network (Self-ONN) is proposed to classify the reconstructed microwave brain (RMB) images into six classes. Initially, an experimental antenna sensor-based microwave brain imaging (SMBI) system was implemented, and RMB images were collected to create an image dataset. It consists of a total of 1320 images: 300 images for the non-tumor, 215 images for each single malignant and benign tumor, 200 images for each double benign tumor and double malignant tumor, and 190 images for the single benign and single malignant tumor classes. Then, image resizing and normalization techniques were used for image preprocessing. Thereafter, augmentation techniques were applied to the dataset to make 13,200 training images per fold for 5-fold cross-validation. The MBINet model was trained and achieved accuracy, precision, recall, F1-score, and specificity of 96.97%, 96.93%, 96.85%, 96.83%, and 97.95%, respectively, for six-class classification using original RMB images. The MBINet model was compared with four Self-ONNs, two vanilla CNNs, ResNet50, ResNet101, and DenseNet201 pre-trained models, and showed better classification outcomes (almost 98%). Therefore, the MBINet model can be used for reliably classifying the tumor(s) using RMB images in the SMBI system.

1. Introduction

At present, brain tumors are a serious cause of death worldwide. The expansion of abnormal cells that develop inside the brain results in a brain tumor. It causes harm to the brain’s major tissues and develops into cancer. It poses a threat to human life, has a deadly prognosis, and has a significant impact on quality of life. However, due to the unrestrained progression of brain tumors, the possibility of developing brain cancer is increasing day by day. The probability of developing brain cancer is increasing over time as a result of the tumors’ unrestrained growth, and it is the tenth leading cause of death in people [1]. Typically, there are two main categories for brain tumors: (i) benign tumors and (ii) malignant tumors [2,3]. Globally, the incidence of brain tumors is rising alarmingly rapidly. The National Brain Tumor Society (NBTS) estimates that 88,970 Americans were living with a major brain cancer diagnosis in 2022, with 63,040 of those individuals suffering from benign tumors and 25,930 from cancerous tumors [2]. According to the report, the survival rate of the patients is only 36%. The benign tumor is made up of non-cancerous cells with a uniform structure and a consistent shape [4,5,6]. It does not expand to other bodily areas or encroach on nearby tissue. A cancerous tissue of the brain with a heterogeneous composition and an irregular form makes up a malignant tumor [4,5,6]. In contrast to malignant tumors, which grow uncontrollably, benign tumors grow relatively slowly. The invasive properties of the tumors increase the death rate, but early diagnosis, monitoring, categorization, and appropriate examination can lower the mortality rate and raise the survival percentage. Furthermore, automatic brain tumor classification from medical images is important for clinical evaluation and treatment planning of brain cancers. In addition, brain tumor analysis, classification, and detection are severe issues for radiologists and medical doctors. The accurate and timely investigation of brain cancer is imperious for the appropriate treatment of this disease. Brain tumor classification is an important technique in medical imaging applications that classifies specific tumors based on head images. Currently, different types of imaging technologies, including PET (positron emission tomography), MRI (magnetic resonance imaging), ultrasound screening, X-ray screening, and CT (computed tomography), are utilized to diagnose brain tumors in advanced medical facilities [4,7,8,9]. These imaging standards assist medical doctors and radiologists to identify different types of health-related diseases, such as brain cancer. These imaging methods’ major downsides include increased cancer risk due to high dose radioactivity, decreased susceptibility, extreme ionizing of brain cells, high cost, and risk for pregnant women and elderly patients [5,6,7,9,10,11,12]. Further, microwave imaging (MWI) has recently attracted a lot of attention from researchers for medical applications because of its remarkable characteristics, such as non-ionizing radioactivity, penetration capability with low power, non-invasiveness, ionization risk-free for the human body, and cost-effectiveness with a low profile [10,13,14,15,16]. Nowadays, researchers have been using microwave imaging technology to overcome the drawbacks of traditional medical imaging modalities [16,17,18,19,20,21]. Antennas play an important role in microwave brain imaging (MBI) technology, where a single antenna acts as a transmitter and others as receivers. Receivers receive the backscattered biosignals, which are then post-processed by the image reconstruction process. The data is then post-processed using the image generation procedure to produce reconstructed images. Different image reconstruction algorithms have been used in microwave head imaging modalities to detect brain tumors [15,18,20,21,22,23,24,25,26]. However, the main limitations of the developed MBI modalities are (i) noisy, blurry, and low-resolution images created by the system; (ii) identification of the tumor and its location is complicated for a non-expert physician and radiologist; and (iii) difficulty in classifying the tumor in RMB images. In order to overcome such limitations, researchers have been applying deep learning techniques to microwave imaging systems [27,28,29,30,31,32].
Deep learning (DL) is a subdomain of machine learning that uses the convolutional neural network (CNN) model to classify images. The CNN has convolutional layers for feature extraction and densely connected layer(s) for classification. Recently, image classification has become an essential role of medical image analysis, in which deep convolutional neural networks (DCNNs) have been used for the last ten decades. The image classification identifies whether the target object or disease is present or not in the image of the investigation. A fine-tuned DenseNet201 (FT-DenseNet201) deep learning model was developed to classify the MRI tumor images [33]. The model achieved 95% accuracy, but it shows less accuracy when classifying the small-sized tumor images. In another study, a pretrained DenseNet201 (PT-DenseNet201) model was proposed to classify the tumors [34]. It is based on multilevel features and concatenation characteristics that can diagnose the tumor at an early stage. The approach achieved 99.34% testing accuracy, but the precision and specificity scores were only 92% and 83%, respectively. The dual pathway DenseNet (DP-DenseNet) architecture model was proposed in [35] to classify tumors. The architecture was evaluated on the BRATS 2017 MRI dataset. The reported precision, F1 score, and dice score were 85%, 88%, and 89%, respectively. The network model can only classify large-sized tumor-based images but not small-sized tumor images, resulting in comparatively poor classification performances. A deep neural network with a generative adversarial network (DGAN) model was proposed in [36] to classify brain tumors based on MR images. The DGAN model used 64 × 64 sized images as input, and it achieved 93% accuracy. It showed low classification performance due to input image size limitations. A pre-trained Inception-v3 classification model was used to classify brain tumors. The approach employed the concatenation method with the Softmax classification technique to classify tumors from the MRI images. The differential deep convolutional neural network (differential-DCNN) model was presented in [37]. The model was evaluated using 17,600 MRI brain images for the classification of the different types of tumors. As a classification performance, the model achieved 95% sensitivity and 93% specificity. In [38], a conventional multi-pathway CNN (CMP-CNN) architecture was presented for tumor classification by using MRI brain images. The model assessed 3064 images for three types of tumor classification and attained 94% sensitivity. A multi-class tumor image classification by ResNet-50 was proposed in [39]. The model used a global average pooling mechanism to enhance the classification accuracy, but it achieved 97.08% mean accuracy and a 90.02% F1 score. A fine-tuned ResNet101 model was presented in [40] for brain tumor classification. The model used a differential evaluation method and a swarm optimization algorithm to improve classification performance. The classification accuracy achieved by this model was 94.4%. Nevertheless, the model may fail to classify noisy, blurry, and small sized images. However, the pros and cons of existing models are summarized in the following Table 1.
In the last couple of years, deep learning-based transformer architecture has been employed for classifying medical images instead of CNN models. An nn-TransUNet model has been used for MRI medical image segmentation tasks [41]. The model used vision transformers and convolution layers in the encoder for enhancing the segmentation and classification performance. The main benefit of the model is that it reduces time complexity and makes it possible to manually tune the hyperparameters for improved training accuracy, but it requires a large amount of memory and a high-performance GPU to train the model. Another transformer architecture-based model, Vision Transformer (ViT), has been used for classifying images and image recognition [42]. The ViT model used linear projections of flattened patches for image classification. The encoder of the transformer model uses multiheaded self-attention and residual connections in every block for increasing classification performance. A residual vision transformer (ResViT) architecture model has been applied for multimodal medical image classification [43]. In [44], the authors have used a zero-shot learned adversarial transformer (SLATER). The SLATER model combines a deep adversarial network with cross-attention transformers to reduce noise in medical MRI images and enhance image classification performances. It is a suitable network architecture for high-performance MRI image acceleration, but it takes a long time to train the model.
Recently, operational neural networks (ONNs) have been applied as a diverse network model for image analyzing, classification, and processing due to their non-linear properties, low computational complexity, simplicity in structure, and high performances. A self-organized ONN (Self-ONN) model was proposed in [45,46] to classify the biomedical images. It is seen that the Self-ONN model can perform better than conventional CNN models if the model architecture and parameters can be tweaked carefully.
Since all the above-mentioned works used deeper architectures, it is natural that these networks require longer training and inference times and are not suitable for portable device deployment. Therefore, there is a demand to design a lightweight deep learning-based classification model to classify the RMB images with better classification performance. The main contributions of this work are specified below:
  • According to the authors’ knowledge, this is the first study to propose a lightweight classification model called microwave brain image network” (MBINet) to classify RMB tumor images using a new machine learning paradigm called the Self-organized operational neural network (Self-ONN) architecture.
  • The proposed MBINet model is implemented and investigated on the RMB tumor images to classify the brain images into six classes: non-tumor (NT), single benign tumor (BT), single malignant tumor (MT), double benign tumor (BBT), double malignant tumor (MMT), and single benign and single malignant tumor (BMT).
  • The Implementation of a sensor-based microwave brain imaging (SMBI) system with a fabricated tissue-imitating brain phantom model to investigate the imaging performance for generating the RMB tumor image dataset.
  • A new Self-ONN model, MBINet, four other Self-ONN models, two conventional CNN models, and three pretrained models (DenseNet201, ResNet50, and ResNet101) are investigated on the RMB tumor images to classify six classes to show the usefulness of the suggested MBINet classification model.
  • The proposed MBINet model is compared with the seven most recent state-of-the-art models to verify the classification outcomes.
The remaining part of the manuscript is structured as follows: Section 2 explains the SMBI implementation setup and image reconstruction process. The research methodology and materials, including dataset preparation and experimental methods, are discussed in Section 3. Section 4 discusses the results of the classification models for the raw RMB images. In the end, the paper is concluded in Section 5.

2. Stacked Antenna Sensor-Based Microwave Brain Imaging (SMBI) System Development and Image Reconstruction Process

2.1. Design and Development Process of the Sensor-Based Stacked Antenna

An experimental stacked antenna sensor-based microwave brain imaging (SMBI) system has been developed in this research to reconstruct microwave brain (RMB) images and examine system performance. It is worth mentioning here that a wideband antenna sensor with high gain and unidirectional characteristics is required with a frequency band of 1 GHz to 4 GHz for the MBI system [10,14,17,18,20,21,25,47,48,49]. A new metamaterial (MTM)-inspired 3D wideband stacked antenna sensor has been printed on inexpensive Rogers RO4350B and RT5880 substrate materials. Three (03) substrate layers, together with two air gaps, comprise the antenna sensor. Double-sided foam tape is used to attach the layers together. The bottom layer (BL) is printed on an RO4350B substrate, while the top and middle layers are printed on a RT5880. The air gap in the middle is 2 mm. In the top and middle layers, a single 1 × 4 MTM array component is employed, while a single 3 × 2 MTM array component is employed in the BL. In order to improve antenna performance in terms of effectiveness, realized gain, bandwidth, radiation directionality in open space, and near proximity to the head model, MTM array elements are utilized in layers. The optimized dimension of the antenna sensor is 50 × 40 × 8.66 mm3. The antenna was measured in both free space and near the head model to ensure antenna performance. The measurements reveal that the antenna has an appropriate field penetration in the head, a fractional bandwidth (FBW) of 79.20 percent (1.37 to 3.16 GHz), 93 percent radiation efficiency, a 98 percent maximum fidelity factor, and 6.67 dBi gain. The fabricated antenna and reflection coefficient (S-parameters) measurements are shown in Figure 1. These results ensure that the antenna is able to produce the desired RMB images from the implemented SMBI system. We utilized our new MTM loaded 3D wideband stacked antenna in the SMBI system framework to generate RMB images for this research.

2.2. Phantom Model Fabrication Process and SMBI System Implementation Process

At first, a six-layered, tissue-imitating phantom with tumor (benign and malignant) tissues was made for validating the SMBI system. The layers and tumors were constructed as stated in the recipe in [50]. The electrical characteristics of the tumors (i.e., malignant and benign) were deemed to be stated quantities in [51]. The malignant tumor was formed in an irregular elliptical and triangle shape, whereas the benign tumor was created in a roughly round form with a typical shape [51]. Later, layer by layer, the tissue-imitating phantoms and tumors were added to the 3D head model. For the purpose of performing the measurement using the brain imaging equipment, the benign and malignant tumors were put into the model at various locations. The benign and malignant tumors were embedded into the skull model in various places. The simulated and measured S-parameters of the antenna sensor with the formulated tissue-imitating head model, including tumors, are shown in Figure 2a.
Additionally, nine antenna sensor arrays set up inside the helmet are shown in Figure 2b. Figure 2c indicates the complete experimental SMBI system. The system comprised nine stacked antenna sensor arrays, a stepper motor, an adjustable stand, a custom-made helmet, an RF switch, a PNA E8358A transceiver, and a microcontroller. The movable platform, to which the stepper motor is attached, turns counterclockwise with a 7.2° angle at each step to cover the entire (360°) area. The helmet is tightly attached to the motor’s shaft. The diameter of the helmet is 300 mm. Through the use of double-sided foam tape (DSFT), the antenna sensor is secured inside the helmet. In order to cover the entire system, there must be a 40° angle between each antenna. The sensor location is set 100 mm up from the lowest point of the helmet to fine-tune the phantom head position. Through the GPIB port, the PNA is connected to the computer. Port A is linked to the transmitting antenna, while Port B is linked to an RF switch to receive backscattered signals. For assessing the effectiveness of the system, a six-layered 3D phantom model is constructed and mounted in the middle of the helmet. The PNA accumulated the reflected sensor signals (S21, S31, … S91) after each 7.2-degree rotation.
Thereafter, the obtained sensor waves were processed and operated by the iteratively corrected delay-multiply-and-sum (IC-DMAS) image generation algorithm [52] to produce RMB images. It is notable that recently, in vivo imaging technology has been applied in medical science for brain tumor diagnostic purposes. Due to the fact that it is non-invasive, it can visualize living organs (for example, brain tumor imaging) and accurately detect target object locations in the region of interest. However, a live object such as an animal or human body is needed to produce an image, which is very challenging and a medical permission issue. In this research, we cannot use a live brain or human body for generating tumor-based images because of clinical trial limitations in our research lab. As a result, we are unable to collect in vivo image samples. Thus, we used a fabricated tissue-mimicking brain phantom model, which has the actual properties of a real brain, and collected all data from the system. Following that, all of the phantom-based data was analyzed and processed in order to generate brain tumor-based images for further evaluation. However, our future aim is to collect in vivo images for further evaluation, which will help medical doctors classify the tumor easily.

2.3. Illustration of RMB Image Samples

Figure 3 depicts samples of the tissue-mimicking phantom, different tumor scenarios, and corresponding RMB images. Figure 3a depicts the layout of the phantom model for visual comprehension. In addition, the tumors were positioned at various locations on the head model. Figure 3a–f depict the six image classes: the NT image, a single BT image, a single MT image, a double BBT image, a double MMT image, and a single benign and malignant tumor (BMT) image. After that, we collected nine hundred twenty (920) samples, including all classes, to make the original RMB image dataset. Later, different pre-processing steps were applied to the images to train, validate, and test the classification models. The proposed classifier was investigated by utilizing the original RMB images. Image preprocessing and augmentation processes were applied to the collected image samples to produce a large enough training dataset.

3. Methodology and Materials

This segment explains the methodology, dataset clarification, preprocessing method, data augmentation processes, and investigational analysis. The summary of the overall research methodology is stated in Figure 4. This research work utilized reconstructed microwave brain (RMB) images, which were collected from two data sources: (i) this research work (implemented experimental MBI system), and (ii) our previous research work [53] to enrich the dataset for classification models. Generally, the research has primarily focused on two distinct categories of images, such as (i) healthy brain (i.e., non-tumor (NT) images) and (ii) unhealthy brain (i.e., tumor-based images). The unhealthy brain images are categorized into five classifications: (i) single benign tumor (BT) images, (ii) single malignant tumor (MT) images, (iii) two benign tumor (BBT) images, (iv) two malignant tumor (MMT) images, and (v) single benign and single malignant tumor (BMT) images. The work explored the proposed lightweight microwave brain image network (MBINet) classification model; four Self-ONN classification models; two CNN-based models; and two pretrained models were used to inspect the classification performances for the six class classifications: NT, BT, MT, BBT, MMT, and BMT.

3.1. Preparation of Image Dataset

In this research, the dataset is prepared by collecting the image data from two sources: our currently implemented MBI system, and our previous research [53]. The combination of two datasets enriches the training dataset for better classification performance. The dataset consists of a total of 1320 original RMB images, where there are three hundred (300) images for the NT class, two hundred fifteen (215) images for each BT and MT class, two hundred (200) images for each BBT and MMT class, and one hundred ninety (190) images for the BMT class. Some samples of the original RMB images for all classes are displayed in Figure 5. The source code and original RMB image dataset can be found at: https://github.com/Amran038/Microwave-Brain-Image, accessed on 20 November 2022.

3.2. Data Preprocessing and Augmentation Process

This section illuminates the data preprocessing and dataset formulation for the experimental deep learning models. The data preprocessing is the starting stage of a DL model because of model’s input constraints. The different classification network models have different input size requirements. Thus, image data is pre-processed (resized and normalized) before training the models. For the raw RMB tumor image classification purposes, the images are resized to 224 × 224 pixels for four Self-ONNs, two vanilla CNNs, two pretrained models (DenseNet201 and ResNet50), and the proposed MBINet models. The original dataset’s images are normalized using the z-score normalization approach, the mean (M), and standard deviation (SD) of each image. Deep learning models typically require a large image dataset to effectively train the model to classify the images.
The image augmentation technique is employed in this study to build a large training dataset because the small dataset is insufficient for training the models. Instead of gathering additional information or samples, image augmentation might improve the performance of the models. It can significantly increase the diversity of data available for training the models as well as create a rich dataset from a small sample image dataset, which helps to enhance network performance. Different types of image augmentation techniques can be used to enrich the training dataset, such as rotation, scaling, translating, horizontal and vertical flipping, zooming, cropping, anatomically guided distortion elimination, etc. In addition, in particular, realistic augmentation based on recent state-of-the-art image synthesis techniques can be considered for image or data augmentation. This augmentation approach can be very useful for multimodal MRI-CT and multi-contrast MRI medical images (i.e., MRI, CT, PET, X-ray mammography, etc.). For example, SynDiff is the adversarial diffusion model, which is used for medical image synthesis and translation purposes [54]. This model is also used for image sampling. A diffusion probabilistic model was used for image synthesis and data scaling as an augmentation process [55]. In data scaling, authors have assumed that an image dataset consists of integers in the range of 0 to 255 as a pixel value, which scales linearly to [−1, 1] for reducing the data dimensionality [55].
Furthermore, in our microwave imaging technology, the produced images are two-dimensional and almost noiseless, and the target object’s (i.e., tumor) perceptibility is good. As a result, anatomically guided distortion and adversarial diffusion augmentation processes are not needed to apply to RMB images. For this reason, four different image augmentation techniques (e.g., rotation, scaling, translating, and flipping) were used in this investigation to generate the training dataset.
The images are rotated in both clockwise and counterclockwise directions at an angle ranging from 10 to 40 degrees. The tumor objects are thus relocated at various locations within the images. Scaling is the process of reducing or enlarging an image. Here, 10% to 12% image magnifications are employed. The image translation technique shifts the tumor objects to different locations in the images by translating the images by 10% to 15% in both the vertical and horizontal directions. The vertical flipping method is also used as an augmentation technique. This study used a five-fold cross-validation technique for training, validation, and testing purposes. Eighty percent of the total images were utilized for training, and twenty percent were used for testing in order to do five-fold cross-validation. Additionally, 20% of the training dataset, which comprises 80 percent of the dataset, is used for validation to prevent overfitting. After augmentation, 13,200 images were created for training, 264 images for testing, and 231 images for validation per fold. The complete dataset explanation is shown in Table 2. However, after pre-processing and augmentation, samples of the augmented images for all classes are demonstrated in Figure 6.

3.3. Experiments

This study uses the PyTorch library with the Python 3.7 version to construct and run nine alternative classification models, including the proposed MBINet model, on the Anaconda distribution platform. The tests are carried out on a 64-bit version of Windows 10 with 128 GB of RAM and a 3.30 GHz 64-bit Intel(R) Xeon(R) W-2016 processor. Additionally, the network training performance is accelerated using an 11 GB NVIDIA GeForce GTX 1080Ti GPU. The classification performance metrics of the five folds were calculated.

3.4. Proposed Microwave Brain Image Network (MBINet) Model—Brain Tumor Classification Model

Recently, an operational neural network (ONN)-based model was established in [56] to overcome the linear nature of the CNN. The ONN is a heterogeneous network that has demonstrated promising performance in a number of applications (image denoising, image restoration, and image classification) [46,57]. It uses a fixed set of nonlinear operators to learn complicated patterns from any input. However, the fixed set of operator libraries restricts ONN’s ability to learn. Self-ONN (Self-organized ONN) is employed to address this issue [58]. Instead of a fixed collection of operator libraries, Self-ONN inevitably discovers the best set of operators throughout the training phase. This develops a more vigorous model that can carry out a wider range of events and generalize effectively in practical situations. Self-ONN networks choose the best set of operators during the training process, which can be a combination of any standard function or some other function that we do not know. The output O k L at k th neuron of L th layer of any ONN can be illustrated as follows:
O k L = b k L + i = 1 N L 1 Ψ k i L ( w k i L , y i L 1 )
where, b k L and w k i L are the weights and biases connected to that layer and neuron, y i L 1 is the input from the previous layer, N L 1 is the kernel size of layers, Ψ k i L is the nodal operator of neurons and layers. If Ψ k i L is linear than the equation simply corresponds to conventional CNN. In ONN, the combined operator Ψ can be formed by a set of standard functions as follows:
Ψ ( w , y ) = w 1 s i n ( w 2 y ) + w 3 e x p ( w 4 y ) + + w q y
Here, w denotes the q-dimensional array of parameters, which is composed of internal parameters of the individual functions and weights. The combined nodal operator Ψ can be built using a Taylor series function rather than a predefined set of operators. The Taylor series is a function f ( x ) , near point, x = a is stated by the following equation:
f ( x ) = f ( a ) + f ( a ) 1 ! ( x a ) + f ( a ) 2 ! ( x a ) 2 + f ( a ) 3 ! ( x a ) 3 + + f n ( a ) n ! ( x a ) n
The Equation (3) can be used to construct the nodal operator as follows:
Ψ ( w , y ) = w 0 + w 1 ( y a ) + w 2 ( y a ) 2 + + w q ( y a ) q
Here, w q = f ( n ) ( a ) q ! denotes the q th parameter of the q th order polynomial. In Self-ONN, Tanh has been employed as an activation function that has a [−1, 1] range-boundary. So, for tanh, a is equal to zero in Equation (4).
We developed a new lightweight classification model called microwave brain image Network (MBINet) for classifying the RMB tumor images into six classes: (i) NT, (ii) BT, (iii) MT, (iv) BBT, (v) MMT, and (vi) BMT. The MBINet is constructed by utilizing a self-organized operational neural network (Self-ONN) architecture. The overall architecture of the MBINet classification model is depicted in Figure 7. From Figure 7, it is observed that the MBINet has a total of eight layers, including seven Self-ONN layers and one MLP (multilayer perception) layer. In the architecture, the first five layers have eight neurons, two have 16 neurons, and one has 32 neurons, respectively. Through the self-organization of its nodal operators, it can accomplish the required non-linear transformations to extract optimal features from the brain tumor images. The kernel sizes are set to ×3 for all Self-ONN layers. The kernel size of max pooling layers is set to 2 × 2 for the 1st, 3rd, 5th, and 7th layers, and 3 × 3 for the 8th layer to make it lightweight. Moreover, the Q value is set to 3 as the order of q th order polynomial for all operational layers. The input image dimension is set to 224 × 224 with 3 channels that are fed to the input layer of the model. Images are propagated through the Self-ONN and Max polling layers, and features are extracted into various feature maps. A flattening layer with 512 neurons is used to convert the output of the convolutional layer into a one-dimensional feature vector and apply it to the final MLP layer. The MLP layer is the final classifier of the network. It uses six neuron layers followed by the SoftMax activation function to classify the upcoming images into six classes: NT, BT, MT, BBT, MMT, and BMT classes.

Experimental Analysis of the Classification Models

In this section, we discuss six classification experiments to investigate the classification performances of the proposed MBINet by using the RMB images. However, the proposed MBINet model and four variations of the Self-ONN-based model, such as two Self-ONN models with 4 operational layers and two with 6 operational layers (Self-ONN4L1DN, Self-ONN4L, Self-ONN6L, and Self-ONN6L1DN), as well as two vanilla CNN models with 6 and 8 layers (Vanilla CNN6L and Vanilla CNN8L), and two pretrained models (DenseNet201 and ResNet50), were investigated and the results were compared separately using the raw RMB tumor images. In the model names, “4L” means the model consists of four layers, “6L” means the model consists of six layers, and “1DN” means the model consists of one dense layer in the final stage. The training was executed using a learning rate (LR) of 0.0005 for a maximum of 30 epochs, a batch size of 16, utilized the Adam optimizer for network optimization, and set stop criteria based on training loss. The Q order value is a significant factor during training the models; Q = 1 is set to train the two vanilla CNNs and the three pretrained models, and Q = 3 is set for the Self-ONN and MBINet models. The hyperparameters for the classification models are presented in Table 3.
It is obviously true that overfitting is a major issue in machine learning models, which degrades the performance of the model. Overfitting occurs when the selected model fits more data than is required and tries to capture every piece of data fed to the model. Hence, the model starts capturing noise, imbalance, and inaccurate data or images from the dataset, which reduces its performance and shows high variance and low bias. In order to avoid overfitting problems, four criteria were used in the proposed model: (i) the cross-validation method, (ii) training with more data samples, (iii) stop criteria based on validation loss, and (iv) epochs of patience. In this experiment, a five-fold cross-validation dataset was used, along with random shuffling of the dataset. Following that, the dataset is split using five-fold stratified cross-validation for training, validation, and testing. Eighty percent of the total images were utilized for training, and twenty percent were used for testing in order to do five-fold cross-validation. Additionally, 20% of the training dataset, which comprises 80 percent of the dataset, is used for validation purposes. As a result, the dataset was more generalized, which reduced the biasedness of the model during training and testing result evaluation. If validation loss remained constant after 5 epochs of training, the training process was terminated. Otherwise, training of the model continued up to 30 epochs. Then, the fold-wise performance of the MBINet model was observed. The training, validation, and testing accuracy and loss plots for five folds with respect to epochs of the proposed MBINet model are illustrated in Figure 8. It is seen from Figure 8a that when the model was trained with the Fold-1 dataset, it got saturated after 10 epochs and stopped training after 15 epochs. However, the model achieved very low testing and validation accuracy. On the other hand, the model achieved high testing and validation losses (Figure 8b), but was not overfit. Moreover, when the model was trained with the Fold-2, Fold-3, and Fold-4 datasets, it showed that the testing and validation accuracy gradually increased, and the corresponding losses gradually decreased. Additionally, when the model trained with the Fold-5 dataset, it achieved high training, validation, and testing accuracy with low losses, which are shown in Figure 8i,j. In Fold-5, the model performance gets saturated after 5 epochs, and the model is not overfitted and converges well. Furthermore, the model tried to capture every image fed to it to enhance the classification performance. As a result, the training, testing, and validation accuracy and corresponding losses are almost similar, which ensures the better performance of the model. It is notable that the data splits may permit an unbiased evaluation of the model. As a result, the proposed model is able to classify the RMB images reliably with high classification performance.

3.5. Evaluation Matrix for the Classification Model

The classification performance of the various CNN and Self-ONN models is assessed by the five assessment matrices, such as: (i) overall accuracy (A), (ii) weighted recall or sensitivity (R), (iii) weighted specificity (S), (iv) weighted precision (P), and (v) weighted F1-score (Fs). The evaluation metrics are computed by applying the subsequent equations:
A = ( N T P + N T N ) ( N T P + N F N ) + ( N F P + N T N )
R = N T P ( N T P + N F N )
S = N T N ( N F P + N T N )
P = N T P ( N T P + N F P )
F s = ( 2 × N T P ) ( 2 × N T P + N F N + N F P )
where NTP denotes the number of tumor samples detected as tumors, NTN represents the number of non-tumor image samples recognized as non-tumors, NFP denotes the number of samples incorrectly recognized as tumors, NFN denotes the number of samples with tumors that were missed by the network.

4. Results and Discussion

4.1. Raw RMB Image Classification Performances

In this study, the main advantages of the MBINet model are: (i) a lightweight architecture that uses non-linear operations to increase network diversity as well as classification effectiveness; (ii) the ability to optimize the learning weight of each layer during the training process; and (iii) the ability to achieve superior classification performances while significantly reducing computational complexity compared to conventional CNN models. This section discusses the four Self-ONNs (Self-ONN4L1DN, Self-ONN4L, Self-ONN6L, and Self-ONN6L1DN), two vanilla CNNs (vanilla CNN6L and vanilla CNN8L), three pretrained models (DenseNet201, ResNet50, and ResNet101), and proposed MBINet classification models to investigate the classification effectiveness by applying the original RMB images. The classification models can classify the images into six classes: non-tumor (NT), single benign tumor (BT), single malignant tumor (MT), two benign tumors (BBT), two malignant tumors (MMT), and single benign and single malignant (BMT) tumor classes.
In addition, all classification models were trained using the raw RMB brain tumor images. The comparative classification performance outcomes of the models for the raw RMB brain tumor images are presented in Table 4. It was found that conventional deeper CNN networks have achieved lower performances than the four Self-ONNs models, but the proposed MBINet model was the best model among all networks and achieved the highest performances. The MBINet has exhibited accuracy, precision, recall, specificity, and an F1 score of 96.97%, 96.93%, 96.85%, 97.95%, and 96.83%, respectively, for the raw RMB brain images. In addition, the mean (M) and standard deviations (STD) were considered for quantitative assessments. Further, in these assessments, the test dataset was split into five folds, where every fold consists of 264 images. The calculated average M and STD for five folds are presented in Table 5. It is noted from Table 5 that the proposed model showed lower STD values than other models, which means there was very little variance. Therefore, it is concluded that the proposed classification model exhibited better performance than the other models for classifying the RMB brain images into six classes.

4.2. Receiver Operating Characteristics (ROC) Analysis

In multi-class classification issues, the ROC curve is a crucial evaluation metric. A classification model’s performance across all thresholds can be seen using the ROC curve. Additionally, it demonstrates the ability to distinguish across classes. Figure 9 depicts the classification ROC and area under the curve (AUC) for all classification models across all thresholds. Figure 9 exhibited the ROC with AUC for raw RMB image classification and showed that the proposed MBINet model performed better. The computed AUCs for Self-ONN4L1DN, Self-ONN4L, Self-ONN6L, Self-ONN6L1DN, vanilla CNN8L, vanilla CNN6L, DenseNet201, ResNet50, and the proposed MBINet are 88.49%, 88.60%, 90.03%, 88.41%, 89.70%, 89.55%, 91.83%, 95.13%, and 97.21%, respectively. It is observed that the MBINet model performed better than other state-of-the-art models, and it can be reliable for RMB image classification.

4.3. Performance Analysis

It was determined from the classification performances in Table 4 and Table 5 that the best classification model was MBINet for classifying the RMB images. The overall classification accuracy was 96.97% for the raw images. For classification results, the confusion matrix of the MBINet model for the raw RMB brain images is illustrated in Figure 10. It is observed from the confusion matrix that the model has been classified at 100%, 97.67%, 97.20%, 96.00%, 96.50%, and 96.84% for NT, BT, MT, BBT, MMT, and BMT classification. It is illustrated that a total of 32 images were misclassified out of 920 images during the testing of the model. For instance, eight misclassified images from four classes (BT, MT, BBT, and MMT) are illuminated in Figure 11.
Figure 10 shows that the BT images were incorrectly classified as NT, MT, and BBT classes; the MT images were incorrectly classified as NT, BT, MMT, and BBT classes; and the MMT images were incorrectly classified as MT, BBT, and BMT classes. In addition, the misclassification rate was comparatively low for the proposed model. However, through the training of Self-ONNs, the optimum non-linear parameters can be learned to exploit the learning performance and attain a superior classification performance in terms of non-tumor and tumor images. Finally, the comparison outcomes of the proposed model with the existing models by applying the same dataset (i.e., the experimental image dataset) are presented in Table 6. The performance metrics: Accuracy (Acc.), Precision (Prc.), Recall (Rec.), Specificity (Spec.), F1-score (Fs), and Overall Classification Performance (OCP) of existing models were calculated and presented in Table 6. It is observed from Table 6 that the proposed MBINet model performed better and showed satisfactory outcomes than other models for the RMB tumor image classification. Finally, it is concluded that the MBINet classification model improved the classification performance and is applicable in the SMBI system for classifying the RMB tumor images into six classes.

5. Conclusions and Future Directions

This paper presents the brain tumor classification from the RMB tumor images through a lightweight, deep learning-based microwave brain image network (MBINet) model. The MBINet is based on a self-organized operational neural network. In the beginning, a compact 3D stacked wideband nine antenna array sensor was utilized to implement the SMBI system that produced reconstructed microwave brain (RMB) images, and then 920 raw RMB image samples were collected. In addition, another RMB dataset was collected from our previous work to enrich the training dataset. Later, a lightweight microwave brain image network (MBINet) classifier model was applied to classify the raw RMB images into six classes (NT, BT, MT, BBT, MMT, and BMT). MBINet uses non-linear operations to boost network diversity, increase computational effectiveness, and attain superior classification performance. Furthermore, the MBINet, four Self-ONN classification models, two conventional CNN models, and three pretrained models were examined using the original RMB images, and the classification outcomes were compared. Compared with the state-of-the-art models, the proposed MBINet classification model performed better. The achieved accuracy, precision, recall, specificity, and F1 score of the MBINet model are 96.97%, 96.93%, 96.85%, 97.95%, and 96.83%, respectively, for six class classifications using the raw RMB images. The MBINet model showed better classification results than other models. Further, it is concluded that the MBINet model can be used for consistently classifying the brain tumor(s) from the RMB images and can be utilized in the SMBI system.

5.1. Research Shortfalls and Future Improvement

We used the M-DMAS image reconstruction algorithm in this study, which can only reconstruct non-tumor images and two tumor-based images, which is one of the algorithm’s shortfalls. However, if more than two tumors or any other types of tumors, such as meningiomas, pituitary adenomas, craniopharyngiomas, etc., are formed in the brain, the algorithm will not reconstruct the images. On the other hand, in the proposed classification model, the learning outcomes of the MBINet depend on the nodal operators and Q-order parameter values, which must be fixed in advance, which is another shortcoming of the model. In other words, if the right operator setting for proper learning is lacking, the learning outcomes will decrease. Moreover, there is an inadequate discrepancy due to the usage of one nodal operator set for every one of the neurons in a hidden layer. Keeping in mind the mentioned limitations, we can focus on improving the following for our future work: (i) Implementation of a new image reconstruction algorithm that will reconstruct different types of tumors with high-resolution images; (ii) Proper ingredient selection and quantity for fabricating the different types of tumors; (iii) Assessment of the classification performance of the proposed model for classifying different types of tumors by optimizing learning parameters and Q-order.

5.2. Future Directions

Based on our evaluations of the research that has already been done, we can suggest a few directions for future research: (i) There is an opportunity to adapt a better optimization algorithm (i.e., AdaGad [59], SGD [60], Adam [61], RMSProp [62]) for training, which should be a modified MBINet model for proper functioning, (ii) Determine an optimal Q-order value that can be automatically used in layers for enhancing the classification performance instead of fixing the Q-value; (iii) Computational complexity is the crucial issue for the Self-ONN model, so finding a computational complexity and inference time reduction mechanism is another research opportunity; and (iv) The model can be assessed by using a large multi-modal or 3D microwave brain image dataset.

Author Contributions

A.H. made extensive contributions to conception, program analysis and design, data acquisition, planning, result investigation, measurement, experimental validation, writing the original draft and revision of the article. Program analysis, result investigation, measurement, and experimental validation were done by S.K.A.R., T.R., M.A.R., A.K., M.A.A., M.E.H.C., A.H. and M.T.I. Experiments were performed by A.H., T.R. and M.E.H.C.; H.A., M.T.I. and M.E.H.C. participated in the supervision, resource management, and funding acquisition. M.E.H.C., S.K.A.R., M.A.A. and H.A. participated in the intellectual concept, editing, and review of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Universiti Kebangsaan Malaysia project grant code DIP-2021-024. This work was also supported by Grant NPRP12S-0227-190164 from the Qatar National Research Fund, a member of the Qatar Foundation, Doha, Qatar, and the claims made herein are solely the responsibility of the authors. Open access publication is supported by the Qatar National Library.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wyant, Y.; Ogoro, R.A.C. Key Statistics for Brain and Spinal Cord Tumors. Available online: https://www.cancer.org/cancer/brain-spinal-cord-tumors-adults/about/key-statistics.html (accessed on 7 September 2021).
  2. Chaturia Rouse, H.G.; Ostrom, Q.T.; Kruchko, C.; Barnholtz-Sloan, J.S. Brain Tumor Facts. Available online: https://braintumor.org/brain-tumors/about-brain-tumors/brain-tumor-facts (accessed on 20 November 2022).
  3. Tariq, M.; Siddiqi, A.A.; Narejo, G.B.; Andleeb, S. A cross sectional study of tumors using bio-medical imaging modalities. Curr. Med. Imaging 2019, 15, 66–73. [Google Scholar] [CrossRef] [PubMed]
  4. Ahmad, H.A.; Yu, H.J.; Miller, C.G. Medical imaging modalities. In Medical Imaging in Clinical Trials; Springer: Berlin/Heidelberg, Germany, 2014; pp. 3–26. [Google Scholar]
  5. Chakraborty, S.; Chatterjee, S.; Ashour, A.S.; Mali, K.; Dey, N. Intelligent computing in medical imaging: A study. In Advancements in Applied Metaheuristic Computing; IGI Global: Hershey, PA, USA, 2018; pp. 143–163. [Google Scholar]
  6. Frangi, A.F.; Tsaftaris, S.A.; Prince, J.L. Simulation and synthesis in medical imaging. IEEE Trans. Med. Imaging 2018, 37, 673–679. [Google Scholar] [CrossRef] [PubMed]
  7. Adamson, E.B.; Ludwig, K.D.; Mummy, D.G.; Fain, S.B. Magnetic resonance imaging with hyperpolarized agents: Methods and applications. Phys. Med. Biol. 2017, 62, R81. [Google Scholar] [CrossRef] [PubMed]
  8. Cazzato, R.L.; Garnon, J.; Shaygi, B.; Koch, G.; Tsoumakidou, G.; Caudrelier, J.; Addeo, P.; Bachellier, P.; Namer, I.J.; Gangi, A. PET/CT-guided interventions: Indications, advantages, disadvantages and the state of the art. Minim. Invasive Ther. Allied Technol. 2018, 27, 27–32. [Google Scholar] [CrossRef] [PubMed]
  9. Jacobs, M.A.; Ibrahim, T.S.; Ouwerkerk, R. MR imaging: Brief overview and emerging applications. Radiographics 2007, 27, 1213–1229. [Google Scholar] [CrossRef]
  10. Alqadami, A.S.; Bialkowski, K.S.; Mobashsher, A.T.; Abbosh, A.M. Wearable electromagnetic head imaging system using flexible wideband antenna array based on polymer technology for brain stroke diagnosis. IEEE Trans. Biomed. Circuits Syst. 2018, 13, 124–134. [Google Scholar] [CrossRef]
  11. Dougeni, E.; Faulkner, K.; Panayiotakis, G. A review of patient dose and optimisation methods in adult and paediatric CT scanning. Eur. J. Radiol. 2012, 81, e665–e683. [Google Scholar] [CrossRef]
  12. Jones, K.M.; Michel, K.A.; Bankson, J.A.; Fuller, C.D.; Klopp, A.H.; Venkatesan, A.M. Emerging magnetic resonance imaging technologies for radiation therapy planning and response assessment. Int. J. Radiat. Oncol. 2018, 101, 1046–1056. [Google Scholar] [CrossRef]
  13. Hossain, A.; Islam, M.T.; Almutairi, A.F.; Singh, M.S.J.; Mat, K.; Samsuzzaman, M. An octagonal ring-shaped parasitic resonator based compact ultrawideband antenna for microwave imaging applications. Sensors 2020, 20, 1354. [Google Scholar] [CrossRef]
  14. Hossain, A.; Islam, M.T.; Chowdhury, M.E.; Samsuzzaman, M. A grounded coplanar waveguide-based slotted inverted delta-shaped wideband antenna for microwave head imaging. IEEE Access 2020, 8, 185698–185724. [Google Scholar] [CrossRef]
  15. Mobashsher, A.; Bialkowski, K.; Abbosh, A.; Crozier, S. Design and experimental evaluation of a non-invasive microwave head imaging system for intracranial haemorrhage detection. PLoS ONE 2016, 11, e0152351. [Google Scholar] [CrossRef]
  16. Tobon Vasquez, J.A.; Scapaticci, R.; Turvani, G.; Bellizzi, G.; Joachimowicz, N.; Duchêne, B.; Tedeschi, E.; Casu, M.R.; Crocco, L.; Vipiana, F. Design and experimental assessment of a 2D microwave imaging system for brain stroke monitoring. Int. J. Antennas Propag. 2019, 2019, 8065036. [Google Scholar] [CrossRef]
  17. Hossain, A.; Islam, M.T.; Islam, M.T.; Chowdhury, M.E.; Rmili, H.; Samsuzzaman, M. A planar ultrawideband patch antenna array for microwave breast tumor detection. Materials 2020, 13, 4918. [Google Scholar] [CrossRef]
  18. Islam, M.S.; Islam, M.T.; Hoque, A.; Islam, M.T.; Amin, N.; Chowdhury, M.E. A portable electromagnetic head imaging system using metamaterial loaded compact directional 3D antenna. IEEE Access 2021, 9, 50893–50906. [Google Scholar] [CrossRef]
  19. Mobashsher, A.T.; Abbosh, A.M.; Wang, Y. Microwave system to detect traumatic brain injuries using compact unidirectional antenna and wideband transceiver with verification on realistic head phantom. IEEE Trans. Microw. Theory Tech. 2014, 62, 1826–1836. [Google Scholar] [CrossRef]
  20. Salleh, A.; Yang, C.; Alam, T.; Singh, M.; Samsuzzaman, M.; Islam, M. Development of microwave brain stroke imaging system using multiple antipodal vivaldi antennas based on raspberry Pi technology. J. Kejuruterran 2020, 32, 39–49. [Google Scholar]
  21. Stancombe, A.E.; Bialkowski, K.S.; Abbosh, A.M. Portable microwave head imaging system using software-defined radio and switching network. IEEE J. Electromagn. RF Microw. Med. Biol. 2019, 3, 284–291. [Google Scholar] [CrossRef]
  22. Chew, K.M.; Yong, C.Y.; Sudirman, R.; Wei, S.T.C. Bio-signal processing and 2D representation for brain tumor detection using microwave signal analysis. In Proceedings of the 2018 IEEE Symposium on Computer Applications & Industrial Electronics (ISCAIE), Penang Island, Malaysia, 28–29 April 2018; pp. 303–309. [Google Scholar]
  23. Rezaeieh, S.A.; Zamani, A.; Abbosh, A. 3-D wideband antenna for head-imaging system with performance verification in brain tumor detection. IEEE Antennas Wirel. Propag. Lett. 2014, 14, 910–914. [Google Scholar] [CrossRef]
  24. Rodriguez-Duarte, D.O.; Vasquez, J.A.T.; Scapaticci, R.; Crocco, L.; Vipiana, F. Brick-shaped antenna module for microwave brain imaging systems. IEEE Antennas Wirel. Propag. Lett. 2020, 19, 2057–2061. [Google Scholar] [CrossRef]
  25. Rokunuzzaman, M.; Ahmed, A.; Baum, T.C.; Rowe, W.S. Compact 3-D antenna for medical diagnosis of the human head. IEEE Trans. Antennas Propag. 2019, 67, 5093–5103. [Google Scholar] [CrossRef]
  26. Ullah, M.; Alam, T.; Islam, M. A 3D directive microwave antenna for biomedical imaging application. Bull. Pol. Acad. Sci. Tech. Sci. 2018, 66, 355–360. [Google Scholar]
  27. Gerazov, B.; Conceicao, R.C. Deep learning for tumour classification in homogeneous breast tissue in medical microwave imaging. In Proceedings of the IEEE EUROCON 2017-17th International Conference on Smart Technologies, Ohrid, North Macedonia, 6–8 July 2017; pp. 564–569. [Google Scholar]
  28. Khoshdel, V.; Asefi, M.; Ashraf, A.; LoVetri, J. Full 3D microwave breast imaging using a deep-learning technique. J. Imaging 2020, 6, 80. [Google Scholar] [CrossRef]
  29. Rana, S.P.; Dey, M.; Tiberi, G.; Sani, L.; Vispa, A.; Raspa, G.; Duranti, M.; Ghavami, M.; Dudley, S. Machine learning approaches for automated lesion detection in microwave breast imaging clinical data. Sci. Rep. 2019, 9, 10510. [Google Scholar] [CrossRef]
  30. Salucci, M.; Polo, A.; Vrba, J. Multi-step learning-by-examples strategy for real-time brain stroke microwave scattering data inversion. Electronics 2021, 10, 95. [Google Scholar] [CrossRef]
  31. Shah, P.; Moghaddam, M. Super resolution for microwave imaging: A deep learning approach. In Proceedings of the 2017 IEEE International Symposium on Antennas and Propagation & USNC/URSI National Radio Science Meeting, San Diego, CA, USA, 9–14 July 2017; pp. 849–850. [Google Scholar]
  32. Shao, W.; Du, Y. Microwave imaging by deep learning network: Feasibility and training method. IEEE Trans. Antennas Propag. 2020, 68, 5626–5635. [Google Scholar] [CrossRef] [PubMed]
  33. Sharif, M.I.; Khan, M.A.; Alhussein, M.; Aurangzeb, K.; Raza, M. A decision support system for multimodal brain tumor classification using deep learning. Complex Intell. Syst. 2022, 8, 3007–3020. [Google Scholar] [CrossRef]
  34. Noreen, N.; Palaniappan, S.; Qayyum, A.; Ahmad, I.; Imran, M.; Shoaib, M. A deep learning model based on concatenation approach for the diagnosis of brain tumor. IEEE Access 2020, 8, 55135–55144. [Google Scholar] [CrossRef]
  35. Ding, Y.; Li, C.; Yang, Q.; Qin, Z.; Qin, Z. How to improve the deep residual network to segment multi-modal brain tumor images. IEEE Access 2019, 7, 152821–152831. [Google Scholar] [CrossRef]
  36. Ghassemi, N.; Shoeibi, A.; Rouhani, M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed. Signal Process. Control. 2020, 57, 101678. [Google Scholar] [CrossRef]
  37. Abd El Kader, I.; Xu, G.; Shuai, Z.; Saminu, S.; Javaid, I.; Salim Ahmad, I. Differential deep convolutional neural network model for brain tumor classification. Brain Sci. 2021, 11, 352. [Google Scholar] [CrossRef]
  38. Díaz-Pernas, F.J.; Martínez-Zarzuela, M.; Antón-Rodríguez, M.; González-Ortega, D. A deep learning approach for brain tumor classification and segmentation using a multiscale convolutional neural network. Healthcare 2021, 9, 153. [Google Scholar] [CrossRef]
  39. Kumar, R.L.; Kakarla, J.; Isunuri, B.V.; Singh, M. Multi-class brain tumor classification using residual network and global average pooling. Multimed. Tools Appl. 2021, 80, 13429–13438. [Google Scholar] [CrossRef]
  40. Zahid, U.; Ashraf, I.; Khan, M.A.; Alhaisoni, M.; Yahya, K.M.; Hussein, H.S.; Alshazly, H. BrainNet: Optimal deep learning feature fusion for brain tumor classification. Comput. Intell. Neurosci. 2022, 2022, 1465173. [Google Scholar] [CrossRef]
  41. Zhao, L.; Zhou, D.; Jin, X.; Zhu, W. nn-TransUNet: An Automatic Deep Learning Pipeline for Heart MRI Segmentation. Life 2022, 12, 1570. [Google Scholar] [CrossRef]
  42. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  43. Dalmaz, O.; Yurt, M.; Çukur, T. ResViT: Residual vision transformers for multimodal medical image synthesis. IEEE Trans. Med. Imaging 2022, 41, 2598–2614. [Google Scholar] [CrossRef]
  44. Korkmaz, Y.; Dar, S.U.; Yurt, M.; Özbey, M.; Cukur, T. Unsupervised MRI reconstruction via zero-shot learned adversarial transformers. IEEE Trans. Med. Imaging 2022, 41, 1747–1763. [Google Scholar] [CrossRef]
  45. Devecioglu, O.C.; Malik, J.; Ince, T.; Kiranyaz, S.; Atalay, E.; Gabbouj, M. Real-time glaucoma detection from digital fundus images using self-onns. IEEE Access 2021, 9, 140031–140041. [Google Scholar] [CrossRef]
  46. Kiranyaz, S.; Malik, J.; Abdallah, H.B.; Ince, T.; Iosifidis, A.; Gabbouj, M. Self-organized operational neural networks with generative neurons. Neural Netw. 2021, 140, 294–308. [Google Scholar] [CrossRef]
  47. Alqadami, A.S.; Nguyen-Trong, N.; Mohammed, B.; Stancombe, A.E.; Heitzmann, M.T.; Abbosh, A. Compact unidirectional conformal antenna based on flexible high-permittivity custom-made substrate for wearable wideband electromagnetic head imaging system. IEEE Trans. Antennas Propag. 2019, 68, 183–194. [Google Scholar] [CrossRef]
  48. Hossain, A.; Islam, M.T.; Islam, M.S.; Chowdhury, M.E.; Almutairi, A.F.; Razouqi, Q.A.; Misran, N. A YOLOv3 deep neural network model to detect brain tumor in portable electromagnetic imaging system. IEEE Access 2021, 9, 82647–82660. [Google Scholar] [CrossRef]
  49. Hossain, A.; Islam, M.T.; Misran, N.; Islam, M.S.; Samsuzzaman, M. A mutual coupled spider net-shaped triple split ring resonator based epsilon-negative metamaterials with high effective medium ratio for quad-band microwave applications. Results Phys. 2021, 22, 103902. [Google Scholar] [CrossRef]
  50. Mobashsher, A.; Abbosh, A. Three-dimensional human head phantom with realistic electrical properties and anatomy. IEEE Antennas Wirel. Propag. Lett. 2014, 13, 1401–1404. [Google Scholar] [CrossRef]
  51. Cheng, Y.; Fu, M. Dielectric properties for non-invasive detection of normal, benign, and malignant breast tissues using microwave theories. Thorac. Cancer 2018, 9, 459–465. [Google Scholar] [CrossRef]
  52. Islam, M.T.; Samsuzzaman, M.; Kibria, S.; Misran, N.; Islam, M.T. Metasurface loaded high gain antenna based microwave imaging using iteratively corrected delay multiply and sum algorithm. Sci. Rep. 2019, 9, 17317. [Google Scholar] [CrossRef]
  53. Hossain, A.; Islam, M.T.; Almutairi, A.F. A deep learning model to classify and detect brain abnormalities in portable microwave based imaging system. Sci. Rep. 2022, 12, 6319. [Google Scholar] [CrossRef]
  54. Özbey, M.; Dar, S.U.; Bedel, H.A.; Dalmaz, O.; Özturk, Ş.; Güngör, A.; Çukur, T. Unsupervised medical image translation with adversarial diffusion models. arXiv 2022, arXiv:2207.08208. [Google Scholar]
  55. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  56. Kiranyaz, S.; Ince, T.; Iosifidis, A.; Gabbouj, M. Operational neural networks. Neural Comput. Appl. 2020, 32, 6645–6668. [Google Scholar] [CrossRef]
  57. Malik, J.; Kiranyaz, S.; Gabbouj, M. Self-organized operational neural networks for severe image restoration problems. Neural Netw. 2021, 135, 201–211. [Google Scholar] [CrossRef]
  58. Malik, J.; Kiranyaz, S.; Gabbouj, M. Operational vs convolutional neural networks for image denoising. arXiv 2020, arXiv:2009.00612. [Google Scholar]
  59. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  60. Qian, N. On the momentum term in gradient descent learning algorithms. Neural Netw. 1999, 12, 145–151. [Google Scholar] [CrossRef] [PubMed]
  61. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  62. Kurbiel, T.; Khaleghian, S. Training of deep neural networks based on distance measures using RMSProp. arXiv 2017, arXiv:1708.01911. [Google Scholar]
Figure 1. (a) Fabricatted 3D stacked antenna, (b) Measured and simulated reflection coefficient.
Figure 1. (a) Fabricatted 3D stacked antenna, (b) Measured and simulated reflection coefficient.
Biosensors 13 00238 g001
Figure 2. Antenna sensor measurements and experimental setup for the SMBI system: (a) Measured and simulated results of the antenna sesnor with a fabricated head phantom model, (b) MTM loaded 3D stacked antenna sensor inside the helmet, (c) Overall SMBI system model.
Figure 2. Antenna sensor measurements and experimental setup for the SMBI system: (a) Measured and simulated results of the antenna sesnor with a fabricated head phantom model, (b) MTM loaded 3D stacked antenna sensor inside the helmet, (c) Overall SMBI system model.
Biosensors 13 00238 g002
Figure 3. Samples of formulated tissue-imitating head phantom models with RMB images: (a) Non-tumor (NT), (b) Single BT image, (c) Single MT image, (d) Two BBT image, (e) Two MMT image, (f) Single benign and single malignant tumor (BMT).
Figure 3. Samples of formulated tissue-imitating head phantom models with RMB images: (a) Non-tumor (NT), (b) Single BT image, (c) Single MT image, (d) Two BBT image, (e) Two MMT image, (f) Single benign and single malignant tumor (BMT).
Biosensors 13 00238 g003
Figure 4. The flowchart for the complete research process.
Figure 4. The flowchart for the complete research process.
Biosensors 13 00238 g004
Figure 5. Randomly selected RMB image samples from the original dataset: (a) NT class, (b) Single BT class, (c) Single MT class, (d) BBT class, (e) MMT class, (f) BMT class.
Figure 5. Randomly selected RMB image samples from the original dataset: (a) NT class, (b) Single BT class, (c) Single MT class, (d) BBT class, (e) MMT class, (f) BMT class.
Biosensors 13 00238 g005
Figure 6. Samples of the augmented training set (six classes): (af) Pre-processed original images, (gl) After rotating all class images by 30 degrees both clockwise and counterclockwise, (mr) All class images after 10% horizontal and vertical, and 8% horizontal and 12% vertical translation, and (sx) All class images after vertical flipping.
Figure 6. Samples of the augmented training set (six classes): (af) Pre-processed original images, (gl) After rotating all class images by 30 degrees both clockwise and counterclockwise, (mr) All class images after 10% horizontal and vertical, and 8% horizontal and 12% vertical translation, and (sx) All class images after vertical flipping.
Biosensors 13 00238 g006aBiosensors 13 00238 g006b
Figure 7. Proposed Microwave Brain Image Network (MBINet) model using Self-ONN.
Figure 7. Proposed Microwave Brain Image Network (MBINet) model using Self-ONN.
Biosensors 13 00238 g007
Figure 8. Training, validation, and testing accuracy and loss graphs for five-fold cross-validation dataset: (a,b) Fold-1, (c,d) Fold-2, (e,f) Fold-3, (g,h) Fold-4, (i,j) Fold-5.
Figure 8. Training, validation, and testing accuracy and loss graphs for five-fold cross-validation dataset: (a,b) Fold-1, (c,d) Fold-2, (e,f) Fold-3, (g,h) Fold-4, (i,j) Fold-5.
Biosensors 13 00238 g008aBiosensors 13 00238 g008b
Figure 9. Receiver Operating Characteristics (ROC) curve with AUC.
Figure 9. Receiver Operating Characteristics (ROC) curve with AUC.
Biosensors 13 00238 g009
Figure 10. The confusion matrix of the proposed MBINet classification model for the raw RMB brain images.
Figure 10. The confusion matrix of the proposed MBINet classification model for the raw RMB brain images.
Biosensors 13 00238 g010
Figure 11. Some misclassified image samples by the MBINet model for the raw RMB images: (a) BT images were misclassified as a NT class, (b) MT images were misclassified as a BT class, (c) BBT images were misclassified as a NT class, and (d) MMT images were misclassified as a BBT class.
Figure 11. Some misclassified image samples by the MBINet model for the raw RMB images: (a) BT images were misclassified as a NT class, (b) MT images were misclassified as a BT class, (c) BBT images were misclassified as a NT class, and (d) MMT images were misclassified as a BBT class.
Biosensors 13 00238 g011aBiosensors 13 00238 g011b
Table 1. Summarizes the pros and cons of the existing models mentioned in the literature review.
Table 1. Summarizes the pros and cons of the existing models mentioned in the literature review.
Ref.Name of the
Existing Model
ProsCons
[33]FT-DenseNet201The Entropy Kurtosis based technique is used in this model for feature extraction to improve the accuracy of the system and reduce the time of classification.The fusion process of the model increases the computational time and shows less accuracy in classifying the small-sized tumor images.
[34]PT-DenseNet201The multi-level features Information is extracted from different bottom layers of the model, which enhances its capability to classify the tumors.It requires a large amount of computational time and shows a low precision score and specificity score for noisy images.
[35]DP-DenseNetThe model is a combination of residual networks and dilated convolutional layers that can solve the vanishing gradient problem. It enhances image resolution and classification accuracy for multimodal brain tumor samples.The network model can only classify large-sized tumor-based images but not small-sized tumor images, resulting in comparatively poor classification performances. Further, it increases the computational time.
[37]Differential-DCNNThe model uses a differential operator with contrast calculation for analyzing the pixel-directional pattern of images. It is very good at accurately classifying a large set of tumor images.The model may fail to achieve high performance due to large dataset constraints, which increases the testing loss. As a result, the model failed to classify the small-sized tumor images.
[38]CMP-CNNThe CMP-CNN model employs three convolutional pathways for extracting discriminant texture features of different kinds of tumors and uses a multi-scale processing strategy for improving the tumor classification performance.It takes a long computational time to train the model and has caused false positives in a number of testing images due to the lack of variability among the three tumor types.
[39]ResNet-50The global average pooling method is used to solve the problems of vanishing gradient and overfitting in the deep network-based ResNet-50 model.It is computationally expensive as it uses a deep neural network. The classification accuracy is slightly lesser when compared to other classes.
[40]Fine-tuned ResNet101The model utilizes the differential evaluation method and particle swarm optimization algorithms to reduce redundant features and computational overhead.The computational time increases during the testing process because of the fusion process, which achieves less accuracy in classifying the small tumor images.
Table 2. Description of the training, testing, and validation datasets.
Table 2. Description of the training, testing, and validation datasets.
DatasetNumber of Original ImagesImage ClassesTraining Dataset
Number of Images Per ClassAugmented Train Images Per FoldTesting Images Per FoldValidation Image Per Fold
This WorkRef. [53]Total
Raw RMB Image Samples1320Non-Tumor (NT)20010030030006048
Single Benign Tumor (BT)1407521521504335
Single Malignant Tumor (MT)1407521521504335
Two Benign Tumors (BBT)1505020020004032
Two Malignant Tumors (MMT)1505020020004032
Single Benign and Single Malignant Tumor (BMT)1405019019003831
Total920400132013,200264213
Table 3. Hyperparameters for all classification models.
Table 3. Hyperparameters for all classification models.
Parameter’s NameAssigned ValueParameter’s NameAssigned Value
Input Channels for Color Image3Q order1 for CNN, 3 for Self-ONNs
OptimizerAdamBatch Size16
Image Size224Stop CriteriaLoss
Maximum Number of Epochs30Epochs Patience5
Maximum Epochs Stop10Learning Factor0.2
Number of Folds5Learning Rate (LR)0.0005
Standard Deviation (STD)[0.4116, 0.3645, 0.2597]Mean (M)[0.2552, 0.4666, 0.8804]
Table 4. Classification results of all models for the RMB brain images. Bold represents the best-performing model.
Table 4. Classification results of all models for the RMB brain images. Bold represents the best-performing model.
Image TypeName of the Network ModelOverallWeighted
Accuracy (A)Precession (P)Recall (R)Specificity (S)F1 Score (Fs)
RMB ImagesSelf-ONN4L1DN93.9093.4893.7794.3493.50
Self-ONN4L93.7693.5693.7694.2893.75
Self-ONN6L94.1994.8594.2995.4794.58
Self-ONN6L1DN95.5095.5395.2096.1195.31
Vanilla CNN8L92.9592.8992.9293.9592.78
Vanilla CNN6L92.8392.7692.9093.8792.45
DenseNet20194.5894.5594.2895.8494.80
ResNet5095.8995.9495.2996.8195.16
ResNet10195.9095.9695.8995.8995.86
MicrowaveBrainImage Network (MBINet)96.9796.9396.8597.9596.83
Table 5. Quantitative assessments of all models. Bold represents the best-performing model.
Table 5. Quantitative assessments of all models. Bold represents the best-performing model.
Name of the
Network Model
Accuracy
(A)
Precession
(P)
Recall
(R)
Specificity
(S)
F1 Score
(Fs)
MSTDMSTDMSTDMSTDMSTD
Self-ONN4L1DN0.93900.01290.93480.01330.93770.01300.94340.01250.93500.0133
Self-ONN4L0.93760.01300.93560.01320.93760.01300.94280.01250.93750.0131
Self-ONN6L0.94190.01260.94850.01190.94290.01250.95470.01120.94580.0122
Self-ONN6L1DN0.95500.01120.95530.01110.95200.01150.96110.01040.95310.0114
Vanilla CNN8L0.92950.01380.92890.01390.92920.01380.93950.01290.92780.0140
Vanilla CNN6L0.92830.01390.92760.01400.92900.01390.93870.01290.92450.0143
DenseNet2010.94580.01220.94550.01220.94280.01250.95840.01080.94800.0120
ResNet500.95890.01070.95940.01060.95290.01140.96810.00950.95160.0116
ResNet1010.95900.01070.95960.01060.95890.01070.95890.01070.95860.0107
MBINet0.96970.00920.96930.00930.96850.00940.97950.00760.96830.0095
Table 6. Comparison results of the proposed model with the existing models. Bold represents the best-performing model.
Table 6. Comparison results of the proposed model with the existing models. Bold represents the best-performing model.
Ref.YearName of the
Existing Models
Acc.
(%)
Prc.
(%)
Rec.
(%)
Spec.
(%)
Fs
(%)
OCP
(%)
[33]2022FT-DenseNet20194.5894.5594.2895.8494.8094.81
[34]2020PT-DenseNet20194.6694.6294.7694.6893.9594.53
[35]2019DP-DenseNet93.1093.8894.0994.8794.9894.18
[37]2021Differential-DCNN95.8095.7395.6295.6195.1195.57
[38]2021CMP-CNN91.8791.8591.8292.9591.7292.04
[39]2021ResNet-5095.8995.9495.2995.7195.1695.59
[40]2022Fine-tuned ResNet10195.9095.9695.8995.8995.8695.82
Proposed2023MBINet96.9796.9396.8597.9596.8397.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossain, A.; Islam, M.T.; Abdul Rahim, S.K.; Rahman, M.A.; Rahman, T.; Arshad, H.; Khandakar, A.; Ayari, M.A.; Chowdhury, M.E.H. A Lightweight Deep Learning Based Microwave Brain Image Network Model for Brain Tumor Classification Using Reconstructed Microwave Brain (RMB) Images. Biosensors 2023, 13, 238. https://doi.org/10.3390/bios13020238

AMA Style

Hossain A, Islam MT, Abdul Rahim SK, Rahman MA, Rahman T, Arshad H, Khandakar A, Ayari MA, Chowdhury MEH. A Lightweight Deep Learning Based Microwave Brain Image Network Model for Brain Tumor Classification Using Reconstructed Microwave Brain (RMB) Images. Biosensors. 2023; 13(2):238. https://doi.org/10.3390/bios13020238

Chicago/Turabian Style

Hossain, Amran, Mohammad Tariqul Islam, Sharul Kamal Abdul Rahim, Md Atiqur Rahman, Tawsifur Rahman, Haslina Arshad, Amit Khandakar, Mohamed Arslane Ayari, and Muhammad E. H. Chowdhury. 2023. "A Lightweight Deep Learning Based Microwave Brain Image Network Model for Brain Tumor Classification Using Reconstructed Microwave Brain (RMB) Images" Biosensors 13, no. 2: 238. https://doi.org/10.3390/bios13020238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop