Next Article in Journal
Non-Ventilated Patients with Spontaneous Pneumothorax or Pneumomediastinum Associated with COVID-19: Three-Year Debriefing across Five Pandemic Waves
Previous Article in Journal
Evaluation of the Loss of Strength, Resistance, and Elasticity in the Different Types of Intraoral Orthodontic Elastics (IOE): A Systematic Review of the Literature of In Vitro Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimized Convolutional Fusion for Multimodal Neuroimaging in Alzheimer’s Disease Diagnosis: Enhancing Data Integration and Feature Extraction

by
Modupe Odusami
1,
Rytis Maskeliūnas
1 and
Robertas Damaševičius
2,*
1
Department of Multimedia Engineering, Kaunas University of Technology, 51423 Kaunas, Lithuania
2
Department of Applied Informatics, Vytautas Magnus University, 53361 Kaunas, Lithuania
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2023, 13(10), 1496; https://doi.org/10.3390/jpm13101496
Submission received: 16 September 2023 / Revised: 28 September 2023 / Accepted: 12 October 2023 / Published: 14 October 2023

Abstract

:
Multimodal neuroimaging has gained traction in Alzheimer’s Disease (AD) diagnosis by integrating information from multiple imaging modalities to enhance classification accuracy. However, effectively handling heterogeneous data sources and overcoming the challenges posed by multiscale transform methods remains a significant hurdle. This article proposes a novel approach to address these challenges. To harness the power of diverse neuroimaging data, we employ a strategy that leverages optimized convolution techniques. These optimizations include varying kernel sizes and the incorporation of instance normalization, both of which play crucial roles in feature extraction from magnetic resonance imaging (MRI) and positron emission tomography (PET) images. Specifically, varying kernel sizes allow us to adapt the receptive field to different image characteristics, enhancing the model’s ability to capture relevant information. Furthermore, we employ transposed convolution, which increases spatial resolution of feature maps, and it is optimized with varying kernel sizes and instance normalization. This heightened resolution facilitates the alignment and integration of data from disparate MRI and PET data. The use of larger kernels and strides in transposed convolution expands the receptive field, enabling the model to capture essential cross-modal relationships. Instance normalization, applied to each modality during the fusion process, mitigates potential biases stemming from differences in intensity, contrast, or scale between modalities. This enhancement contributes to improved model performance by reducing complexity and ensuring robust fusion. The performance of the proposed fusion method is assessed on three distinct neuroimaging datasets, which include: Alzheimer’s Disease Neuroimaging Initiative (ADNI), consisting of 50 participants each at various stages of AD for both MRI and PET (Cognitive Normal, AD, and Early Mild Cognitive); Open Access Series of Imaging Studies (OASIS), consisting of 50 participants each at various stages of AD for both MRI and PET (Cognitive Normal, Mild Dementia, Very Mild Dementia); and whole-brain atlas neuroimaging (AANLIB) (consisting of 50 participants each at various stages of AD for both MRI and PET (Cognitive Normal, AD). To evaluate the quality of the fused images generated via our method, we employ a comprehensive set of evaluation metrics, including Structural Similarity Index Measurement (SSIM), which assesses the structural similarity between two images; Peak Signal-to-Noise Ratio (PSNR), which measures how closely the generated image resembles the ground truth; Entropy (E), which assesses the amount of information preserved or lost during fusion; the Feature Similarity Indexing Method (FSIM), which assesses the structural and feature similarities between two images; and Edge-Based Similarity (EBS), which measures the similarity of edges between the fused and ground truth images. The obtained fused image is further evaluated using a Mobile Vision Transformer. In the classification of AD vs. Cognitive Normal, the model achieved an accuracy of 99.00%, specificity of 99.00%, and sensitivity of 98.44% on the AANLIB dataset.

1. Introduction

Alzheimer’s Disease (AD) is a neurodegenerative disorder characterized by progressive cognitive decline and memory impairment. Early and accurate diagnosis of AD is crucial for effective intervention and treatment planning. Patients with AD prolong to dementia and lose physiological functions, eventually leading to death. An estimated 55 million people worldwide have dementia, and more than 60% of them live in low- and middle-income countries [1]. It is anticipated that this will increase to 78 million by 2030 and rise to 139 million by 2050 [1]. Metabolic changes in the brain and significant atrophy contribute to the neurodegenerative processes observed in AD. Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET) have become very useful for studying the structural and functional changes linked to Alzheimer’s Disease in the neuroimaging field [2].
The pathogenic nature of Alzheimer’s disease manifests in the brain as structural alterations, including anatomical location, cortical thickness, volumetry, and other morphological features [3]. The capacity to quantify these morphological traits using MRI has resulted in an explosion of methodological research in predicting and categorizing AD. MR images show exceptional anatomical information, and hippocampus shrinkage evaluated on a high-resolution T1-weight MRI is an important criterion for the clinical diagnosis of Alzheimer’s disease [4]. For example, volumetric features of sMRI data were utilized for the classification of Early Mild Cognitive Impairment (EMCI) vs. Normal Cognitive (NC) [5]. sMRI cortical thickness and its underlying geometric information were employed for the early detection of AD [6]. The most useful spatial features of GM were extracted from sMRI and further segmented into ninety regions for Late Mild Cognitive Impairment (LMCI) vs. EMCI classification [7]. Extracted gray matter (GM) images from sMRI using CNN architecture were used for the diagnosis and classification of the CN, EMCI, and LMCI groups [8]. While structural imaging captures downstream pathological changes, it is not appropriate for reflecting changes that precede protein deposition [9]. PET imaging with 18F-fluorodeoxyglucose (FDGPET) imaging can capture brain metabolism characteristics to aid in the detection of lesions for AD classification [10]. For example, FDGPET data was used for the automated classification of AD groups [11]. The risk of AD was predicted based on the deep learning model by extracting FDG PET image features [12]. The fusion of various imaging modalities in multimodal neuroimaging holds the promise of offering comprehensive insights into the metabolic and structural changes occurring in AD [9,13,14].
In recent years, there has been a growing interest in leveraging multimodal neuroimaging data to enhance the accuracy of AD classification. Combining information from multiple imaging modalities can provide a more comprehensive understanding of AD [7,15,16,17] in capturing complementary aspects of brain alterations that may not be evident in a single modality. However, effectively integrating these heterogeneous data sources presents a considerable challenge.
A rising number of studies have used MRI and PET data to discover multilevel and multimodal properties by translating regional brain images into higher-level, more compact characteristics. For example, to classify AD, a new composite image was generated by blending the Gray Matter (GM) tissue region of the brain in both MRI and FDG-PET images using mask and registration coding techniques [18]. Likewise, researchers have classified individuals with AD according to their GM density and glucose utilization from MRI and PET, allowing for a more comprehensive and accurate diagnosis of AD [19]. Furthermore, a novel multimodal image-fusion technique designed to merge PET and MRI data was introduced, in which the extracted features are subsequently input into an ensemble classifier [17]. While the automatic pipeline method described in their study utilized techniques such as Free Surfer and affine registration for pixel-level fusion, achieving precise alignment and ensuring that the combined information accurately reflects the underlying neurobiological changes was a challenge. Furthermore, the extraction of relevant features from these fused modalities introduced complexities in terms of feature selection and interpretability. Although the study employed a range of techniques, including ANOVA, scalar methods, and machine learning classifiers, to identify prominent features, the process of discerning which specific features contribute most significantly to the accurate classification of AD stages remained challenging. The three-channel phase feature learning model demonstrated promise in integrating and learning latent representations from multimodal neuroimaging data, even in the presence of data heterogeneity [10]. However, the successful partial resolution of the heterogeneity issue highlighted the complexity of reconciling distinct characteristics of PET and MRI data within the unified framework.
The multimodality latent space-inducing ensemble Support Vector Machine classifier demonstrated the potential for improved AD classification accuracy. The study’s exploration of an ensemble SVM classifier that induced a latent space via multimodality MRI and PET inputs gave a promising avenue for enhancing AD classification accuracy. However, the intricacy lay in the requirement of effectively reconciling and encapsulating the inherent interrelationships present in MRI and PET modalities within this latent space [20]. An image fusion method to combine MRI and PET images into a composite Gray Matter–PET modality for AD diagnosis has been proposed [21]. While the image fusion method demonstrated superior overall performance compared to unimodal methods, it was noted that its performance in terms of sensitivity and specificity sometimes fell short of optimal levels. The fusion process might introduce subtle distortions or uncertainties that affect diagnostic accuracy.
Additionally, researchers have explored the use of multiscale transform approaches to integrate information from multiple modalities consisting of MRI and PET imaging data in the field of multimodal neuroimaging for the diagnosis of AD. Wavelet-based fusion was used to bring together information from the MRI and PET scans to improve spatial resolution and yielded an image with metabolic and anatomical detail coupled with the finest resolution. Despite the coregistration and alignment process, variability in image resolution and information content between MRI and PET was still a challenge [22]. Ensuring accurate registration and fusion requires overcoming differences in image resolution, contrast, and acquisition protocols [23]. The computational complexity of processing and analyzing data across multiple dimensions is a further disadvantage of multiscale transform methods. These methods frequently involve intricate mathematical algorithms and intensive computational operations, making them time-consuming and resource-intensive [24]. In contrast to multilevel feature learning and multiscale transform approaches for multimodal neuroimaging fusion, our proposed method leverages pre-trained convolution neural networks. Consequently, it requires neither a specific dataset for training nor a specialized network. The proposed network is fed with source images, and optimized feature maps are extracted at layer 1 [25]; Maximum Fusion (MF) is used to fuse the extracted feature. Vision transformers have garnered significant attention in the field of computer vision due to their remarkable performance in image classification tasks, demonstrating their ability to capture long-range dependencies within images [26]. Simple ViT models trained faster and better than the original [27]; the performance of ViTs saturated faster when scaled to be deeper and improved image classification accuracy [28], and Mobile ViT allowed for light-weight global processing of information with transformers [29]. Leveraging the success of vision transformers, we employ fused images as inputs to train a vision transformer model for AD classification.
The research paper’s primary contributions are as follows: (1) An MF strategy is designed to fuse the same depth of feature maps of MRI and PET. (2) An optimized convolution technique including variations in kernel size to increase receptive field and the addition of instance normalization, which ensured no bias towards the modality, is designed to improve alignment and integration of MRI and PET data. (3) A novel fusion model of MRI and PET images using MS strategy and optimized convolution network is proposed, which overcomes the shortcomings of multiscale transform fusion methods and enhances the receptive field of the network.

2. Materials and Methods

The Laplacian transform, a mathematical technique for accentuating intricate image details and extracting edge and texture characteristics [30,31], emerges as a pivotal tool in the fusion of MRI and PET images. In this study, MRI and FDG-PET images are collected from three databases (ADNI, OASIS, and AANLIB), with the Laplacian transform demonstrating its significance in efficiently revealing finer aspects of the images, thereby contributing to the improved fusion of MRI and PET modalities. This technique’s potential is well-founded in its ability to capture and highlight subtle features within medical images, aligning with the demands of contemporary image fusion methodologies [32]. This paper uses Laplace sharpening to obtain the fine details of the image. Transposition convolution was optimized by varying the kernel size, and instance normalization is used to extract relevant feature maps from MRI and PET images. The feature maps obtained via MRI and PET are fused using MF to emphasize the strongest activations between the MRI and PET modalities. The proposed fusion model is shown in Figure 1.

2.1. Datasets

This study utilized MRI and PET images obtained from the official website of Harvard University (http://www.med.harvard.edu/AANLIB/home.html (accessed on 15 September 2023)), the ADNI website (https://adni.loni.usc.edu (accessed on 15 September 2023)), and the OASIS website. The brain images under consideration are categorized into two distinct stages, namely Cognitive Normal (CN) and Alzheimer’s Disease (AD). 50 images of each of the stages are downloaded from each website, making a total of 300 images. PET images are in red, green, and blue (RGB) while MRI images are in black and white. Figure 2, Figure 3 and Figure 4 show sample the datasets used from AANLIB, ADNI, and OASIS database, respectively.

2.2. Laplace Sharpening

Fine details are obtained from the source image S n ( n = 1 , which means MRI image; n = 2 , which means PET image). The S n with pixel intensities is represented as S n x , y , c , where ( x , y ) are the coordinates and c is the color channel [30]. A Gaussian blur kernel is applied to the source image S n ( x , y , c ) . This is represented by Equation (1):
K g a u s s i a n = 1 16 1 2 1 2 4 2 1 2 1 ,
The blurred image is then obtained by convolving the source image with the Gaussian blur kernel. This process is expressed in Equation (2): The obtained image (blurred image) after Equation (1) is applied to Si, as is represented in Equation (2).
B l u r r e d i m a g e x , y , c = i = 1 1 j = 1 1 S n x + i , y + j , c . K g a u s s i a n ,
where: i and j iterate over the 3 × 3 neighborhood of the pixel at position ( x , y , c ) .
To obtain the finer details, a Laplacian kernel, given by Equation (3), is utilized.
K l a p l a c i a n = 1 16 1 1 1 1 8 1 1 1 1 a ,
By convolving the Laplacian kernel from Equation (3) with the previously obtained blurred image, we generate the Laplacian image. This process is outlined in Equation (4):
L a p l a c i a n i m a g e x , y , c = i = 1 1 j = 1 1 B i m a g e x + i , y + j , c . K l a p l a c i a n i , j ,
The sharpened image is derived by combining the source image Sn (x,y,c) with the Laplacian-filtered image. This combination is controlled by an enhancement factor ‘k’. The resulting equation is given by:
S h a r p e n e d i m a g e x , y , c = S n   ( x , y , c ) + k .   L a p l a c i a n i m a g e x , y , c ,
Lastly, to ensure that pixel values are within the valid range of [0, 1], the sharpened image is clipped using the following equation:
S h a r p e n e d i m a g e x , y , c = clip ( S h a r p e n e d i m a g e x , y , c , 0,1 ) ,
Sampled sharpened images from the AANLIB database are shown in Figure 5.

2.3. Basic Image Feature Map Extraction Based on Optimized Transposition Convolution

The proposed method leverages the VGG19 network as a backbone for feature extraction due to its exceptional performance in various computer vision tasks. Previous research in the field of image fusion has reported successful results using VGG19 for image fusion tasks. For instance, infrared and visible images were integrated using VGG19 to extract relevant features, and the MF rule was utilized for the final fused image [33,34]. Likewise, transposed convolution was used for upsampling in image resolution to enhance effective feature extraction [35]. In our proposed study, the initial step involves applying transposed convolution after the first convolution layer [36] by varying kernel sizes. Traditional transposed convolution uses fixed-size kernels for upsampling, which might not capture all levels of detail effectively. By varying the kernel size, the convolution operation can adapt to different spatial scales present in the input image [37,38]. This adaptability is particularly useful for retaining fine-grained information while ensuring that larger structures are also captured [39]. Equation (7) gives the transposition operation on the input images.
T r a n s i , j = m = k k n = k k K e r n e l m , n . X i i n p u t i m , j n ,
In Equation (7), X can be either MRI or PET, and Equation (7) effectively captures the process of obtaining the feature maps from both images using the same convolutional operation with K e r n e l ( m , n ) , which is kernel-centered at the position ( m , n ) , Trans represents the output of transposition process, while i and j represent the row and column indices of the output feature map, respectively. The variable k is the half-size of the kernel (kernel radius), indicating the distance from the center pixel which must be considered in the summation, and I n p u t ( i m ,   j n ) refers to the pixel value of the input image at the relative position ( i m ,   j n ) .
After the transpose convolution operation, instance normalization without learnable parameters is applied across the height and width dimensions to the feature maps corresponding to MRI and PET modalities generated at the higher resolution. Instance normalization without learnable parameters is a variant of instance normalization, in which the scaling and shifting factors are not learned but are instead fixed and applied in a predetermined manner [40,41]. In this study, instance normalization helps normalize the activations of individual instances independently, without introducing any learnable parameters. Normalizing each modality ensures that the fusion process is not biased towards the modality due to differences in intensity, contrast, or scale. Also, since this instance of normalization does not require the learning of scaling and shifting parameters, it can lead to a reduction in model complexity and improve the performance of image fusion [42].
The instance normalization is represented as follows:
Step 1: The mean and variance of the feature map Trans (i, j) across the height (H) and width (W) dimensions for each instance (t) are calculated separately. Let us denote them as μ_ti and σ_ti, respectively:
μ t i = 1 H W l = 1 W m = 1 H T r a n s ( i + m , j + n ) a   =   1 ,
σ t i 2 = 1 H W l = 1 W m = 1 H ( T r a n s ( i + m , j + n ) μ t i ) 2 a   =   1 ,
Step 2: The feature map T r a n s ( i ,   j ) is normalized using the computed mean and variance in Equations (8) and (9) as follows:
y i j k = ( T r a n s ( i + m , j + n μ t i ) ρ t i 2 + a   =   1 ,
where y i j k is the normalized output at position ( i ,   j ) and channel ( k ) ,   T r a n s ( i + m ,   j + n ) represents the pixel value of the feature map at the relative position ( i + m ,   j + n ) ,   μ t i and ρ t i 2 are the mean and variance computed in step 1, and ε is a small constant added to the denominator to avoid division by zero. Note: the normalization is applied separately to both the MRI and PET feature maps to ensure that each feature map is normalized across its channel dimensions.

2.4. Basic Image Fusion Strategy Using the MS

After normalizing the feature maps separately for MRI and PET using instance normalization, MS is applied to combine the normalized feature maps. This strategy aims to leverage complementary information from both MRI and PET modalities. The basic idea is to take the maximum value between the corresponding elements of the normalized MRI and PET feature maps, pixel-wise, to create a fused feature map [43].
For each pixel position ( i ,   j ) and channel ( k ) , the maximum value between the normalized MRI and PET feature maps is:
F u s e d f e a t u r e m a p i j k = max ( y m r i i j k , y p e t i j k ) ,
This Maximum Strategy ensures that the fused feature map retains the strongest features from both the MRI and PET modalities, leveraging the strengths of each image while minimizing the impact of less relevant information. The fused feature map is further fed into a Vision Transformer for the classification of AD stages.

3. Experiment and Result Analysis

To verify the effectiveness and advancement of the proposed fusion method in different scenarios, we conducted a rigorous set of evaluations and experiments that leveraged its intrinsic adaptability and generalization capabilities. Notably, our approach capitalizes on the unique ability to fuse information from single data instances, rendering it suitable for scenarios where large training datasets might be unavailable or impractical to create. By focusing on single data instances without requiring prior training, we ensured that the method’s performance was consistently assessed across three different databases. We adopted the same network size in [36] for the fusion process. To effectively assess its performance, we adopted a holistic approach based on quantitative metrics assessments. Quantitative analysis involves calculating relevant performance metrics, such as PSNR, SSIM, FSIM, E, and EBS, among others [44,45]. These metrics allowed us to quantitatively measure the fidelity, preservation of salient features, and alignment with ground truth information. The PSNR values indicate the quality of the fused images in terms of noise and distortion, where higher values suggest better quality. The SSIM values indicate the structural similarity between the original and fused images, with values closer to 1 indicating better similarity. The metrics E, FSIM, and EBS provide insights into various aspects of image quality, such as edge preservation and structural information. These metrics assess the performance of the fusion strategy in different dimensions. At the same time, the proposed algorithm will be compared with the results of the other two typical fusion methods on the dataset from the selected databases. Sample fused images are displayed in Figure 6.
All experiments were conducted using the Python programming language and executed on a GPU-accelerated system. Table 1, Table 2 and Table 3 depict the quantitative evaluation values of the fusion images corresponding to twelve typical source images from AD stages in the selected databases. The overall evaluation metrics evaluate the SSIM, FSIM, E, EBS, and PSNR of the proposed fusion model, and multiscale transform methods are depicted in Figure 7, Figure 8 and Figure 9. To assess the generalization capability and performance of the fused data on individual modalities, MViT is trained and validated using the fused data and then tested with unseen fused data from ADNI and AANLIB. The MViT model is trained on the fused data from ADNI and AANLIB separately, i.e., 150 fused data from ADNI and 100 fused data from AANLIB, as shown in Table 4. This study leveraged the hyperparameters in [29] and finetuned them for the proposed model AD stage classification. Hyperparameters used have a learning rate of 0.0002 and a weight decay of 0.01. The stochastic gradient descent optimizer performed better than the adaptive optimizer with weight decay in our study.
A set of image transformations was applied to the fused images to augment the dataset to minimize overfitting. The transformations were implemented in the Pytorch transforms module to provide a range of image manipulation. The training and validation accuracy/loss is depicted in Figure 10, Figure 11 and Figure 12. The Confusion matrix is used to provide insight into the model’s classification accuracy and potential misclassifications across different classes. Figure 13 depicts the confusion matrix of the proposed model. Simultaneously, the ROC curve is depicted in Figure 14, which illustrates the trade-off between a true positive rate and a false positive rate, aiding in the determination of a suitable decision threshold.

4. Discussion

From Table 1, it can be observed that the proposed fusion strategy generally performs better compared with DWT and LPG in terms of most metrics across both datasets. This indicates that the proposed fusion method leveraging the VGG19 model tends to preserve more image details, maintain better structural similarity, and produce higher-quality fused images. Furthermore, the results demonstrate that the proposed fusion strategy exhibits effectiveness and advancement across AD and CN without the need for explicit model training. Likewise, the results in Table 2 also show consistency in the performance trend across different stages of AD from ADNI. This suggests the robustness and general applicability of the proposed fusion method across diverse datasets, showcasing its effectiveness for various image fusion scenarios.
Figure 7, Figure 8 and Figure 9 show the proposed model achieves a PSNR of 32.55, a SIMM of 0.83, an E of 4.13, an FSIM of 0.96, and an EBS of 0.75 on the AANLIB dataset. Comparing these metrics with the other fusion models, the proposed model outperforms DWT and GLP across most metrics, indicating superior image quality and feature preservation. On the ADNI dataset, the proposed model achieves a PSNR of 30.39, a SIMM of 0.70, an E of 6.78, an FSIM of 0.93, and an EBS of 0.67. Again, the proposed model generally performs better than the other methods, showcasing its potential for accurately fusing MRI and PET data. The proposed model scores a PSNR of 28.58, a SIMM of 0.67, an E of 6.83, an FSIM of 0.85, and an EBS of 0.71 on the OASIS dataset. While still performing favorably, the differences in metrics between the models are less pronounced on this dataset. Across all three datasets, the proposed model consistently outperforms both DWT and GLP. This implies that the proposed fusion method captures more meaningful information from both MRI and PET modalities, resulting in images that are better suited for subsequent analysis or diagnosis. While DWT and GLP might have their advantages, the results indicate that the proposed model is more effective for the specific task of fusing MRI and PET data for AD vs. CN classification.
The clinical implication of these results lies in the ability of the image fusion model to aid in the classification of individuals with AD vs. those who are CN. Generally, higher performance metrics (like higher PSNR, SIMM, and FSIM values and lower E and EBS values) indicate better image fusion quality. This better fusion quality can potentially enhance the ability to detect patterns and features that are crucial for accurate classification. Figure 12 shows that the performance of the fused image in the classification of AD vs. CN using MRI test data from AANLIB provides a precision of 100% (AD) and 98% (CN), with a recall of 97% (AD), and 100% (CN), and an F1-score of 99% (AD), with 99% (CN) indicating that the model performs very well on the AANLIB MRI test dataset. It has high precision, recall, and F1-score values for both classes, which suggests that the model is effective in correctly classifying instances from both classes. The model performance using the AANLIB PET test dataset is like the AANLIB MRI dataset. The model’s performance on the ADNI MRITEST dataset is good but not as high as the AANLIB datasets. The performance of the fused image in the classification of AD vs. CN using MRI test data from ADNI provides precision values of 91% (AD) and 100% (CN), recalls of 100% (AD) and 93% (CN), and F1-scores of 95% (AD) and 96% (CN). The model’s performance on the ADNI PET test dataset is like that of the ADNI MRI test dataset. It achieves high precision and recall for both classes, but the recall for class AD is comparatively low (88%), affecting the overall F1-score for that class.
However, there are some variations in performance across datasets. The model performs slightly better on the AANLIB datasets compared with the ADNI datasets. Additionally, the performance metrics for class AD are relatively low compared with those for class CN in some cases, suggesting that the model might struggle more with the AD class. Figure 14 shows that the AUC values, which measure the model’s ability to discriminate between positive and negative classes, are generally high. The AANLIB MRI and AANLIB PET datasets have the highest AUC values, suggesting excellent discriminative power. The ADNI MRITEST dataset has perfect AUC (100%) but slightly lower accuracy compared with the AANLIB datasets. Likewise, the OASIS MRI Test data have perfect AUC (100%). The ADNI PET dataset has a relatively low AUC (90%) compared with the other datasets, which might indicate that the model has slightly more difficulty distinguishing between classes. The high recall values indicate that the model is effective at identifying most samples of the positive class, which is crucial in medical diagnosis in order to avoid false negatives. The high AUC values imply that the model can effectively distinguish between different classes, which is vital for reliable diagnostic decisions. However, while the model’s performance is promising, in clinical decisions, the model should be used as a supportive tool for medical professionals to aid in diagnosis and decision making. A summary of the proposed model performance is depicted in Table 5.
The results achieved in this article are compared and validated with the recent research conducted on AD detection using multimodal neuroimaging in Table 5.
Table 6 presents a comparison between the proposed method and several existing methods for binary classification tasks in AD vs. CN classification using the ADNI database. The proposed method demonstrates competitive performance compared with existing methods. While some methods achieve higher accuracy, the proposed method maintains a good balance between specificity and sensitivity, which is crucial in medical diagnosis scenarios. Additionally, the proposed method’s performance on the AANLIB dataset suggests its potential for generalizability across different datasets.

5. Conclusions

In this research study, we proposed a novel fusion method for multimodal neuroimaging data consisting of MRI and PET images to enhance the accuracy of Alzheimer’s Disease classification. The method leveraged a Maximum Fusion strategy and an optimized convolution network, effectively combining the complementary information from both modalities. The fusion approach demonstrated its effectiveness across multiple datasets and AD stages without the need for explicit model training. Our comprehensive experiments and result analysis showcased the superior performance of the proposed fusion method compared with traditional fusion methods such as Discrete Wavelet Transform and Laplacian Pyramid Gaussian in terms of various quantitative metrics. The fused images were used to train the Mobile Vision Transformer for the classification of Alzheimer’s Disease vs. Cognitive Normal classification, and the proposed model’s classification accuracy, precision, recall, F1-score, and AUC values demonstrated its effectiveness.
Comparisons with existing methods highlighted the competitive nature of our proposed approach, maintaining a good balance between the specificity and sensitivity crucial for medical diagnosis scenarios. The fusion model’s clinical implication lies in its potential to aid medical professionals in the accurate classification of AD vs. CN patients. One major limitation is the absence of real clinical data for validation, which could impact the model’s performance when applied to real-world scenarios. Additionally, the proposed model’s performance might be influenced by variations in data acquisition protocols and scanner characteristics in clinical settings. Future research could focus on validating the method using diverse clinical datasets and addressing the limitations observed in this study.

Author Contributions

Conceptualization, R.M.; methodology, R.M.; software, M.O.; validation, M.O., R.M. and R.D.; formal analysis, M.O., R.M. and R.D.; investigation, M.O., R.M. and R.D.; resources, R.M.; data curation, M.O.; writing—original draft preparation, M.O. and R.M.; writing—review and editing, R.D.; visualization, M.O.; supervision, R.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arvanitakis, Z.; Shah, R.C.; Bennett, D.A. Diagnosis and Management of Dementia: A Review. JAMA 2019, 322, 1589–1599. [Google Scholar] [CrossRef] [PubMed]
  2. Young, P.N.E.; Estarellas, M.; Coomans, E.; Srikrishna, M.; Beaumont, H.; Maass, A.; Venkataraman, A.V.; Lissaman, R.; Jiménez, D.; Betts, M.J.; et al. Imaging biomarkers in neurodegeneration: Current and future practices. Alzheimer’s Res. Ther. 2020, 12, 49. [Google Scholar] [CrossRef] [PubMed]
  3. Márquez, F.; Yassa, M.A. Neuroimaging Biomarkers for Alzheimer’s Disease. Mol. Neurodegener. 2019, 14, 21. [Google Scholar] [CrossRef] [PubMed]
  4. Ledig, C.; Schuh, A.; Guerrero, R.; Heckemann, R.A.; Rueckert, D. Structural brain imaging in Alzheimer’s disease and mild cognitive impairment: Biomarker analysis and shared morphometry database. Sci. Rep. 2018, 8, 11258. [Google Scholar] [CrossRef] [PubMed]
  5. Jiang, J.; Kang, L.; Huang, J.; Zhang, T. Deep learning based mild cognitive impairment diagnosis using structure MR images. Neurosci. Lett. 2020, 730, 134971. [Google Scholar] [CrossRef] [PubMed]
  6. Wee, C.-Y.; Liu, C.; Lee, A.; Poh, J.S.; Ji, H.; Qiu, A. Cortical graph neural network for AD and MCI diagnosis and transfer learning across populations. NeuroImage Clin. 2019, 23, 101929. [Google Scholar] [CrossRef] [PubMed]
  7. Yue, L.; Gong, X.; Chen, K.; Mao, M.; Li, J.; Nandi, A.K.; Li, M. Auto-Detection of Alzheimer’s Disease Using Deep Convolutional Neural Networks. In Proceedings of the 2018 14th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Huangshan, China, 28–30 July 2018; pp. 228–234. [Google Scholar] [CrossRef]
  8. Gorji, H.T.; Kaabouch, N. A Deep Learning approach for Diagnosis of Mild Cognitive Impairment Based on MRI Images. Brain Sci. 2019, 9, 217. [Google Scholar] [CrossRef] [PubMed]
  9. Kim, J.; Jeong, M.; Stiles, W.R.; Choi, H.S. Neuroimaging Modalities in Alzheimer’s Disease: Diagnosis and Clinical Features. Int. J. Mol. Sci. 2022, 23, 6079. [Google Scholar] [CrossRef] [PubMed]
  10. Odusami, M.; Maskeliūnas, R.; Damaševičius, R.; Misra, S. Explainable Deep-Learning-Based Diagnosis of Alzheimer’s Disease Using Multimodal Input Fusion of PET and MRI Images. J. Med. Biol. Eng. 2023, 43, 291–302. [Google Scholar] [CrossRef]
  11. Ozsahin, I.; Sekeroglu, B.; Mok, G.S.P. The use of back propagation neural networks and 18F-Florbetapir PET for early detection of Alzheimer’s disease using Alzheimer’s Disease Neuroimaging Initiative database. PLoS ONE 2019, 14, e0226577. [Google Scholar] [CrossRef] [PubMed]
  12. Yang, Z.; Liu, Z. The risk prediction of Alzheimer’s disease based on the deep learning model of brain 18F-FDG positron emission tomography. Saudi J. Biol. Sci. 2020, 27, 659–665. [Google Scholar] [CrossRef] [PubMed]
  13. Uludağ, K.; Roebroeck, A. General overview on the merits of multimodal neuroimaging data fusion. NeuroImage 2014, 102, 3–10. [Google Scholar] [CrossRef] [PubMed]
  14. Loftus, J.R.; Puri, S.; Meyers, S.P. Multimodality imaging of neurodegenerative disorders with a focus on multiparametric magnetic resonance and molecular imaging. Insights Into Imaging 2023, 14, 8. [Google Scholar] [CrossRef] [PubMed]
  15. Meng, X.; Liu, J.; Fan, X.; Bian, C.; Wei, Q.; Wang, Z.; Liu, W.; Jiao, Z. Multi-Modal Neuroimaging Neural Network-Based Feature Detection for Diagnosis of Alzheimer’s Disease. Front. Aging Neurosci. 2022, 14, 911220. Available online: https://www.frontiersin.org/articles/10.3389/fnagi.2022.911220 (accessed on 15 October 2022). [CrossRef] [PubMed]
  16. Ismail, W.N.; Rajeena, P.F.; Ali, M.A.S. MULTforAD: Multimodal MRI Neuroimaging for Alz-heimer’s Disease Detection Based on a 3D Convolution Model. Electronics 2022, 11, 3893. [Google Scholar] [CrossRef]
  17. Shukla, A.; Tiwari, R.; Tiwari, S. Alzheimer’s Disease Detection from Fused PET and MRI Modalities Using an Ensemble Classifier. Mach. Learn. Knowl. Extr. 2023, 5, 512–538. [Google Scholar] [CrossRef]
  18. Das, A.; Bhattacharya, M. Effective image fusion method to study Alzheimer’s disease using MR, PET images. In Proceedings of the 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Washington, DC, USA; 2015; pp. 1603–1607. [Google Scholar] [CrossRef]
  19. Shao, W.; Peng, Y.; Zu, C.; Wang, M.; Zhang, D.; Alzheimer’s Disease Neuroimaging Initiative. Hypergraph based multi-task feature selection for multimodal classification of Alzheimer’s disease. Comput. Med. Imaging Graph. 2020, 80, 101663. [Google Scholar] [CrossRef]
  20. Zhou, T.; Thung, K.-H.; Liu, M.; Shi, F.; Zhang, C.; Shen, D. Multi-modal latent space inducing ensemble SVM classifier for early dementia diagnosis with neuroimaging data. Med. Image Anal. 2020, 60, 101630. [Google Scholar] [CrossRef] [PubMed]
  21. Song, J.; Zheng, J.; Li, P.; Lu, X.; Zhu, G.; Shen, P. An Effective Multimodal Image Fusion Method Using MRI and PET for Alzheimer’s Disease Diagnosis. Front. Digit. Health 2021, 3, 637386. Available online: https://www.frontiersin.org/article/10.3389/fdgth.2021.637386 (accessed on 14 March 2022). [CrossRef]
  22. Dwivedi, S.; Goel, T.; Tanveer, M.; Murugan, R.; Sharma, R. Multimodal Fusion-Based Deep Learning Network for Effective Diagnosis of Alzheimer’s Disease. IEEE Multimed. 2022, 29, 45–55. [Google Scholar] [CrossRef]
  23. Odusami, M.; Maskeliūnas, R.; Damaševičius, R. Pixel-Level Fusion Approach with Vision Transformer for Early Detection of Alzheimer’s Disease. Electronics 2023, 12, 1218. [Google Scholar] [CrossRef]
  24. He, B. Research on Multi-modal Medical Image Fusion Based on Multi-Scale Attention Mechanism. J. Phys. Conf. Ser. 2022, 2173, 012037. [Google Scholar] [CrossRef]
  25. Odusami, M.; Maskeliūnas, R.; Damaševičius, R. An Intelligent System for Early Recognition of Alzheimer’s Disease Using Neuroimaging. Sensors 2022, 22, 740. [Google Scholar] [CrossRef]
  26. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in vision: A survey. ACM Comput. Surv. (CSUR) 2022, 54, 1–41. [Google Scholar] [CrossRef]
  27. Beyer, L.; Zhai, X.; Kolesnikov, A. Better plain ViT baselines for ImageNet-1k. arXiv 2022, arXiv:2205.01580. [Google Scholar]
  28. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. Deepvit: Towards deeper vision transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  29. Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
  30. Pham, T.D. Kriging-Weighted Laplacian Kernels for Grayscale Image Sharpening. IEEE Access 2022, 10, 57094–57106. [Google Scholar] [CrossRef]
  31. Iwanowski, M. Image Contrast Enhancement Based on Laplacian-of-Gaussian Filter Combined with Morphological Reconstruction, in Progress in Computer Recognition Systems. In Advances in Intelligent Systems and Computing; Burduk, R., Kurzynski, M., Wozniak, M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 305–315. [Google Scholar] [CrossRef]
  32. Shan, S.; Rossi, E.A.; Snyder, V.; Chhablani, J.; Lejoyeux, R.; Sahel, J.-A.; Dansingani, K.K. Laplacian feature detection and feature alignment for multimodal ophthalmic image registration using phase correlation and Hessian affine feature space. Signal Process. 2020, 177, 107733. [Google Scholar]
  33. Zhou, J.; Ren, K.; Wan, M.; Cheng, B.; Gu, G.; Chen, Q. An infrared and visible image fusion method based on VGG-19 network. Optik 2021, 248, 168084. [Google Scholar] [CrossRef]
  34. Lu, Y.; Qiu, Y.; Gao, Q.; Sun, D. Infrared and visible image fusion based on tight frame learning via VGG19 network. Digit. Signal Process. 2022, 131, 103745. [Google Scholar] [CrossRef]
  35. Sahito, F.; Zhiwen, P.; Sahito, F.; Ahmed, J. Transpose convolution based model for super-resolution image reconstruction. Appl. Intell. 2023, 53, 10574–10584. [Google Scholar] [CrossRef]
  36. Odusami, M.; Maskeliūnas, R.; Damaševičius, R. Pareto Optimized Adaptive Learning with Transposed Convolution for Image Fusion Alzheimer’s Disease Classification. Brain Sci. 2023, 13, 1045. [Google Scholar] [CrossRef] [PubMed]
  37. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. Available online: https://openaccess.thecvf.com/content_CVPR_2019/html/Li_Selective_Kernel_Networks_CVPR_2019_paper.html (accessed on 15 August 2023).
  38. Nevavuori, P.; Narra, N.; Lipping, T. Crop yield prediction with deep convolutional neural networks. Comput. Electron. Agric. 2019, 163, 104859. [Google Scholar] [CrossRef]
  39. Chi, L.; Jiang, B.; Mu, Y. Fast Fourier Convolution. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2020; pp. 4479–4488. Available online: https://proceedings.neurips.cc/paper/2020/hash/2fd5d41ec6cfab47e32164d5624269b1-Abstract.html (accessed on 15 August 2023).
  40. Zhu, Q.; Bai, H.; Sun, J.; Cheng, C.; Li, X. LPAdaIN: Light Progressive Attention Adaptive Instance Normalization Model for Style Transfer. Electronics 2022, 11, 2929. [Google Scholar] [CrossRef]
  41. Chen, L.; Lu, X.; Zhang, J.; Chu, X.; Chen, C. Hinet: Half instance normalization network for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 182–192. [Google Scholar]
  42. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance normalization: The missing ingredient for fast stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  43. Cheng, C.; Wu, X.J.; Xu, T.; Chen, G. Unifusion: A lightweight unified image fusion network. IEEE Trans. Instrum. Meas. 2021, 70, 5016614. [Google Scholar] [CrossRef]
  44. Singh, V. Image Fusion Metrics—A Systematic Study. In Proceedings of the 2nd International Conference on IoT, Social, Mobile, Analytics & Cloud in Computational Vision & Bio-Engineering (ISMAC-CVB 2020), Rochester, NY, USA, 21 November 2020; 2020. [Google Scholar] [CrossRef]
  45. Hou, S.; Fu, M.; Song, W. Joint Learning of Image Deblurring and Depth Estimation Through Adversarial Multitask Network. IEEE Trans. Circuits Syst. Video Technol. 2023, 9, 607–619. [Google Scholar]
Figure 1. Framework of Proposed Fusion Method.
Figure 1. Framework of Proposed Fusion Method.
Jpm 13 01496 g001
Figure 2. Sample Images from the AANLIB database.
Figure 2. Sample Images from the AANLIB database.
Jpm 13 01496 g002
Figure 3. Sample Images from the ADNI database.
Figure 3. Sample Images from the ADNI database.
Jpm 13 01496 g003
Figure 4. Sample Images from OASIS Database.
Figure 4. Sample Images from OASIS Database.
Jpm 13 01496 g004
Figure 5. Sample Images after Laplace Sharpening.
Figure 5. Sample Images after Laplace Sharpening.
Jpm 13 01496 g005
Figure 6. Sample of Fused Images.
Figure 6. Sample of Fused Images.
Jpm 13 01496 g006
Figure 7. Average Metrics Value of AANLIB.
Figure 7. Average Metrics Value of AANLIB.
Jpm 13 01496 g007
Figure 8. Average Metrics Value of ADNI.
Figure 8. Average Metrics Value of ADNI.
Jpm 13 01496 g008
Figure 9. Average Metrics Value of OASIS.
Figure 9. Average Metrics Value of OASIS.
Jpm 13 01496 g009
Figure 10. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from AANLIB.
Figure 10. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from AANLIB.
Jpm 13 01496 g010
Figure 11. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from ADNI.
Figure 11. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from ADNI.
Jpm 13 01496 g011
Figure 12. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from OASIS.
Figure 12. Training Accuracy, Validation Accuracy, Training Loss, and Validation Loss using Fused Data from OASIS.
Jpm 13 01496 g012
Figure 13. Confusion Matrix for (a) MRI Test Data from AANLIB, (b) PET Test Data from AANLIB, (c) MRI Test Data from ADNI, (d) PET Test Data from ADNI, (e) MRI Test Data from OASIS, (f) PET Test Data from OASIS.
Figure 13. Confusion Matrix for (a) MRI Test Data from AANLIB, (b) PET Test Data from AANLIB, (c) MRI Test Data from ADNI, (d) PET Test Data from ADNI, (e) MRI Test Data from OASIS, (f) PET Test Data from OASIS.
Jpm 13 01496 g013aJpm 13 01496 g013b
Figure 14. ROC Curve for Proposed Model for Classification of AD vs. CN: (a) AANLIB MRI Test Data, (b) AANLIB PET Test Data, (c) ADNI MRI Test Data, (d) ADNI PET Test Data, (e) OASIS MRI Test Data, (f) OASIS PET Test Data.
Figure 14. ROC Curve for Proposed Model for Classification of AD vs. CN: (a) AANLIB MRI Test Data, (b) AANLIB PET Test Data, (c) ADNI MRI Test Data, (d) ADNI PET Test Data, (e) OASIS MRI Test Data, (f) OASIS PET Test Data.
Jpm 13 01496 g014
Table 1. Quantitative Evaluation Values of Fused Images from AANLIB Database.
Table 1. Quantitative Evaluation Values of Fused Images from AANLIB Database.
ImagePSNRSSIMEFSIMEBS
DWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposed
AD
131.9530.2933.300.660.500.743.985.894.110.830.900.990.730.500.86
231.8030.2631.400.660.500.804.025.844.150.830.900.990.700.470.76
331.8030.2832.900.540.500.903.985.844,120.830.900.900.680.510.79
431.9030.4432.900.540.500.903.925.844.140.830.900.990.670.480.69
531.8030.2832.900.550.500.804.025.894.110.830.800.880.700.510.73
631.8030.4432.500.550.550.833.985.894.140.830.920.990.670.480.72
731.8030.2832.000.580.500.833.985.844.110.830.800.890.680.480.75
831.9030.4432.000.580.500.833.925.844.140.830.900.990.670.510.69
931.8030.2632.900.550.500.873.925.894.140.830.800.990.690.500.79
1031.9530.4432.000.700.550.874.025.894.110.820.900.990.730.500.80
1131.9530.3932.900.720.540.864.025.844.140.830.900.980.710.500.75
1231.9530.3932.900.700.500.833.985.844.140.820.900.990.720.510.75
CN
130.9531.6530.990.520.600.624.744.295.450.930.981.000.630.660.70
230.8831.6531040.520.600.624.744.295.050.930.970.990.640.660.72
330.8430.9531.800.540.660.684.744.405.000.930.970.990.630.660.73
430.9531.0031.800.520.660.684.744.295.080.980.980.990.630.660.73
530.9530.9531.800.550.670.684.744.405.180.980.980.990.630.660.70
631.0830.9531.670.520.680.674.744.295.180.930.980.990.640.660.73
731.0831.0031.670.520.660.674.744.445.240.940.980.990.640.660.73
830.8531.0031.080.520.650.674.744.295.150.940.970.990.630.670.70
930.9031.0531.080.520.660.624.744.295.330.930.981.000.630.670.73
1030.9531.0531.650.520.650.654.754.295.140.940.981.000.630.660.73
1130.9531.0531.750.520.650.654.744.245.000.940.980.990.630.660.70
1230.9531.0531.040.520.600.604.744.305.250.940.970.990.630.660.70
PSNR = Peak Signal-To-Noise Ratio (PSNR), SSIM = Structural Similarity Index Measure (SSIM), E = Entropy, FSIM = Feature Similarity Indexing Method, EBS = Edge-Based Similarity, DWT = Discrete Wave Transform, LPG = Laplacian Gaussian Pyramid, AD = Alzheimer’s Disease, and CN = Cognitive Normal.
Table 2. Quantitative Evaluation Values of Fused Images from the ADNI Database.
Table 2. Quantitative Evaluation Values of Fused Images from the ADNI Database.
ImagePSNRSSIMEFSIMEBS
DWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposed
AD
128.6029.9829.000.680.530.706.505.906.900.800.930.850.710.710.70
228.3228.8329.010.680.600.706.505.906.900.800.970.990.710.730.69
328.3228.8330.000.680.600.706.506.046.700.770.950.900.730.710.69
428.8229.0830.250.680.600.706.506.046.600.770.930.990.720.730.65
528.8229.0930.560.680.640.706.506.056.900.780.930.900.710.750.65
628.8229.1430.560.690.600.706.506.056.900.780.980.900.710.710.69
728.8229.0730.000.690.600.706.506.206.800.780.900.950.710.600.69
828.8229.0630.050.690.620.706.506.206.750.790.900.960.710.710.65
929.1029.0830.980.690.600.706.506.206.750.790.900.950.710.740.69
1029.1029.1430.980.710.600.716.506.206.750.780.900.950.710.600.65
1129.1029.0830.980.710.600.716.506.206.750.790.900.950.710.730.69
1229.1029.0730.980.710.600.716.506.206.750.790.950.900.710.740.69
CN
131.0530.5031.250.750.650.895.306.206.400.830.970.990.720.760.70
229.2429.5331.090.720.600.876.306.906.360.800.960.970.700.730.69
331.0530.7931.900.740.650.834.867.005.000.690.960.990.700.760.77
431.2231.1531.550.780.620.795.086.305.000.700.960.990.760.760.76
529.9429.4332.050.670.640.766.336.306.500.810.960.990.780.760.79
630.4430.3531.050.700.600.845.356.325.530.780.980.970.700.760.69
730.7330.6130.920.700.600.735.806.326.940.810.980.990.700.760.78
831.0630.8730.940.720.610.765.266.415.600.740.980.960.750.760.74
930.5530.6130.570.700.600.785.466.206.660.750.980.980.750.760.68
1030.4930.6131.730.700.600.715.336.305.530.800.960.990.700.760.69
1130.5030.6131.050.700.650.775.256.306.060.750.960.980.700.760.79
1231.6030.6431.900.700.660.815.286.306.900.890.960.980.720.780.73
EMCI
130.9530.4532.050.740.630.885.456.156.600.820.960.980.740.770.71
230.7530.6031.950.730.620.865.406.106.550.810.970.980.730.760.70
331.1530.7532.150.750.640.895.556.206.700.830.970.990.750.780.72
430.8530.6531.750.7290.610.855.506.056.500.800.960.980.720.750.69
531.0530.5032.000.740.630.875.606.256.650.820.970.990.740.770.71
630.9530.7031.900.730.620.885.586.186.630.810.970.980.730.760.70
731.1030.7532.100.750.640.895.636.236.680.830.970.990.750.780.72
830.9030.6031.800.720.610.865.556.086.530.800.960.980.720.750.69
931.0030.5531.950.740.630.885.586.156.600.820.970.990.740.770.71
1031.0530.6032.050.740.630.895.606.206.650.830.970.990.740.770.72
1130.8530.6531.750.720.610.865.536.106.550.800.960.980.720.750.70
1231.2030.7032.200.750.640.895.656.256.700.830.970.990.750.780.72
PSNR = Peak Signal-To-Noise Ratio (PSNR), SSIM = Structural Similarity Index Measure (SSIM), E = Entropy, FSIM = Feature Similarity Indexing Method, EBS = Edge-Based Similarity, DWT = Discrete Wave Transform, LPG = Laplacian Gaussian Pyramid, AD = Alzheimer’s Disease, CN = Cognitive Normal, and EMCI = Early Mild Cognitive Normal.
Table 3. Quantitative Evaluation Values of Fused Images from the OASIS Database.
Table 3. Quantitative Evaluation Values of Fused Images from the OASIS Database.
ImagePSNRSSIMEFSIMEBS
DWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposedDWTLPGProposed
VMD
126.6027.9826.900.580.430.576.004.906.100.700.900.840.590.710.75
226.3226.8326.500.580.500.625.505.005.900.800.870.840.600.730.70
326.3226.8326.00.580.500.575.505.046.720.870.800.870.600.710.70
426.8027.0829.800.580.500.605.505.046.030.770.800.880.520.700.75
526.8227.0929.800.580.540.805.505.056.030.880.900.870.600.750.65
626.8027.1429.700.570.500.605.505.057.900.700.800.750.600.710.79
726.8227.0728.900.590.500.605.505.206.710.880.700.930.600.520.70
826.8227.0628.850.670.520.715.505.207.720.790.800.870.600.710.69
927.1027.0629.150.690.500.806.005.206.710.890.700.870.600.710.76
1027.2027.1029.150.610.500.715.505.207.750.780.800.860.700.540.69
1127.1027.0029.500.610.500.786.005.206.760.890.600.820.600.730.69
1227.1027.0128.800.610.500.706.005.207.710.700.650.820.700.710.68
ND
130.0529.5031.060.650.600.705.005.205.000.800.800.860.700.660.69
228.2428.5330.200.620.580.746.005.905.600.700.800.880.710.630.69
330.0529.7029.900.640.600.794.566.005.400.600.800.850.710.660.70
430.2229.1529.850.700.600.755.005.305.400.600.860.870.750.660.70
528.9028.4030.000.570.600.646.305.306.400.710.700.880.760.660.70
628.0029.3030.000.600.580.785.305.325.600.700.800.860.710.660.70
728.7029.6029.900.690.570.655.205.325.700.710.900.790.710.660.70
830.0029.8030.000.620.600.695.205.415.600.700.780.890.730.660.69
929.0029.6029.500.690.500.665.405.205.700.700.800.860.740.660.69
1029.4029.6029.630.650.580.604.335.305.730.700.860.870.710.660.69
1129.4029.6029.000.650.600.775.205.305.660.700.800.860.740.660.69
1230.4029.5429.800.650.600.785.205.305.700.800.860.790.700.680.75
MD
130.0029.8030.250.600.500.635.205.805.510.700.760.870.600.400.59
230.0028.8030.900.600.500.595.205.806.020.850.860.920.600.400.67
330.0129.8029.500.700.500.565.785.805.050.850.900.750.520.430.70
429.5029.8930.500.610.520.505.005.805.210.700.900.810.500.500.70
529.5029.0030.600.610.520.695.405.055.060.700.850.810.500.500.69
629.5031.7031.200.600.500.685.205.856.020.800.750.810.500.500.68
730.5028.1029.900.650.500.685.206.055.240.700.700.880.500.500.66
830.5030.7029.920.790.500.696.205.065.030.700.850.810.500.500.65
930.0531.7030.450.500.400.614.305.074.060.700.900.920.500.500.66
1030.5029.0028.350.620.500.706.305.895.990.850.850.860.500.500.66
1130.5030.0029.350.630.500.695.305.565.990.710.900.810.500.500.65
1230.5030.0030.250.700.550.693.305.865.070.900.880.750.500.500.65
PSNR = Peak Signal-To-Noise Ratio (PSNR), SSIM = Structural Similarity Index Measure (SSIM), E = Entropy, FSIM = Feature Similarity Indexing Method, EBS = Edge-Based Similarity, DWT = Discrete Wave Transform, LPG = Laplacian Gaussian Pyramid, VMD = Very Mild Dementia, ND = Non Dementia, MD = Mild Dementia.
Table 4. Details of Source Images, Fused Images, and Augmented Images.
Table 4. Details of Source Images, Fused Images, and Augmented Images.
ImageNo of Images
ADNIAANLIBOASIS
ADCNEMCIADCNCNVMDMD
Fused5050505050505050
Augmented10001000100010001000100010001000
Table 5. Summary of Proposed Model Performance.
Table 5. Summary of Proposed Model Performance.
AANLIB MRI Test DataAANLIB PET Test DataADNI MRI Test DataADNI PET Test DataOASIS
MRI Test
Data
OASIS
PET Test
Data
Accuracy (%)99.0099.0095.9195.7498.3096.30
Precision (%)99.0098.5095.5097.0098.0097.00
Recall (%)98.5098.5096.5094.0098.5095.50
F1-Score (%)99.9998.5095.5095.5098.0096.00
Sensitivity (%)96.8810010088.2410094.29
Specificity (%)10097.3793.1010096.43100
Table 6. Comparison of Proposed Method with Existing Methods.
Table 6. Comparison of Proposed Method with Existing Methods.
ReferenceDatabaseMethodBinary ClassificationAccuracy (%)Specificity (%)Sensitivity (%)
[17]ADNIFeature level fusion + Ensemble ClassifierAD vs. CN99.00--
[18]ADNIHypergraph-based RegularizationAD vs. CN92.5190.4494.08
[19]ADNI3D CNN with Sparse AutoencoderAD vs. CN93.2195.4291.43
[22]ADNIInception–Resnet with DWTAD vs. CN97.0097.0097.00
[23]ADNIDWT + Pretrained ViTAD vs. EMCI93.75--
ProposedADNIOptimized Transposition + Mobile ViTAD vs. CN96.0097.0094.12
AANLIB99.0099.0098.44
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Odusami, M.; Maskeliūnas, R.; Damaševičius, R. Optimized Convolutional Fusion for Multimodal Neuroimaging in Alzheimer’s Disease Diagnosis: Enhancing Data Integration and Feature Extraction. J. Pers. Med. 2023, 13, 1496. https://doi.org/10.3390/jpm13101496

AMA Style

Odusami M, Maskeliūnas R, Damaševičius R. Optimized Convolutional Fusion for Multimodal Neuroimaging in Alzheimer’s Disease Diagnosis: Enhancing Data Integration and Feature Extraction. Journal of Personalized Medicine. 2023; 13(10):1496. https://doi.org/10.3390/jpm13101496

Chicago/Turabian Style

Odusami, Modupe, Rytis Maskeliūnas, and Robertas Damaševičius. 2023. "Optimized Convolutional Fusion for Multimodal Neuroimaging in Alzheimer’s Disease Diagnosis: Enhancing Data Integration and Feature Extraction" Journal of Personalized Medicine 13, no. 10: 1496. https://doi.org/10.3390/jpm13101496

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop