Next Article in Journal
Investigating the Effects of Transcranial Alternating Current Stimulation on Cortical Oscillations and Network Dynamics
Previous Article in Journal
Prior Engagement in Physical Activity Correlates with Enhanced Quality of Life Perceptions among Older Adults during COVID-19 Lockdown
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ASD-GANNet: A Generative Adversarial Network-Inspired Deep Learning Approach for the Classification of Autism Brain Disorder

by
Naseer Ahmed Khan
and
Xuequn Shang
*,†
School of Computer Science and Technology, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Current address: School of Computer Science and Technology, Changan Campus, Northwestern Polytechnical University, Xi’an 710072, China.
Brain Sci. 2024, 14(8), 766; https://doi.org/10.3390/brainsci14080766
Submission received: 8 July 2024 / Revised: 23 July 2024 / Accepted: 24 July 2024 / Published: 29 July 2024
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
The classification of a pre-processed fMRI dataset using functional connectivity (FC)-based features is considered a challenging task because of the set of high-dimensional FC features and the small dataset size. To tackle this specific set of FC high-dimensional features and a small-sized dataset, we propose here a conditional Generative Adversarial Network (cGAN)-based dataset augmenter to first train the cGAN on computed connectivity features of NYU dataset and use the trained cGAN to generate synthetic connectivity features per category. After obtaining a sufficient number of connectivity features per category, a Multi-Head attention mechanism is used as a head for the classification. We name our proposed approach “ASD-GANNet”, which is end-to-end and does not require hand-crafted features, as the Multi-Head attention mechanism focuses on the features that are more relevant. Moreover, we compare our results with the six available state-of-the-art techniques from the literature. Our proposed approach results using the “NYU” site as a training set for generating a cGAN-based synthetic dataset are promising. We achieve an overall 10-fold cross-validation-based accuracy of 82%, sensitivity of 82%, and specificity of 81%, outperforming available state-of-the art approaches. A sitewise comparison of our proposed approach also outperforms the available state-of-the-art, as out of the 17 sites, our proposed approach has better results in the 10 sites.

1. Introduction

Autism spectrum disorder (ASD), as described by the American Psychiatric Association (APA) in the Diagnostic Statistical Manual (DSM-5) [1], is a neurological disorder characterized by a lack of communication, restrained social skills, and visible signs of repetitive behaviors, like the stacking of objects, in children who have ASD. ASD is called a “spectrum” disorder because no two people having ASD exhibit the same phenotype, as most ASD conditions vary from person to person and ASD subjects can have variations in skills, symptoms, and cognitive abilities [2]. The prevalence of ASD worldwide is around 1% [3], while estimates are even higher in high-income countries; for example, in the US [4], about 2.3% of children aged 8 years and about 2.2% of adults are diagnosed with ASD, which makes ASD a global economic burden. The early diagnosis of ASD is considered crucial because early diagnosis and the timely start of therapies and interventions in ASD subjects have shown improvement in the quality of life of people with ASD [5]. The challenging issue is that the early diagnosis of ASD is difficult because of the lack of trained ASD specialists worldwide, demographic variations across countries, and universal ASD screening tests. Most importantly, the interventions and therapies that exist around the world to target ASD are mostly designed for the adult population [6,7]. In light of this evidence, it is of significant interest to find psychological interventions, biological drugs, and artificial intelligence-based methods to tackle and treat ASD in general and find those neuro-markers and genetic factors related to ASD that can reduce the economic burden of ASD and improve the quality of lives of people suffering with ASD in particular.

1.1. Motivations

The functional magnetic resonance imaging (fMRI) modality has been receiving more attention from researchers in the field of neuro-science because of its non-invasiveness, high spatial resolution, and lower cost spent over time. fMRI measures the neuronal activity in the brain by measuring the magnitude of the blood-oxygenated-level-dependent (BOLD) signals [8,9]. A BOLD signal reflects the changes in blood oxygenation levels in response to neuronal activity in a brain region. The key idea of fMRI is to measure these changes, which are indicative of increased blood flow and oxygenation following neuronal activation. A range of brain disorders, including Alzheimer’s [10], attention deficit hyperactivity disorder [11], autism spectrum disorder [12], epilepsy [13], major depressive disorders [14], and Parkinson’s [15], have now been explored using fMRI, which makes fMRI a widely researched modern modality for studying various brain disorders. A plethora of fMRI studies on various brain disorders, the non-invasive nature of fMRI, and the high spatial resolution of fMRI make this imaging modality a promising research direction for pursuing and exploring brain disorders.
Of late, deep learning (DL)-based approaches [16] have revolutionized the field of artificial intelligence and are being used in a variety of tasks, such as classification, detection, segmentation, and natural language processing. The underlying idea of DL, which is also called representational learning, is to map the underlying data distribution to a non-linear latent dimension by first using a multi-layered neural network and then training the network using the back-propagation algorithm. After training, the trained neural networks are tested on unseen or testing data. DL-based approaches have shown huge potential in the fields of image classification [17], brain tumor detection [18], drug discovery [19], and natural language processing (NLP) [20], which makes DL an active area of research for many open problems, consequently motivating a community of researchers to use the ideas of DL in their respective fields. In the following, we briefly review the related articles specific to our proposed approach, with which we compare our results.
Conditional Generator Adversarial Networks (cGANs) [21] comprise an unsupervised category or class-embedded approach that is used to generate a synthetic dataset when the dataset is rare and the problem of over-fitting the model is to be reduced using the synthetically generated dataset. cGAN consists of two neural networks called the “Generator” and “Discriminator”, where the task of the Generator is to generate real data using the random vector and a class-embedding vector to fool the Discriminator, and the task of the Discriminator is to identify if the data generated by the Generator are real or fake using the real and fake data with class-embedding information. Our motivations for using the cGAN comes from its ability to generate synthetic data to reduce the problem of over-fitting because of highly dimensional connectivity features. We use the cGAN to generate connectivity features across the category of subjects so that more of the dataset on connectivity features can be generated, and our model has solved the issue of over-fitting.

1.2. Related State of the Art

RFEGNN [22], a “Recursive Features Extraction-based Graph Neural Network” approach, was proposed to classify ASD using spatial features extracted from recursive feature elimination and then concatenating them with the phenotype information on the subjects. The authors achieved promising results, as their fusion-based approach achieved an overall accuracy of 80%. A potential drawback of their approach is the use of a feature selection approach and a concatenated approach to fuse two sets of features. Our proposed approach neither relies on those hand-crafted features nor uses the phenotype information and still outperforms this approach, which shows the robustness of our approach.
MHSA [23] is a “Multi-Head Self Attention”-based promising approach that uses the architecture of transformer-based Multi-Head attention and a data augmentation module. The authors achieved promising results with their end-to-end approach that did not require hand-crafted features with 81.47% accuracy, 83.8% sensitivity, and 80.16% specificity. But the main weaknesses of their approach was the augment module where the authors used a sliding window approach and a large number of parameters needed for selection and experimentation. To this end, our proposed approach of cGAN-powered data augment module solves this problem by not relying on the sliding window to augment the dataset.
MVES [24] is a “Multi-View Ensemble”-based approach to tackle the classification of ASD by first extracting the mean time series of the fMRI data and then by a selection of low-/high-level functional connectivity. It features using PCA and autoencoder and finally an ensemble-based model for the classification. The authors reported a classification accuracy of 72% and a highest classification accuracy of 92.9% on the CMU site. In MVES, the authors used a fixed-length window approach to augment the data, which is also the main limitation of their approach, because by using the fixed-length window, the inherent meaningful structure of the full-length fMRI is ignored. To this end, we again firmly believe that our proposed approach is superior to their fixed-length approach because our cGAN-based data augmentation approach tackles data augmentation by generating fixed-length correlation features that do not depend on the varying time series of a subject.
NVS [25], a “Novel Features Selection”-based approach that uses a pre-trained variational autoencoder, achieved a 10-fold accuracy of 78.12% on the overall dataset. The authors also proposed an innovative activation function and devised a normalization pipeline. But their approach used the hand-crafted feature selection approach that is now considered obsolete. Our end-to-end tackles this limitation well because, it does not rely on a pre-selection of features. Moreover, the authors did not report sitewise metrics for the 17 sites, which again limits the usefulness of their approach as it might not work well on sitewise data because some sites have significantly smaller data points, which make classification a challenging task on individual sites.
MSC [26] is a “Multi-Site Clustering”-based approach that tackles the classification task using nested feature selection, where the authors first divide the dataset into two categories of ASD and HC, and then the features are selected from the two sites using single-value decomposition, resulting in an accuracy of 68.42%. The major limitations of their work are the hand-crafted features and the reporting of accuracy on fewer sites instead of all the 17 sites.
DeepGCN [27], a “Deep Graph Convolution Neural Network”-based approach, uses an end-to-end idea using Graph Convolution Network (GCN) that has achieved promising results on the ABIDE dataset and the authors have reported a cross-validated accuracy of 73.7%. The main limitation of their approach is that, although the authors used the end-to-end approach for the classification of ABIDE, they did not report a sitewise comparison of the accuracy, which limits the usefulness of their approach because the sitewise high accuracy is considered more challenging due to lesser data points, as previously discussed.

1.3. Contributions

The main contributions of the proposed approach are as follows:
  • A cGAN-based synthetic dataset generator is proposed to augment the connectivity features of a subject. The reason for the generation of fixed-length connectivity features per subject using Pearson correlation is to avoid the variable time dimension of the subjects’ time series data.
  • A Multi-Head-attention-based deep learning model is proposed here using the idea of multi-channel keys, query, and value to capture diverse aspects of the features. The motivation for this idea is because we do not want to use hand-crafted features for the classification as a Multi-Head-based attention mechanism with skip connections, tackling the task of selecting useful features in an end-to-end manner.
  • A pre-trained autoencoder trained on the training subjects’ data is used to train the cGAN because we want to generate the subjects’ features which are as close to the original subjects’ distribution as possible. Therefore, when training the cGAN, in addition to the losses of the Discriminator and Generator, we also focus on the pre-trained autoencoder loss to verify if the generated subjects follow the distribution of the original datasets or not.

1.4. Objectives

We use correlation-based connectivity features computed using Pearson’s correlation to tackle the task of varying time series on subjects. The connectivity features have a dimension of 6670 and they pose a problem of over-fitting in the classification model because of a much smaller size of dataset. The main objectives of the proposed deep learning-based approach are to first come up with a synthetic subject’s data on ASD and HC and use an end-to-end framework for the classification of subjects into ASD and HC without using hand-crafted features. Since the ASD dataset on some sites contain very few subjects which make accurate classification very challenging, our cGAN-based data augmentation tackles this problem in an ingenious way. Moreover, The Multi-Head attention mechanism that we propose tackles the problem of over-fitting by focusing on only those areas of the brain regions which actually play their roles in the ASD, underlying neural activity patterns.

2. Materials and Methods

We now describe the public dataset on AN ASD-resting-state fMRI, challenges related to the dataset, and the proposed methodology in detail.

2.1. Dataset

Autism Brain Imaging Data Exchange (ABIDE) [28] is a public consortium that has provided the resting-state fMRI dataset on the individuals with ASD, which is pre-processed, as described in [29], where pre-processing includes steps like slice-timing correction, motion realignment, normalization, and registration. The original dataset includes 1112 individuals with 539 suffering from the ASD and 573 are typical controls, but as per the latest updates, the dataset now contains 884 total subjects as shown in Table 1. Moreover, we use Automated Anatomical Labeling (AAL) [30]-based regions of interests .
The distribution of ASD and HC on the whole dataset and across the sites is shown in the following Figure 1. It can be seen from the figure that although the overall distribution of ASD and HC is not imbalanced, when observing the distribution sitewise, there are drastic variations in the number of ASD and HC across sites. For example, the site CMU does not even have a total of 10 data points, whereas the NYU site has more than 100 data points. These observations make stratified split of training and testing more logical because otherwise it might be possible that no training data are included from the CMU site, leading to biases in the model.

2.2. Stratified Training and Testing Split

The imbalance nature of the dataset as we have just explained in the previous section necessitates us to use a stratified-based partition of the dataset into training and testing ones, as shown in Figure 2. The reason to use stratified-based splitting of dataset for training and testing is because of the varying sizes of ASD and HC across the sites. The site “SBL” has only 26 subjects while the site “NYU” has 117 subjects; so when using cross-validation-based on no-stratification, the “SBL” site might have no representation in the training set, resulting in poor performance predictions on the “SBL” testing dataset. Therefore, in the rest of study, we use a stratified-based split of dataset into the training and testing ones so that every site receives its representation in the training and testing datasets.

2.3. Methodology

In Figure 3, we now explain our proposed model for the classification of ASD on the ABIDE pre-reprocessed rs-fMRI dataset. The proposed model consists of three modules which we call the Features Module, Data Augment Module, and a Classification Module. The general flow of the proposed approach is as follows. Firstly, the connectivity features will be calculated for both categories of subjects that is ASD and HC. Secondly, the connectivity features will be used to train a cGAN to generate synthetic connectivity features and class per subjects so that the problem of over-fitting while training can be reduced. Finally, a classification module will be used for the classification of subjects to ASD and HC using the Multi-Head attention mechanism.
In the following subsections, we briefly describe each of the three modules that are shown in Figure 3.

2.3.1. Feature Module

The purpose of the this module as shown in Figure 4 is to convert the time series data on the brain regions across subjects to a fixed-length dimension that can be processed by the “Data Augment Module” and subsequently by the “Classification Module”. The time series data on the 116 brain regions vary sometimes within site and also across sites. Tackling the varying dimension of the time series data is a significant challenge because taking a fixed-size window for the time series across subjects not only violates the inherent structure of the pre-processed fMRI but also limits generalization over the subjects. To tackle this challenge, we have used a Pearson’s correlation “r”, as given in Equation (1), to measure the functional connectivity between any two brain regions “x” and “y”. Since we are using the AAL-116 template in the study where each subject has 116 brain regions, the correlation will result in a symmetric 116 × 116 matrix, from which we extract the upper diagonal “6670” values excluding the diagonal, because the diagonal is always 1 as the correlation between the two same brain regions will always be 1. Therefore, for each of the subjects, whether ASD or HC, we will have a “6670” dimensional connectivity feature vector, which will be used subsequently in the rest of the modules.
r = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( x i x ¯ ) 2 i = 1 n ( y i y ¯ ) 2
The motivations of using the resting-state functional connectivity (rs-fMRI) features is deeply rooted in how the neuronal activity in a disordered brain takes place as it has been shown that the resting state brain of a subject, when the subject does not perform any task, is different from the neuronal activity of the subjects without a disorder [31].

2.3.2. Data Augment Module

The purpose of this module, as shown in Figure 5, is to generate synthetic 6670 dimensional features across ASD and HC because the size of the dataset when compared to the size of features dimension is a lot lower, leading to the over-fitting of the deep learning models. We use the cGAN-based approach [21] to generate synthetic features across subjects as we not only want the features but we also want to condition the features across the subjects based on the subjects’ conditions, whether belonging to ASD or HC. The objective function of the conditional GAN (cGAN) as explained in [32] is given in Equation (2), where “G” and “D” are the Generator and Discriminator, “x” is the features vector, “y” is the class of the subject and “z” is the random vector. The main purpose of the cGAN min–max approach is to fool the Discriminator over time by the generator by generating subjects as close to the real data.
min G max D V ( D , G ) = E x p d a t a ( x ) [ log D ( x | y ) ] + E z p z ( z ) [ log ( 1 D ( G ( z | y ) ) ) ]
The motivations of using the cGAN-based approach is to overcome the problem of using fixed-size subject time series data and to minimize the problem of over-fitting that is faced when training the DL model.

2.3.3. Classification Module

In this module, as shown in Figure 6, we show how the final objective of the classification will be achieved. Here, first, we will have features and associated labels for the original and synthetic ASD and HC subjects, and then a Multi-Head attention-based module [33] will be used so that focus is on the relevant features, followed by a fully connected layer and a classification head.The Multi-Head attention-based module is designed using Equation (3), where “Q”, “K” and “V” are called the query, key, and value and “d” is a normalization constant.
A t t e n t i o n ( Q , K , V ) = s o f t m a x Q K T d k V

2.3.4. Algorithm

In the following section, we explain the proposed approach using a pseudo-algorithm as written in Algorithm 1.
The proposed algorithm takes an input of the subject data with an associated label on the training dataset and outputs the accuracy of the model on the testing dataset. Algorithm 1 is divided into the three submodules: Feature Module, Data Augment Module, and the Classification Module. The function “Train-Test-Split” as described in line 3 is used to partition the dataset into the training and testing datasets using stratification. The function “PearsonCorr” is the correlation coefficient as mentioned in Equation (1) and the function “UpperTria” is used to extract the unique values from the symmetric 116 × 116 connectivity matrix. The function “Generator” and “Discriminator” are the neural networks used to train the cGAN in an adversarial manner. The function “Multi-Head”, “FC’s”, and “Class” are the Multi-Head-based attention layer, fully connected layers, and the classification layer of the classification module.
Algorithm 1 Proposed cGAN-Powered Multi-Head Attention Approach’s Algorithm
1:
Input ← D, Output ← {}, cGANEpochs, synthEpochs, classEpochs
2:
X,YD (X ← Features Matrix, Y = ASD/HC)
3:
X t r a i n , X t e s t , Y t r a i n , Y t e s t ← Train-Test-Split(X,Y)
4:
F t r a i n ← {}, F t e s t ← {}
5:
(Feature Module):
6:
for  x t r a i n , x t e s t , y t r a i n , y t e s t in ( X t r a i n , X t e s t , T t r a i n , Y t e s t ) do
7:
       s u b j e c t T r a i n 116 x T = x t r a i n , s u b j e c t T e s t 116 x T = x t e s t
8:
       T r a i n 116 × 116 PearsonCorr( s u b j e c t T r a i n 116 x T )
9:
       T e s t 116 × 116 PearsonCorr( s u b j e c t T e s t 116 x T )
10:
     T r a i n 6670 UpperTria( T r a i n 116 × 116 )
11:
     T e s t 6670 UpperTria( T e s t 116 × 116 )
12:
     F t r a i n F t r a i n U { T r a i n 6670 , y t r a i n }
13:
     F t e s t F t e s t U { T e s t 6670 , y t e s t }
14:
end for
15:
(Data Augment Module):
16:
F s y n t h e t i c ← {}
17:
for epoch in cGANEpochs do
18:
     D a t a r a n d o m , l a b e l r a n d o m Random()
19:
     F r e a l , Y r e a l Generator( D a t a r a n d o m , l a b e l r a n d o m )
20:
     F r e a l , Y r e a l Subset( F t r a i n )
21:
    Discriminator( F f a k e , Y f a k e )
22:
    Discriminator( F r e a l , Y r e a l )
23:
end for
24:
for epoch in synthEpochs do
25:
     D s y n t h e t i c , l a b e l s y n t h e t i c Generator()
26:
     F s y n t h e t i c F s y n t h e t i c U { D s y n t h e t i c , l a b e l s y n t h e t i c }
27:
end for
28:
(Classification Module):
29:
for epoch in classEpochs do
30:
     F 1 Multi-Head( F t r a i n , F s y n t h e t i c )
31:
     F 2 F C 1 ( F 1 )
32:
     F 3 F C 2 ( F 2 )
33:
     F 4 F C 3 ( F 3 )
34:
     M o d e l t r a i n e d Class( F 4 )
35:
end for
36:
Output M o d e l t r a i n e d ( F t e s t )

3. Results

We now describe experimental settings related to the autoencoder, cGAN, and Multi-Head attention and the types of experiments related to the combined site and sitewise that we performed in the current study.

3.1. Experimental Settings

We use a “Denoising Autoencoder” to reconstruct connectivity features by mapping the original data into a latent representation. Then, we generate connectivity features using the cGAN, and finally, validate the generated subjects’ data distribution with the original subjects’ feature data by mapping the generated subjects’ features to the latent representation.

3.1.1. Denoising Autoencoder

Denoising Autoencoder has five encoder layers of size 4000, 3000, 2000, 1000, and 500 and five decoder layers of size 1000, 2000, 3000, 4000, and 6670. “Batch Normalization” and “Relu” activation are used after each of the encoder and decoder layers. “Learning Rate” was set to 0.00001, “Epochs” set to 5000, “Batch Size” set to 64 and the “Adam” optimization was used for the training of training while a random “Gaussian” noise was added to the input to make the decoder part of the model more robust.

3.1.2. cGAN

The Generator has a total of five layers, the noise dimension for random vector generation is set to 500, and the condition dimension for ASD and HC is also set to 500 as our purpose is to train the conditional cGAN. Before the first layer, we used the embedding layer for converting the two categories to a combined noise and condition layer, resulting in 1000 dimension. After the embedding layer, the layers have sizes of 2000, 3000, 4000, 5000, and 6670, where each layer is followed by the spectral normalization layer, the batch normalization layer, and a leakyRelu activation of 0.1. The Discriminator has five layers of sizes 4000, 2000, 1000, 500, and 1, where each layer is followed by spectral normalization and a LeakyRelu of 0.1. Before the first layer of the Discriminator, the embedding layer is used to convert the ASD/HC categories to 500 conditional dimension and then the first input layer has an input size of the feature size, that is 6670, and a conditional dimension size of 500. The learning rate for the Discriminator and Generator was set to 0.00001, optimization was set to Adam, batch size to 64, and the total epochs were set to 5000 for the training of the conditional cGAN.

3.1.3. Classification

The classification module consisted of a Multi-Head attention block, which has an input of size 6670, followed by a Multi-Head attention layer with 10 heads, a skip layer from input to the output of the Multi-Head attention layer, and a normalization layer followed by three feed forward layers of size 3000, 2000, and 1000, 1 containing skip connections from previous layer and a normalization layer. We set epochs to 5000, batch size to 64, learning rate to 0.0001, and optimizer to Adam for the training of the classification module.

3.2. Experiments

There are two kinds of experiments that we conducted for the current work. The first was an overall accuracy comparison where the whole dataset was first divided into the training and testing split using stratification. The second was training the model on the combined training data and then testing it on the combined testing dataset. A sitewise comparison, where we used the largest site NYU for the training, trained the model on this site and then tested on the remaining sites, but excluding the CMU site because it contained too few subjects to compare.

3.2.1. Overall Comparison

In the first experiment, we compare the results of the proposed approach with the state-of-the-art approaches as shown in Table 2. It can be seen that our proposed approach accuracy, precision, and recall have outperformed the other approaches, which validates the effectiveness of the proposed approach on the whole data. To analyze the classification model performance, the Receiver Operating Curve (ROC) is performed by varying the thresholds of prediction based on the combined dataset, as shown in Figure 7.

3.2.2. Sitewise Comparison

In the second comparison, we used the “NYU” training total dataset as the training dataset and tested the trained model on the rest of the 15 sites excluding the “CMU” site, as performed in [23]. We have shown the results of the proposed approach with the state-of-the-art approaches and it can be seen that our proposed approach has outperformed the other approaches in 10 out of the 15 sites, validating the effectiveness of the proposed approach (Table 3).

4. Discussion

The generation of the synthetic subject’s connectivity features using cGAN is not a trivial task because we have no prior validation method about the goodness of the synthetic subject’s feature set. To tackle this issue, we sampled a small subset of the training data and trained a Denoising Autoencoder on the training dataset. The objective was that after training the Denoising Autoencoder on a small subset of subjects, it will learn the latent representation of the training dataset. Secondly, during the cGAN training stage, the generated subjects will be fed to the trained Denoising Autoencoder and then the reconstructed subject using the decoder part will be compared to the original input to measure the loss. This step is crucial because even after the minimization of the loss of the Generator and the Discriminator of the cGAN, we should know how close the generated subject is to the distribution of the original subjects. The Denoising Autoencoder loss and the comparison of the original and the reconstructed features are shown in Figure 8.
We have carried out experimentation with a varying number of synthetic subjects as shown in Figure 9. It can be seen that the results improve when the number of the synthetic subjects are 2000. A pattern can be seen in the values of the accuracy across the synthetic subjects, when the number of synthetic subjects are fewer, and the value of accuracy is low, but it gradually increases and peaks when the number of subjects are 2000, which shows that the over-fitting issue is gradually resolved by the model when attaining a sufficient number of subjects. On the other hand, too many synthetic subjects is also not a favorable scenario for the model because, in that case, the model will learn more and more the distribution of synthetic data instead of the original data and will forget about the true distribution, resulting in lower accuracy.

Ablation Studies

We have also compared the performance of our model by taking out the essential components from our proposed approach to analyze how the proposed model behaves in scenarios when those components are present or absent in the model. For the purpose of this comparison, we performed two kinds of experiments where we first only used the cGAN component, named as “cGAN Only”, in which the downstream task of classification was performed using a simple neural network consisting of seven fully connected layers of 4000, 3000, 2000, 1000, 500, 50, and 1, where each layer was followed by “BatchNorm”, “ReLU”, and “Dropout” of 10%. In the second experiment, we only used the “Muti-Head” attention layer named as “Multi-Head Only” and did not use cGAN for synthetic data generation.
For both kinds of experiments, the results of cross-validated accuracy on the whole dataset and sitewise are displayed in Table 4 and Table 5, respectively. It is obvious from both the tables that the accuracy results suffered significantly in both the comparison, which proves that both the cGAN and Multi-Head attention component are essential for the proposed approach.

5. Conclusions

In this paper, we proposed ASD-GANNet, a cGAN-based data augmentation technique for the classification of ASD on the fMRI pre-processed ABIDE dataset. The challenge of synthetic data generation in the presence of varying time points across subjects was tackled by generating the fixed-length connectivity features as connectivity features across subjects have a fixed dimension. Our proposed framework is promising as we not only achieved good performance but also compared our results with the state-of-the-art methods in the literature and proved the effectiveness of the proposed approach.

Limitations

The major limitation of the proposed approach is that we were not able to generate the original time-varying synthetic data using cGAN. There are varying time points across the sites per subject and it is very challenging to condition the cGAN on this additional time condition. In future work, we will explore how the subjects belonging to ASD or HC could be generated instead of the fixed-length correlation features and how the synthetic subjects’ distributions align with the overall original data distribution and the sitewise distribution on the ASD and HC subjects.

Author Contributions

N.A.K. conceived the idea, performed the experiments, performed the detailed analysis of the research work, and prepared the manuscript. X.S. repeatedly analyzed the results, performed critical analysis, reviewed the draft and gave valuable suggestions. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Innovation Capability Support Program of Shaanxi (No:2021TD-06).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

ABIDE pre-processed fMRI dataset is a publicly available dataset maintained at http://preprocessed-connectomes-project.org/abide/ (accessed on 1 May 2024). To download the dataset, users can go to the Github tab of the mentioned url and under the “abide” repo can run the script named “download_abide_preproc.py” to download all the pre-processed fMRI datasets of ASD.

Acknowledgments

We are grateful to China National Natural Science Foundation, China Scholarships Council, Data Mining Laboratory of Northwestern Polytechnical University, China, for providing us with administrative and moral support to pursue the proposed research works.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ABIDEAutism Brain Imaging Data Exchange
ASDAutism Spectrum Disorder
cGANConditional Generative Adversarial Networks
fMRIFunctional Magnetic Resonance Imaging
GANGenerative Adversarial Networks
WGANWasserstein Generative Adversarial Networks
rs-fMRIResting State Functional Magnetic Resonance Imaging

References

  1. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders: DSM-5; American Psychiatric Association: Washington, DC, USA, 2013; Volume 5. [Google Scholar]
  2. Lord, C.; Cook, E.H.; Leventhal, B.L.; Amaral, D.G. Autism spectrum disorders. Neuron 2000, 28, 355–363. [Google Scholar] [CrossRef] [PubMed]
  3. Lord, C.; Elsabbagh, M.; Baird, G.; Veenstra-Vanderweele, J. Autism spectrum disorder. Lancet 2018, 392, 508–520. [Google Scholar] [CrossRef] [PubMed]
  4. Hirota, T.; King, B.H. Autism spectrum disorder: A review. JAMA 2023, 329, 157–168. [Google Scholar] [CrossRef] [PubMed]
  5. Gabbay-Dizdar, N.; Ilan, M.; Meiri, G.; Faroy, M.; Michaelovski, A.; Flusser, H.; Menashe, I.; Koller, J.; Zachor, D.A.; Dinstein, I. Early diagnosis of autism in the community is associated with marked improvement in social symptoms within 1–2 years. Autism 2022, 26, 1353–1363. [Google Scholar] [CrossRef] [PubMed]
  6. Iyama-Kurtycz, T.; Iyama-Kurtycz, T. Obstacles to diagnosing autism spectrum disorder and myths about autism. In Diagnosing and Caring for the Child with Autism Spectrum Disorder: A Practical Guide for the Primary Care Provider; Springer: Cham, Switzerland, 2020; pp. 9–17. [Google Scholar]
  7. McCarty, P.; Frye, R.E. Early detection and diagnosis of autism spectrum disorder: Why is it so difficult? Semin. Pediatr. Neurol. 2020, 35, 100831. [Google Scholar] [CrossRef] [PubMed]
  8. Soares, J.M.; Magalhães, R.; Moreira, P.S.; Sousa, A.; Ganz, E.; Sampaio, A.; Alves, V.; Marques, P.; Sousa, N. A Hitchhiker’s guide to functional magnetic resonance imaging. Front. Neurosci. 2016, 10, 218898. [Google Scholar] [CrossRef] [PubMed]
  9. Chow, M.S.; Wu, S.L.; Webb, S.E.; Gluskin, K.; Yew, D. Functional magnetic resonance imaging and the brain: A brief review. World J. Radiol. 2017, 9, 5. [Google Scholar] [CrossRef] [PubMed]
  10. Lajoie, I.; Nugent, S.; Debacker, C.; Dyson, K.; Tancredi, F.B.; Badhwar, A.; Belleville, S.; Deschaintre, Y.; Bellec, P.; Doyon, J.; et al. Application of calibrated fMRI in Alzheimer’s disease. NeuroImage Clin. 2017, 15, 348–358. [Google Scholar] [CrossRef] [PubMed]
  11. Yap, K.H.; Manan, H.A.; Sharip, S. Heterogeneity in brain functional changes of cognitive processing in ADHD across age: A systematic review of task-based fMRI studies. Behav. Brain Res. 2021, 397, 112888. [Google Scholar] [CrossRef]
  12. Santana, C.P.; de Carvalho, E.A.; Rodrigues, I.D.; Bastos, G.S.; de Souza, A.D.; de Brito, L.L. rs-fMRI and machine learning for ASD diagnosis: A systematic review and meta-analysis. Sci. Rep. 2022, 12, 6030. [Google Scholar] [CrossRef]
  13. Guo, L.; Bai, G.; Zhang, H.; Lu, D.; Zheng, J.; Xu, G. Cognitive functioning in temporal lobe epilepsy: A BOLD-fMRI study. Mol. Neurobiol. 2017, 54, 8361–8369. [Google Scholar] [CrossRef] [PubMed]
  14. Casteràs, D.P.; Cano, M.; Camprodon, J.; Loo, C.; Palao, D.; Soriano-Mas, C.; Cardoner, N. A multimetric systematic review of fMRI findings in patients with MDD receiving ECT. Eur. Neuropsychopharmacol. 2021, 44, S47. [Google Scholar] [CrossRef]
  15. Wolters, A.F.; van de Weijer, S.C.; Leentjens, A.F.; Duits, A.A.; Jacobs, H.I.; Kuijf, M.L. “Resting-state fMRI in Parkinson’s disease patients with cognitive impairment: A meta-analysis”: Answer to Wang and colleagues. Park. Relat. Disord. 2019, 66, 253–254. [Google Scholar] [CrossRef] [PubMed]
  16. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  17. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  18. Nazir, M.; Shakil, S.; Khurshid, K. Role of deep learning in brain tumor detection and classification (2015 to 2020): A review. Comput. Med Imaging Graph. 2021, 91, 101940. [Google Scholar] [CrossRef] [PubMed]
  19. Lavecchia, A. Deep learning in drug discovery: Opportunities, challenges and future prospects. Drug Discov. Today 2019, 24, 2017–2032. [Google Scholar] [CrossRef] [PubMed]
  20. Lauriola, I.; Lavelli, A.; Aiolli, F. An introduction to deep learning in natural language processing: Models, techniques, and tools. Neurocomputing 2022, 470, 443–456. [Google Scholar] [CrossRef]
  21. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  22. Yang, J.; Hu, M.; Hu, Y.; Zhang, Z.; Zhong, J. Diagnosis of Autism Spectrum Disorder (ASD) Using Recursive Feature Elimination–Graph Neural Network (RFE–GNN) and Phenotypic Feature Extractor (PFE). Sensors 2023, 23, 9647. [Google Scholar] [CrossRef]
  23. Zhao, F.; Feng, F.; Ye, S.; Mao, Y.; Chen, X.; Li, Y.; Ning, M.; Zhang, M. Multi-head self-attention mechanism-based global feature learning model for ASD diagnosis. Biomed. Signal Process. Control 2024, 91, 106090. [Google Scholar] [CrossRef]
  24. Kang, L.; Chen, J.; Huang, J.; Jiang, J. Autism spectrum disorder recognition based on multi-view ensemble learning with multi-site fMRI. Cogn. Neurodyn. 2023, 17, 345–355. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, F.; Wei, Y.; Liu, J.; Wang, Y.; Xi, W.; Pan, Y. Identification of Autism spectrum disorder based on a novel feature selection method and Variational Autoencoder. Comput. Biol. Med. 2022, 148, 105854. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, N.; Yao, D.; Ma, L.; Liu, M. Multi-site clustering and nested feature extraction for identifying autism spectrum disorder with resting-state fMRI. Med. Image Anal. 2022, 75, 102279. [Google Scholar] [CrossRef] [PubMed]
  27. Cao, M.; Yang, M.; Qin, C.; Zhu, X.; Chen, Y.; Wang, J.; Liu, T. Using DeepGCN to identify the autism spectrum disorder from multi-site resting-state data. Biomed. Signal Process. Control 2021, 70, 103015. [Google Scholar] [CrossRef]
  28. Di Martino, A.; Yan, C.G.; Li, Q.; Denio, E.; Castellanos, F.X.; Alaerts, K.; Anderson, J.S.; Assaf, M.; Bookheimer, S.Y.; Dapretto, M.; et al. The autism brain imaging data exchange: Towards a large-scale evaluation of the intrinsic brain architecture in autism. Mol. Psychiatry 2014, 19, 659–667. [Google Scholar] [CrossRef]
  29. Craddock, C.; Benhajali, Y.; Chu, C.; Chouinard, F.; Evans, A.; Jakab, A.; Khundrakpam, B.S.; Lewis, J.D.; Li, Q.; Milham, M.; et al. The neuro bureau preprocessing initiative: Open sharing of preprocessed neuroimaging data and derivatives. Front. Neuroinform. 2013, 7, 5. [Google Scholar]
  30. Korhonen, O.; Saarimäki, H.; Glerean, E.; Sams, M.; Saramäki, J. Consistency of regions of interest as nodes of fMRI functional brain networks. Netw. Neurosci. 2017, 1, 254–274. [Google Scholar] [CrossRef] [PubMed]
  31. Smith, S.M.; Vidaurre, D.; Beckmann, C.F.; Glasser, M.F.; Jenkinson, M.; Miller, K.L.; Nichols, T.E.; Robinson, E.C.; Salimi-Khorshidi, G.; Woolrich, M.W.; et al. Functional connectomics from resting-state fMRI. Trends Cogn. Sci. 2013, 17, 666–682. [Google Scholar] [CrossRef]
  32. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
Figure 1. Distribution of ASD and HC: (a) ASD vs. HC on the combined dataset. (b) ASD vs. HC across 17 sites.
Figure 1. Distribution of ASD and HC: (a) ASD vs. HC on the combined dataset. (b) ASD vs. HC across 17 sites.
Brainsci 14 00766 g001
Figure 2. Stratified training and testing split of the dataset.
Figure 2. Stratified training and testing split of the dataset.
Brainsci 14 00766 g002
Figure 3. Architecture of the ASD-GANNet.
Figure 3. Architecture of the ASD-GANNet.
Brainsci 14 00766 g003
Figure 4. Feature module.
Figure 4. Feature module.
Brainsci 14 00766 g004
Figure 5. Data Augment Module.
Figure 5. Data Augment Module.
Brainsci 14 00766 g005
Figure 6. Classification module.
Figure 6. Classification module.
Brainsci 14 00766 g006
Figure 7. ROC curve on the combined dataset.
Figure 7. ROC curve on the combined dataset.
Brainsci 14 00766 g007
Figure 8. Training of the Denoising Autoencoder (a). Loss of the Denoising Autoencoder model. (b) Distribution of the reconstructed and the original features.
Figure 8. Training of the Denoising Autoencoder (a). Loss of the Denoising Autoencoder model. (b) Distribution of the reconstructed and the original features.
Brainsci 14 00766 g008
Figure 9. Accuracy, sensitivity, and specificity metric comparison with the cGAN generated data.
Figure 9. Accuracy, sensitivity, and specificity metric comparison with the cGAN generated data.
Brainsci 14 00766 g009
Table 1. Pre-processed rs-fMRI dataset provided by ABIDE.
Table 1. Pre-processed rs-fMRI dataset provided by ABIDE.
Site NameASD 1HC 2Total
1CALTECH191837
2CMU325
3KKU122739
4LEUVEN273461
5MAXMUN182442
6NYU7398171
7OHSU121123
8OLIN141125
9PITT222345
10SBL141226
11SDSU122133
12STANFORD171936
13TRINITY212344
14UCLA363975
15UM4865113
16USM382361
17YALE222648
TOTAL408476884
1 ASD = Autism spectrum disorder, 2 HC = healthy controls.
Table 2. Tenfold cross-validation based on overall dataset comparison with the state of the art (%).
Table 2. Tenfold cross-validation based on overall dataset comparison with the state of the art (%).
MethodAccuracyPrecisionRecallSensitivitySpecificity
1RFEGNN [22]80.6380.21-76.24-
3MHSA [23]81.4083.8080.1683.8080.16
4MVES [24]72.00----
5NVS [25]78.00--80.0080.19
6MSC [26]68.42--70.0563.64
7DeepGCN [27]73.0272.9768.80--
8ASD-GANNet82.0084.0081.0082.0081.00
Table 3. Site wise dataset comparison with the state-of-the-art approaches (%).
Table 3. Site wise dataset comparison with the state-of-the-art approaches (%).
SiteSizeMHSA [23]MVES [24]ASD-GANNet
1CALTECH3764.6071.4071.60
2KKI3979.6075.0075.40
3LEUVEN6170.4072.1073.20
4MaxMun4266.4061.5066.90
5OUSH2366.00-60.00
6Olin2576.0073.5076.00
7Pitt4565.8074.5073.00
8SBL2689.3064.3065.00
9SDSU3375.7079.3079.70
10Stanford3677.5069.2070.00
11Trinity4473.1073.3074.00
12UCLA7569.3077.6078.50
13USM11376.9087.3088.10
14UM6174.9075.7075.90
15Yale4875.1080.0081.10
Table 4. Tenfold cross-validation-based overall dataset comparison with reduced model’s component (%).
Table 4. Tenfold cross-validation-based overall dataset comparison with reduced model’s component (%).
MethodAccuracyPrecisionRecallSensitivitySpecificity
1cGAN Only66.1068.5068.2069.5055.30
2Multi-Head Only65.0060.3059.6055.5060.00
3ASD-GANNet (GAN + Multi-Head)82.0084.0081.0082.0081.00
Table 5. Sitewise dataset comparison with model’s reduced component (%).
Table 5. Sitewise dataset comparison with model’s reduced component (%).
SiteGAN OnlyMulti-Head OnlyASD-ccGANNet (GAN + Multi-Head)
1CALTECH59.2050.0071.60
2KKI50.4036.0075.40
3LEUVEN48.0030.0073.20
4MaxMun44.0055.6066.90
5OUSH50.0055.3060.00
6Olin59.0058.0076.00
7Pitt48.5051.5073.00
8SBL44.0049.0065.00
9SDSU59.6059.0079.60
10Stanford58.0650.5570.00
11Trinity59.6050.6074.00
12UCLA58.1258.9078.50
13USM50.1450.5588.10
14UM60.1361.6075.90
15Yale59.3650.1581.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, N.A.; Shang, X. ASD-GANNet: A Generative Adversarial Network-Inspired Deep Learning Approach for the Classification of Autism Brain Disorder. Brain Sci. 2024, 14, 766. https://doi.org/10.3390/brainsci14080766

AMA Style

Khan NA, Shang X. ASD-GANNet: A Generative Adversarial Network-Inspired Deep Learning Approach for the Classification of Autism Brain Disorder. Brain Sciences. 2024; 14(8):766. https://doi.org/10.3390/brainsci14080766

Chicago/Turabian Style

Khan, Naseer Ahmed, and Xuequn Shang. 2024. "ASD-GANNet: A Generative Adversarial Network-Inspired Deep Learning Approach for the Classification of Autism Brain Disorder" Brain Sciences 14, no. 8: 766. https://doi.org/10.3390/brainsci14080766

APA Style

Khan, N. A., & Shang, X. (2024). ASD-GANNet: A Generative Adversarial Network-Inspired Deep Learning Approach for the Classification of Autism Brain Disorder. Brain Sciences, 14(8), 766. https://doi.org/10.3390/brainsci14080766

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop