Next Article in Journal
Rapid, Precise and Affordable Estimation of Venlafaxine and Its Metabolites in Highly Polluted Effluent Waters: Proof-of-Concept for Methodology
Next Article in Special Issue
Newly Synthesized Fluorinated Cinnamylpiperazines Possessing Low In Vitro MAO-B Binding
Previous Article in Journal
Are Multicentre Bond Indices and Related Quantities Reliable Predictors of Excited-State Aromaticity?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of the Multiple Stages of Parkinson’s Disease by a Deep Convolution Neural Network Based on 99mTc-TRODAT-1 SPECT Images

1
Department of Medical Imaging and Radiological Science, I-Shou University, No. 8, Yida Road., Jiao-su Village Yan-chao District, Kaohsiung City 82445, Taiwan
2
Department of Anesthesiology, E-DA Cancer Hospital, I-Shou University, No.1, Yida Road, Jiao-su Village, Yan-chao District, Kaohsiung City 82445, Taiwan
3
Department of Information Engineering, I-Shou University, No.1, Sec. 1, Syuecheng Road., Dashu District, Kaohsiung 84001, Taiwan
4
Department of Life Sciences, National Taitung University, No.369, Sec. 2, University Road, Taitung 95092, Taiwan
5
Department of Nuclear Medicine, E-DA Hospital, I-Shou University, No.1, Yida Rd, Jiao-su Village, Yan-chao District, Kaohsiung 82445, Taiwan
*
Authors to whom correspondence should be addressed.
Molecules 2020, 25(20), 4792; https://doi.org/10.3390/molecules25204792
Submission received: 7 September 2020 / Revised: 12 October 2020 / Accepted: 16 October 2020 / Published: 19 October 2020
(This article belongs to the Special Issue Radiolabelled Molecules for Brain Imaging with PET and SPECT II)

Abstract

:
Single photon emission computed tomography (SPECT) has been employed to detect Parkinson’s disease (PD). However, analysis of the SPECT PD images was mostly based on the region of interest (ROI) approach. Due to limited size of the ROI, especially in the multi-stage classification of PD, this study utilizes deep learning methods to establish a multiple stages classification model of PD. In the retrospective study, the 99mTc-TRODAT-1 was used for brain SPECT imaging. A total of 202 cases were collected, and five slices were selected for analysis from each subject. The total number of images was thus 1010. According to the Hoehn and Yahr Scale standards, all the cases were divided into healthy, early, middle, late four stages, and HYS I~V six stages. Deep learning is compared with five convolutional neural networks (CNNs). The input images included grayscale and pseudo color of two types. The training and validation sets were 70% and 30%. The accuracy, recall, precision, F-score, and Kappa values were used to evaluate the models’ performance. The best accuracy of the models based on grayscale and color images in four and six stages were 0.83 (AlexNet), 0.85 (VGG), 0.78 (DenseNet) and 0.78 (DenseNet).

1. Introduction

In recent years, single photon emission computed tomography (SPECT) has been used to estimate tumor growth, genetic treatments, brain function detection and cardiovascular diseases [1,2]. It is a mature imaging tool. In nuclear imaging, γ-rays are emitted from radiopharmaceuticals and received by a gamma camera placed around the object. The signal was then passed through the internal components, including scintillation crystals, photomultiplier tubes, positioning circuits, pulse height analyzers, etc. Finally, a radionuclide species distribution is obtained by image reconstruction algorithms.
However, the imaging process of nuclear medicine images is deeply disturbed by scattering events (called Compton scattering), especially for single photon emission computed tomography, which can cause quantitative image bias, low contrast and affect image quality [3,4], which prevents the fixed image results from appearing correctly. For example, in Siemens’ Symbia® S SPECT imaging the scattering ratio (under 172 keV energy) may reach 50% when using 111-In for trunk imaging (scatter fraction = total counts (172 keV)/total counts (247 keV)) [5]. This means one of every two signals from the camera is a self-scattering event, so the nuclear medicine image is blurred, but nuclear medicine imaging is mainly used to detect radioactive tracer’s spatial distribution in the human body. Moreover, it is particularly sensitive to detecting specific tissues or organs (such as the thyroid gland, brain, heart, liver, kidney, etc.), which is also called physiological imaging modality.
Parkinson’s disease (PD) is mainly caused by insufficient secretion of neurotransmitters (dopamine, acetylcholine) in the brain’s basal nucleus and neurodegeneration in the substantia nigra area. In past research, the dopamine neurons from the substantia nigra area are degraded due to increasing age or external factors [6]. This makes the presynaptic release of dopamine neurons insufficient and decreases dopamine transporter (DAT) levels. This then causes inhibition signals to be transmitted to the globus pallidus by the tail nucleus, and the putamen is reduced. Moreover, the inhibition signals to the ventral anterior thalamus are increased, which leads to insufficient signals to the premotor cortex and causes dyskinesia. Generally, almost 70% to 80% of the dopamine signals are lost by the time clinical signs begin, so early diagnosis of Parkinson’s disease is essential [7]. For the early stages of Parkinson’s disease or senior subjects, dopamine drugs (such as levodopa, dopamine agonists, anticholinergics) could be used to distinguish patients’ symptoms and pathological properties by the reaction after taking the drug. Furthermore, the use of non-invasive imaging diagnosis techniques (such as SPECT) is also an essential reference.
In earlier nuclear medicine 123I-β-CIT was used to detect dopamine transporters. Because of the drug’s uptake in the striatum, it had a significant correlation with the standard Unified Parkinson Disease Rating Scale (UPDRS—the comprehensive Parkinson’s disease rating scale is divided into four parts, namely patient psychology, daily living ability, motor function, and treatment complications) adopted to determine the PD stage [8]. However, the pharmacokinetics of 123I-β-CIT were very slow. It took up to 24 h to reach equilibrium after administering the drug, requiring more examination time. After that, other drugs were developed (such as 99mTc-TRODAT-1, 123I-FP-CIT, 123I-altropane) [9,10,11,12]. Although the imaging time can be shortened, the probability of non-specific binding was increased. Furthermore, the cost of the medicines became more expensive, so they could not be routinely applied in the clinic until the advent of 99mTc- TRODAT-1, which has high sensitivity and specificity for DaTscan. In this study, the PD stage was classified using the Hoehn and Yahr scale (HYS). According to the literature the HYS is highly correlated with UPDRS scores [13]. The following table lists information about the HYS scale (Table 1).
99mTc-TRODAT-1 is a successful preclinical dopamine-labeled drug [14,15]. 99mTc-TRODAT-1 could be used in SPECT brain imaging to evaluate a variety of neurological disorders in the brain (such as Parkinson’s disease, hemichorea-hemiballism, Tourette’s syndrome, multiple system atrophy, Wilson’s disease, depression, etc.) [16,17,18,19]. In addition 99mTc-TRODAT-1 has good performance to detect dopamine transporters in the striatum [20,21]. In the past, most scientists used the region of interest (ROI) method to calculate and extract features for the analysis and classification of SPECT dopamine transporter scans and then utilized statistical classification methods or machine learning methods to analyze the results [22,23,24]. This kind of method was limited by the selected ROI size and the extraction value of the characteristic parameters. This affects the disease stage determination result. The deep convolutional neural network (CNN) method was applied to the image classification area [25,26] by feeding a whole image to the deep network. This can effectively solve the problems in determining the ROI.
This study aimed to explore deep learning algorithm performance for Parkinson disease (PD) classification by SPECT imaging using the 99mTc-TRODAT-1 tracer method. Among deep learning algorithms, five different convolutional neural network architectures such as AlexNet, GoogLeNet, Residual Neural Network, VGG, and DenseNet were utilized to classify and model Parkinson’s disease stages. It is expected this will provide information to support clinical application and imaging diagnosis in the future.

2. Materials and Methods

2.1. Study Population

This experiment was a retrospective study. It collected the 99mTc-TRODAT-1 imaging and diagnostic reports archived in the Picture Archiving and Communication System (PACS) between March 2006 and August 2013. A total of 202 cases were collected. Among these were six healthy patients (three males and three females) and 196 patients with PD (80 males and 116 females). These patients were between the ages of 25 and 91 (Table 2).
According to the disease stage, the number of healthy control (HC) and PD’s HYS stage I to V patients were 6, 22, 27, 53, 87, and 7, respectively. In the study, the category was distinguished into healthy, early (HYS I, HYS II), mid (HYS III), and four late groups (HYS IV, HYS V). Then, five slice images from each patient was chosen for the analysis data set. The total number of images was 1010 (healthy (n = 30), early (HYS I~II, n = 110, 135), mid (HYS III, n = 265), late (HYS IV~V, n = 435, 35).
In Figure 1, “A” shows a healthy image. The white spot area showed high activity of striatum on both sides and had a curved bean shape. “B” corresponds to HYS I. The putamen in one-side of the striatum was decreased. “C” corresponds to HYS II. The putamen in two-sides of the striatum was decreased. “D” corresponded to HYS III. Two sides of the putamen show no activity. “E” corresponds to HYS IV. The two sides of the putamen display no activity and decreased activity on two sides of the caudate is seen. “F” corresponds to HYS V. There was no activity signal in the two sides of both the putamen and caudate.
Cases in which patients received an overall 99mTc-TRODAT-1 dose of 25 to 30 millicuries (mCi) and then underwent imaging within 2.5 to 4 h after administration were included in the study. However, patients who experienced head tremors were excluded, along with patients who received drug treatment during imaging since the drugs’ interference with the efficacy of 99mTc-TRODAT-1 would make outcomes unreliable. The Medical Ethics Committee of E-DA Hospital approved this clinical study (EMRP-100-054(RIII)). All patients signed an informed consent form prior to their participation.

2.2. The Imaging Conditions of 99mTc TRODAT-1 SPECT

A step-by-step (10 min, 32 steps) scan method was adopted to perform a DaTscan using 99mTc-TRODAT-1. The image matrix size was 128 × 128. A total of 64 images were captured at a collection rate of 25 s per image. Filtered back projection (FBP) was adopted for image reconstruction. The filter was a low-pass Butterworth filter with a cutoff frequency of 0.4 and an order of eight. A dual-head SPECT instrument (E. camTM Signature Series Fixed 180; Siemens Medical Solutions Inc., Malvern, PA, USA) equipped with the Siemens E. soft Workstation and a fan beam collimator was employed in this study. The field of view (FOV) of a single detection head was 53.3 × 38.7 cm2, and the diagonal FOV was 63.5 cm. A single detection head is equipped with 59 photomultiplier tubes in a hexagonal arrangement and uses 5/8 inch sodium iodide (NaI(Tl)) crystals with a crystal size of 59 × 44.5 cm2.

2.3. The Deep Learning Method Concept

Deep convolutional neural network technology has five primary layers which are named convolution layer, pooling layer, rectified linear unit (ReLU) layer, fully connected layers, and softmax layer. The above steps are performed to define the feature category that satisfies the input image. This study designed and established a PD staging classification model through the pre-trained network in the above manner. Moreover, it evaluated the model effectiveness with accuracy, recall, precision, F-score, and Kappa. NVIDIA GeForce GTX 1060 6 GB (Santa Clara, CA, USA) hardware was used to train the CNN in this study.

2.4. Popular Pre-trained Models in CNN

In this study, a total of 1010 DaTscan slice images was explored in the CNN analysis using five different pre-trained models. These were the AlexNet, GoogLeNet, Residual Neural Network, VGG, and DenseNet models. There were two types of image data sets: grayscale and pseudocolor images. For each subject only the maximum active slice of the striatum and two other slices, which were the previous and next one to the active slice were obtained in order to simulate how to diagnose and determine the disease in the clinic. Therefore, five images collected from each patient were used in the analysis dataset. The total number of images was 1010 (early (HYS I~II, n = 110, 135), mid (HYS III, n = 265), late (HYS IV~V, n = 435, 35) and healthy (n = 30)). Furthermore, 70% of the data was used for training and 30% of the data was used for verification. The overall structure of the study’s analytic system is illustrated in Figure 2. In Step 1, the 99mTc-Trodat-1 SPECT images, which were already separated into six categories were input. In Step 2, these SPECT images were colored by the pseudocolor technique. The pseudocolor technique was associated with a color table that defines the color displayed for each image pixel. In Step 3, the image augmentation method was used to increase the number of images. In Step 4, we start to train the network using the five pre-trained models. Before the training process begins, each SPECT image needs to be resized to match every model’s input size. In Step 4, the multiple stages of the PD classify model were established. Finally, in Step 5, the testing set data was used for statistical analysis and performance validation.

2.4.1. AlexNet

The architecture of the AlexNet network could be regarded as a large LeNet. The input image size was increased from 28 × 28 to 224 × 224 (RGB). It was divided into a five-layer convolution neural network in the front part and a traditional connection layer. There are a total of eight layers in AlexNet. The input of this architecture was an image and the output was the prediction classification and error rate. The AlexNet network improves the ReLU and pooling layers of the LeNet network by using ReLU to replace the previously used sigmoid and tanh functiond. It can solve the problem of gradient disappearance, make training more efficient and improving the accuracy. In the pooling layer, the LeNet network uses the average pooling method but the features will be blurred, whereas the AlexNet network uses the max-pooling method and chose a small step, which can avoid feature loss.
The AlexNet network uses two methods to reduce overfitting. First, a data augmentation technique [27] is used to increase the amount of training data to avoid overfitting during training, which is due to insufficient images in the training data set. The data augmentation technique includes image flip and horizontal and vertical moves to obtain a different view of the same image to provide more images for training. Second, the dropout method [28] is used to add a dropout in the fully connected layer. Each neuron with the same probability does not participate in the transfer of the next layer. This method makes the network force the current neurons for training that effectively reduce the overfitting. The dropout rate of the AlexNet was 0.5, which means each neuron had a 50% chance of not participating in the transmission of the next layer.
This study utilized a pre-trained AlexNet model. The input image size was 227 × 227, using five convolution, three max-pooling, seven ReLU and two fully connected layers, followed by a softmax output layer (about the convolution parameter and pooling layer as shown in Table 3). A total of 25 layers were discussed for gray and color images in four and six categories simultaneously. Figure 3 shows visually the convolution between first and final image.

2.4.2. GoogLeNet

GoogLeNet first appeared in ILSVRC14 competition and was the champion with a top-5 error of 6.67% [29]. GoogLeNet has a bold structure, with 22 layers in depth. Nevertheless, the parameter number was much smaller than in AlexNet. GoogLeNet has five million parameters and AlexNet has 60 million parameters, or twelve times more that GoogLeNet. Therefore, GoogLeNet is a better choice when memory or computing resources are limited.
Generally, the best way to improve network performance ss to increase the network’s depth and width. Depth refers to the number of network layers and width refers to the number of neurons. However, this method had three problems: (1) If the training data set is not enough and has too many training parameters, there will be overfitting. (2) If we design more layers to be included in the network, more parameters will be needed and the computation will be more complicated. (3) If the network becomes deeper, it will produce a gradient disappearance problem. To solve these problems, the solution is to reduce the parameters while increasing the depth and width of the network.
Google proposed the original inception module structure. The structure uses three convolution layers (1 × 1, 3 × 3, 5 × 5) and one pooling layer (using max pooling) for stacking to increase the network’s width. A ReLU function was required after each convolution layer to increase the non-linear characteristics. However, in the original inception version, the calculation for a 5 × 5 convolution kernel was too large, which caused the feature map dimensions to be too big. To avoid this situation, a 1 × 1 convolution kernel is added before the 3 × 3, 5 × 5 convolutions and after max pooling was produced. This could successfully reduce the feature map dimensions, increase the non-linear characteristics of the network and improve the expressive ability of the network. For example, if the output size of the previous layer was 28 × 28 × 192, after passing through 5 × 5 convolution with 32 channels the output size will become 28 × 28 × 32. Moreover, the parameters of the convolution layer were 192 × 5 × 5 × 32. After passing through the first 64 channels by 1 × 1 convolution and then 32 channels 5 × 5 convolution the output size was still 28 × 28 × 32, but the convolution parameters are reduced to (192 × 1 × 1 × 64) + (64 × 5 × 5 × 32). This is a significant reduction of the number of parameters.
This study utilized a pre-trained GoogLeNet model. The input image size was 224 × 224, using 58 convolutions (in the inception module it had 55 convolutions), 14 max-pooling layers (the inception module had nine pooling layers), and one layer fully connected by a softmax output layer. A total of 144 layers was used for gray and color images in four and six categories simultaneously. Visually the difference between the first and final convolution is shown in Figure 4.

2.4.3. ResNet

In the ILSVRC15 competition, the residual neural network (ResNet) technique won the championship with a top-5 error of 3.57% by using depth layers 152 [30]. Its characteristic was that the neural network did not necessarily need to be executed layer by layer. It could skip to the next layer through jumping. For a CNN, it is very important to identify the depth, but a deeper network will result in more complexity. The common reason is that the backpropagation is hard to update when the network is too deep. At the same time, this will cause the training speed to increase. Besides, it is found that the deep network causes degradation problems. This study utilized a pre-trained ResNet50 model. The input image size was 224 × 224 ResNet50 and ResNet101. (1) ResNet50: using 53 convolution, one max-pooling, and one fully connected layer, followed by a softmax output layer. A total of 117 layers was used for gray and color images in four and six categories simultaneously. The difference between the first and final convolution is visualized in Figure 5 and Figure 6.

2.4.4. VGG

VGG is the abbreviation for the Visual Geometry Group of Oxford University in the United Kingdom. Its main contribution is the use of more hidden layers and more training images, which can improve the accuracy to 90%. The VGG network can be divided into VGG16 and VGG19, which have 16 layers (13 convolution layers and three full connection layers) and 19 layers (16 convolution layers and three full connection layers), respectively [31]. The architecture is described below.
Compared with AlexNet, VGG16 uses several consecutive 3 × 3 convolutions to replace the larger convolutions (11 × 11, 7 × 7, 5 × 5) of the former. Given a local size picture as input multiple small convolutions are used because multiple non-linear layers can increase the network’s depth, which not only ensures the complexity of learning but spends fewer resources. In VGG, three 3 × 3 convolutions were used to substitute for a 7 × 7 convolution, and two 3 × 3 convolutions were used to substitute for a 5 × 5 convolution. The purpose was to ensure that the depth and effect of the network were improved under the same conditions. For example, three 3 × 3 convolutions with a step equal to 1 can be regarded as an input image with a 7 × 7 size in layer operation. The total number of parameters was 27 × C2 (C is the number of input and output channels). Suppose we use a 7 × 7 convolution kernel, then the total number of parameters is 49 × C2. The former not only reduces the parameters but also better maintains the image properties.
The main advantage of VGG is that its structure is straightforward. The entire network uses the same size of convolution (3 × 3) and maximum pooling size (2 × 2). The combination of several small convolution layers (3 × 3) was better than one large convolution layer (5 × 5, 7 × 7). It was verified that the continuously deep network structure could improve the performance. The disadvantage of VGG is that it consumes more computing resources and uses more memory. Most of the parameters are in the fully connected layer, and VGG has three fully connected layers.
This study utilized the VGG19 pre-trained model. The input image size was 224 × 224, using 16 convolution, five max-pooling, and three fully connected layers, followed by a softmax output layer. A total of 47 layers was used for the gray and color images in four and six categories simultaneously. In visualize the result the difference between the first and final convolution is shown in Figure 7.

2.4.5. DenseNet

DenseNet is similar to ResNet. The difference between DenseNet and ResNet is that ResNet calculates summations, but DenseNet calculates by stitching. The input of each layer network includes the previous output. For example, the Lth layer’s input is equal to k0+k(L-1), where k is the number of channels. DenseNet improves the transmission efficiency of information and gradients in the network. Each layer can directly get the gradient from the loss function and get the input signal directly. In this way, the network can be trained deeper. This structure also has the effect of regularization. Relatively, other networks are committed to improving network performance based on the depth and width [32]. The characteristic of DenseNet is that the features trained in each layer are provided repeatedly for use in subsequent layers. This greatly improves the feature utilization rate. The advantages of DenseNet are: (1) It alleviates the vanishing-gradient problem. (2) It strengthens the spread and reuse of features. (3) It significantly reduces the number of parameters. This study utilized the DenseNet201 pre-trained model. The input image size was 224 × 224, 200 convolutions, one max-pooling, and one fully connected layer, followed by a softmax output layer. A total of 709 layers was used for the gray and color images in four and six categories simultaneously. The difference between a first and final convolution is visualized in Figure 8.
This study discusses five commonly used pre-trained models for Parkinson’s disease multistage classification. Table 4 lists the parameters of these six models, including the input image size, depth, layer number, model memory size, and training time based on a batch size of 10 when the epoch was 1.

3. Experimental Results

This study was utilized five pre-trained models—AlexNet, GoogLeNet, Residual Neural Network, VGG, and DenseNet—to classify Parkinson’s disease stages. A total of 202 DaTscan cases were collected in this study. They are separated into two groups. One is healthy, early, mid, and late PD (four categories). The other is healthy and HYS I~V (six categories). There were two kinds of images in the deep CNN training data sets: grayscale and pseudocolor images in order to simulate the clinical diagnosis and determine the disease stage. Only the striatum’s maximum active slice and the active slice’s previous and next two images were obtained for each subject. Therefore, the analysis data set contains five images collected from each patient. The total number of images was 1010 (early (HYS I~II, n = 110, 135), mid (HYS III, n = 265), late (HYS IV~V, n = 435, 35), healthy (n = 30), and then 70% of the data in the data set used for training and 30% of the data used for verification. Due to the fact the number of images in each category was unbalanced an imported image augmentation method (such as cropping, rotating, resizing, translating, and flipping) was applied (Figure 9).
This study uses recall, precision, F-score, accuracy, and Kappa to evaluate the performance of the classification models. Recall means the percentage of positive predictions in the total of positive cases. Precision means how many true positive cases are in the total number of positive predictions. When the recall is low, it may not be possible to judge the category, but it will not be misjudged if the model provides a result. When the precision is low, it means the category cannot be determined correctly. For both evaluation indexed bigger is better. If a result has high precision and low recall, it means the model might be too cautious and almost useless for prediction. If a result has high recall and low precision it means the model might produce more misjudgment results. F-score is the harmonic means between recall and precision. One can utilize this index to roughly evaluate model performance. Accuracy is the total classification performance of the model. Kappa is used to evaluate the agreement of the classification results compared with real cases.
Table 5 shows each pre-trained models’ deep CNN results divided between grayscale and color images in four categories: healthy, early, mid, and late PD. As can be seen from the data AlexNet had the best performance on the grayscale images with accuracy, recall, precision, F-score, and Kappa values of 0.825, 0.753, 0.874, 0.809, and 0.725, respectively, while DenseNet201 had the best performance on the color images with accuracy, recall, precision, F-score, and Kappa values of 0.855, 0.821, 0.903, 0.860, 0.724, respectively.
Table 6 shows the deep CNN results of each pre-trained models between grayscale and color images in four categories: healthy and Parkinson disease stages I to V. AlexNet had the best performance on the grayscale image with accuracy, recall, precision, F-score, and Kappa values of 0.774, 0.742, 0.853, 0.794, and 0.679, respectively, whereas DenseNet201 had the best performance on the color images with accuracy, recall, precision, F-score, and Kappa values of 0.778, 0.696, 0.814, 0.750, 0.680, respectively.
In terms of accuracy, the six pre-trained models were divided into four and six stages for PD. The classification performance using grayscale images was best with AlexNet (four stages) and VGG19 (six stages). In addition, the classification performance using pseudocolor images was best with DenseNet for both the four and six stages (Figure 10).

4. Discussion

Nuclear medical imaging is a kind of functional imaging. The images mainly show the intensity of the radiopharmaceutical activity of the organ. In this study’s deep learning design, using a pre-trained model could reduce the time spent on the CNN model development, but the use of the pre-trained model needs to conform to the model architecture. Furthermore, transfer learning needs to be carried out before training. Two points in the model must be revised: (1) The image size of the source data needs to be adjusted and (2) The number of output categories needs to be corrected. Besides, it can obtain a better model during training by changing parameters such as batch size, epoch and learning rate. Batch size refers to how many data pieces were used to calculate each iteration, and epoch represents the training sample run repeatedly.

4.1. Comparison between Published Literature Methods and the Presented Method

PD patients are commonly into two normal and abnormal groups (Table 7). There are fewer articles that discuss multi-class classification. Therefore, this study tried to use two kinds of techniques to explore multi-class classification. Finally, we successfully found a CNN model to classify the healthy and HYS I~V stages of PD diseases and obtained high accuracy. However, with the CNN model it was more difficult to explain the characteristics adopted by the model and the correlation between the characteristics and PD disease stages.

4.2. The Related Literature on Multiple Stages Classification in Medical Images

Regarding the classification of multi-stage diseases in medical images, Farooq et al. published in 2017 an article titled, “A deep CNN-based multi-class classification of Alzheimer’s disease using MRI” [33]. The authors divided Alzheimer’s disease into four groups: Alzheimer’s, mild cognitive impairment, late mild cognitive impairment and healthy persons. Then by jusing a pre-trained model, they successfully achieved a high classification accuracy rate of 98.8%.
In 2019, Talo et al. published the article “Convolutional neural networks for multi-class brain disease detection using MRI images” [34], which classified five class of MRI images of the brain, namely normal and cerebrovascular, neoplastic, degenerative and inflammatory diseases categories. It also used a pre-trained model and obtained 95.23% classification accuracy. In 2020, Ramzan et al. published “A Deep Learning Approach for Automated Diagnosis and Multi-Class Classification of Alzheimer’s Disease Stages Using Resting-State fMRI and Residual Neural Networks” [35]. It divided the disease into six categories: cognitively normal, significant memory concern, early mild cognitive impairment, mild cognitive impairment, late mild cognitive impairment, and Alzheimer’s disease. Moreover, using the pre-trained model plus transfer learning they obtained the accuracy of 97.92%. At the same time, the conclusiond mentioned that the pre-trained model could be used for multi-stage classification, which is consistent with this study’s conclusion.

4.3. The Presented Deep Learning Method

The images collected for this paper was a DICOM image file after DaTscan. There are two approaches in the literature on deep learning. One is to select a slice image of the striatum for modeling. The other one is to use the entire group of brain images for modeling. This study combines the advantages of both the above methods to design and select a striatum slice and choose the other two slices before and after the striatum. Then we assemble these five slices as an image data set. This can increase the amount of data in the image set and eliminate the noise of the external striatum. This method also corresponds to how clinicians perform interpretation and disease diagnosis. Many studies have pointed out that more training data could provide higher accuracy and stability [36]. However, more data takes more time to train the model and hardware equipment support is required. If the hardware performance is low the training time will be longer. It will be impossible for model training and modeling (due to problems like insufficient memory) in some cases (as shown in Table 8).

5. Conclusions

The assessment of pathophysiological changes using SPECT imaging could be essential for the diagnosis of Parkinson’s disease. The purpose of this study was to classify the multiple stages of PD by using pre-trained deep CNN models. In deep learning, the five CNN models Alex Net, GoogLeNet, Residual Neural Network, VGG, and DenseNet were used for modeling and to determine the accuracy of distinguishing four and six categories in PD staging. When performing deep learning, there were a total of 829 PD images. 70% of the data was used as the training set and 30% of the data was used for verification. The original grayscale image and the pseudo color technique to produce color images were compared with the classification results. The following is a summary of the conclusions obtained in this study.
  • When using deep convolutional neural network technology to classify 99mTc-Trodat-1 PD images for the original grayscale images processed through five pre-trained models (AlexNet, GoogLeNet, VGG19, ResNet, DenseNet201) the highest accuracy was 0.83 for AlexNet. In six categories (healthy, HYS I~V), the best accuracy was 0.78 obtained by VGG19 in four categories (healthy, early, mid, late);
  • For color images, DenseNet201 yielded the highest accuracy of 0.85 in four categories. In six categories, the highest accuracy was 0.78 also obtained using DenseNet201;
  • Overall, the pre-trained models could produce accurate results when using grayscale images. In this case, the pseudocolor images might be non-essential;
  • CNN could obtain high classification accuracy in multiple categories of SPECT PD scans;
  • However, the establishment of the CNN classification model was very time-consuming, and the results had low interpretability in clinic.

Author Contributions

T.-B.C. and W.-C.D. supervised the research and contributed to the manuscript organization. Y.-C.W. and H.-Y.C. contributed to in situ equipment and monitoring data acquisition. Y.-H.H. (Yung-Hui Huang), W.-H.T. and L.-R.Y. gave partially revised suggestions and comments on this manuscript. Y.-H.H. (Yun-Hsuan Hsu) and M.-C.L. acquired image database. S.-Y.H. conceived and developed the research framework, performed the experiments, analyzed the data, and wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors would like to thank the Department of Nuclear Medicine, E-DA Hospital for partially financially supporting this research under Contract No. EDAHP104023, No. EDAHP104011, No. EDAHP105017, No. EDAHP106015. and No. EDAHP109032. Also like to thank the MOST in Taiwan for partial supporting in this study with Contract No:106-2118-M-214-001 and 109-2118-M-214-001.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Popa, T.; Ibanez, L.; Levy, E.; White, A.; Bruno, J.; Cleary, K. Tumor volume measurement and volume measurement comparison plug-ins for VolView using ITK. In Proceedings of the Medical Imaging, MIUA 2017, Edinburgh, UK, 11–13 July 2017; pp. 61411B–61418B. [Google Scholar]
  2. Dutour, A.; Monteil, J.; Paraf, F.; Charissoux, J.L.; Kaletta, C.; Sauer, B.; Naujoks, K.; Rigaud, M. Endostatin cDNA/cationic liposome complexes as a promising therapy to prevent lung metastases in osteosarcoma: Study in a human-like rat orthotopic tumor. Mol. Ther. 2005, 11, 311–319. [Google Scholar] [CrossRef]
  3. Sun, H. An improved positron emission tomography (PET) reconstruction of 2D activity distribution using higher order scattered data. Mater. Sci. 2016. Available online: http://hdl.handle.net/1993/31782 (accessed on 10 October 2020).
  4. Bailey, D.L.; Willowson, K.P. An evidence-based review of quantitative SPECT imaging and potential clinical applications. J. Nucl. Med. 2013, 54, 83–89. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Vandervoort, E.; Celler, A.; Harrop, R. Implementation of an iterative scatter correction, the influence of attenuation map quality and their effect on absolute quantitation in SPECT. Phys. Med. Biol. 2007, 52, 1527. [Google Scholar] [CrossRef] [PubMed]
  6. Huang, W.-S.; Lee, M.-S.; Lin, J.-C.; Chen, C.-Y.; Yang, Y.-W.; Lin, S.-Z.; Wey, S.-P. Usefulness of brain 99m Tc-TRODAT-1 SPET for the evaluation of Parkinson’s disease. Eur. J. Nucl. Med. Mol. Imaging 2004, 31, 155–161. [Google Scholar] [CrossRef] [PubMed]
  7. Galvan, A.; Devergnas, A.; Wichmann, T. Alterations in neuronal activity in basal ganglia-thalamocortical circuits in the parkinsonian state. Front. Neuroanat. 2015, 9, 5. [Google Scholar] [CrossRef] [Green Version]
  8. Pirker, W. Correlation of dopamine transporter imaging with parkinsonian motor handicap: How close is it? Mov. Disord. 2003, 18, S43–S51. [Google Scholar] [CrossRef]
  9. Benamer, H.T.; Patterson, J.; Wyper, D.J.; Hadley, D.M.; Macphee, G.; Grosset, D.G. Correlation of Parkinson’s disease severity and duration with 123I-FP-CIT SPECT striatal uptake. Mov. Disord.Off. J. Mov. Disord. Soc. 2000, 15, 692–698. [Google Scholar] [CrossRef]
  10. Booij, J.; Speelman, J.D.; Horstink, M.W.; Wolters, E.C. The clinical benefit of imaging striatal dopamine transporters with [123 I] FP-CIT SPET in differentiating patients with presynaptic parkinsonism from those with other forms of parkinsonism. Eur. J. Nucl. Med. 2001, 28, 266–272. [Google Scholar] [CrossRef]
  11. Gleave, J.A.; Farncombe, T.H.; Saab, C.; Doering, L.C. Correlative single photon emission computed tomography imaging of [123I] altropane binding in the rat model of Parkinson’s. Nucl. Med. Biol. 2011, 38, 741–749. [Google Scholar] [CrossRef]
  12. Gharibkandi, N.A.; Hosseinimehr, S.J. Radiotracers for imaging of Parkinson’s disease. Eur. J. Med. Chem. 2019, 166, 75–89. [Google Scholar] [CrossRef]
  13. Hoehn, M.M.; Yahr, M.D. Parkinsonism: Onset, progression, and mortality. Neurology 1998, 50, 318. [Google Scholar] [CrossRef]
  14. Lu, C.-S.; Weng, Y.-H.; Chen, M.-C.; Chen, R.-S.; Tzen, K.-Y.; Wey, S.-P.; Ting, G.; Chang, H.-C.; Yen, T.-C. 99mTc-TRODAT-1 imaging of multiple system atrophy. J. Nucl. Med. 2004, 45, 49–55. [Google Scholar] [PubMed]
  15. Trott, C.M.; El Fakhri, G. Sequential and simultaneous dual-isotope brain SPECT: Comparison with PET for estimation and discrimination tasks in early Parkinson disease. Med. Phys. 2008, 35, 3343–3353. [Google Scholar] [CrossRef] [Green Version]
  16. Hung, G.-U.; Chiu, P.-Y. The Value of 99mTc-Trodat-1 SPECT for Discriminating Dementia with Lewy Bodies and Alzheimer’s disease. J. Nucl. Med. 2017, 58, 1279. [Google Scholar]
  17. Maciel, R.; Maia, D.; de Lima, C.; Cardoso, F. Evidence of striatal dopaminergic dysfunction Sydenham’s chorea in remission with (99m) tc-trodat-1 Spect. Mov. Disord. 2016, 31, S351. [Google Scholar]
  18. Lee, L.-T.; Tsai, H.C.; Chi, M.H.; Chang, W.H.; Chen, K.C.; Lee, I.H.; Chen, P.S.; Yao, W.J.; Chiu, N.T.; Yang, Y.K. Lower availability of striatal dopamine transporter in generalized anxiety disorder: A preliminary two-ligand SPECT study. Int. Clin. Psychopharmacol. 2015, 30, 175–178. [Google Scholar] [CrossRef]
  19. Tripathi, M.; Arora, G.; Das, C.J.; Grover, T.; Gupta, R.; Bal, C. Incidental detection of intracranial tuberculomas on (99m) Tc-TRODAT-1 SPECT/CT. Clin. Nucl. Med. 2015, 40, e321–e322. [Google Scholar] [CrossRef]
  20. Lee, J.-D.; Chu, Y.-H.; Chen, C.-W.; Lin, K.-J. Multi-image registration for evaluating the 99mTc-TRODAT-1 of Parkinson’s Rat Model. In Proceedings of the Engineering in Medicine and Biology Society, 2009, EMBC 2009, Annual International Conference of the IEEE, Minneapolis, MN, USA, 3–6 September 2009; pp. 5801–5804. [Google Scholar]
  21. Bernsen, M.R.; Vaissier, P.E.; Van Holen, R.; Booij, J.; Beekman, F.J.; de Jong, M. The role of preclinical SPECT in oncological and neurological research in combination with either CT or MRI. Eur. J. Nucl. Med. Mol. Imaging 2014, 41, 36–49. [Google Scholar] [CrossRef] [Green Version]
  22. Kijewski, M.; El Fakhri, G.; Moore, S. Performance of simultaneous and sequential Tc-99m/I-123 SPECT imaging in estimation of striatal activity. In Proceedings of the 2003 IEEE Nuclear Science Symposium Conference Record, Portland, OR, USA, 19–25 October 2003; pp. 3134–3136. [Google Scholar]
  23. Acton, P.D.; Choi, S.-R.; Plössl, K.; Kung, H.F. Quantification of dopamine transporters in the mouse brain using ultra-high resolution single-photon emission tomography. Eur. J. Nucl. Med. Mol. Imaging 2002, 29, 691–698. [Google Scholar] [CrossRef]
  24. Wu, Y.-C.; Peng, G.-S.; Cheng, C.-A.; Lin, C.-C.; Huang, W.-S.; Hsueh, C.-J.; Lee, J.-T. 99mTc-TRODAT-1 and 123I-IBZM SPECT studies in a patient with extrapontine myelinolysis with parkinsonian features. Ann. Nucl. Med. 2009, 23, 409–412. [Google Scholar] [CrossRef] [PubMed]
  25. Kiryu, S.; Yasaka, K.; Akai, H.; Nakata, Y.; Sugomori, Y.; Hara, S.; Seo, M.; Abe, O.; Ohtomo, K. Deep learning to differentiate parkinsonian disorders separately using single midsagittal MR imaging: A proof of concept study. Eur. Radiol. 2019, 29, 6891–6899. [Google Scholar] [CrossRef] [PubMed]
  26. Ishii, N.; Mochizuki, Y.; Shiomi, K.; Nakazato, M.; Mochizuki, H. Spiral drawing: Quantitative analysis and artificial-intelligence-based diagnosis using a smartphone. J. Neurol. Sci. 2020, 411, 116723. [Google Scholar] [CrossRef] [Green Version]
  27. Cireşan, D.; Meier, U.; Schmidhuber, J. Multi-column deep neural networks for image classification. arXiv 2012. Available online: https://arxiv.org/abs/1202.2745 (accessed on 13 October 2004).
  28. Hinton, G.E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R.R. Improving neural networks by preventing co-adaptation of feature detectors. arXiv 2012. Available online: https://arxiv.org/abs/1207.0580 (accessed on 13 October 2004).
  29. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  31. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014. Available online: https://arxiv.org/abs/1409.1556 (accessed on 15 August 2020).
  32. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4700–4708. [Google Scholar]
  33. Farooq, A.; Anwar, S.; Awais, M.; Rehman, S. A deep CNN based multi-class classification of Alzheimer’s disease using MRI. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; pp. 1–6. [Google Scholar]
  34. Talo, M.; Yildirim, O.; Baloglu, U.B.; Aydin, G.; Acharya, U.R. Convolutional neural networks for multi-class brain disease detection using MRI images. Comput. Med Imaging Graph. 2019, 78, 101673. [Google Scholar] [CrossRef] [PubMed]
  35. Ramzan, F.; Khan, M.U.G.; Rehmat, A.; Iqbal, S.; Saba, T.; Rehman, A.; Mehmood, Z. A Deep Learning Approach for Automated Diagnosis and Multi-Class Classification of Alzheimer’s Disease Stages Using Resting-State fMRI and Residual Neural Networks. J. Med. Syst. 2020, 44, 37. [Google Scholar] [CrossRef]
  36. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 843–852. [Google Scholar]
  37. Prashanth, R.; Roy, S.D.; Mandal, P.K.; Ghosh, S. High-accuracy classification of Parkinson’s disease through shape analysis and surface fitting in 123I-Ioflupane SPECT imaging. IEEE J. Biomed. Health Inform. 2016, 21, 794–802. [Google Scholar] [CrossRef] [Green Version]
  38. Brahim, A.; Khedher, L.; Górriz, J.M.; Ramírez, J.; Toumi, H.; Lespessailles, E.; Jennane, R.; El Hassouni, M. A proposed computer-aided diagnosis system for Parkinson’s disease classification using 123 I-FP-CIT imaging. In Proceedings of the 2017 International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Fez, Morocco, 22–24 May 2017; pp. 1–6. [Google Scholar]
  39. Adeli, E.; Wu, G.; Saghafi, B.; An, L.; Shi, F.; Shen, D. Kernel-based joint feature selection and max-margin classification for early diagnosis of Parkinson’s disease. Sci. Rep. 2017, 7, 41069. [Google Scholar] [CrossRef]
  40. Rumman, M.; Tasneem, A.N.; Farzana, S.; Pavel, M.I.; Alam, M.A. Early detection of Parkinson’s disease using image processing and artificial neural network. In Proceedings of the 2018 Joint 7th International Conference on Informatics, Electronics & Vision (ICIEV) and 2018 2nd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Kitakyushu, Japan, 25–29 June 2018; pp. 256–261. [Google Scholar]
Figure 1. SPECT images correspond to HYS stage of PD ((A) = Healthy case, (BF) = HYS I~V). The Upper row is original grayscale images. The lower images have pseudocolor mapping.
Figure 1. SPECT images correspond to HYS stage of PD ((A) = Healthy case, (BF) = HYS I~V). The Upper row is original grayscale images. The lower images have pseudocolor mapping.
Molecules 25 04792 g001
Figure 2. The study flow chart in multiple stages classification of Parkinson’s disease.
Figure 2. The study flow chart in multiple stages classification of Parkinson’s disease.
Molecules 25 04792 g002
Figure 3. Visualization of the activation of the convolution layer in AlexNet (Left: first convolution, Right: final convolution).
Figure 3. Visualization of the activation of the convolution layer in AlexNet (Left: first convolution, Right: final convolution).
Molecules 25 04792 g003
Figure 4. Visualization of the activation of the convolution layer in GoogLeNet (Left: first convolution, Right: final convolution).
Figure 4. Visualization of the activation of the convolution layer in GoogLeNet (Left: first convolution, Right: final convolution).
Molecules 25 04792 g004
Figure 5. Visualization of the activation of the convolution layer in ResNet50 (Left: first convolution, Right: final convolution).
Figure 5. Visualization of the activation of the convolution layer in ResNet50 (Left: first convolution, Right: final convolution).
Molecules 25 04792 g005
Figure 6. Visualization of the activation of the convolution layer in ResNet101 (Left: first convolution, Right: final convolution).
Figure 6. Visualization of the activation of the convolution layer in ResNet101 (Left: first convolution, Right: final convolution).
Molecules 25 04792 g006
Figure 7. Visualization of the activation of the convolution layer in VGG19 (Left: first convolution, Right: final convolution)
Figure 7. Visualization of the activation of the convolution layer in VGG19 (Left: first convolution, Right: final convolution)
Molecules 25 04792 g007
Figure 8. Visualization of the activation of the convolution layer in DenseNet201 (Left: first convolution, Right: final convolution).
Figure 8. Visualization of the activation of the convolution layer in DenseNet201 (Left: first convolution, Right: final convolution).
Molecules 25 04792 g008
Figure 9. Example of image augmentation by reflection, rotation (−180, 180), X-axis translation (−3, 30), Y-axis translation (−3, 3).
Figure 9. Example of image augmentation by reflection, rotation (−180, 180), X-axis translation (−3, 30), Y-axis translation (−3, 3).
Molecules 25 04792 g009
Figure 10. The accuracy of the six pre-trained models on four and six stages of PD1.
Figure 10. The accuracy of the six pre-trained models on four and six stages of PD1.
Molecules 25 04792 g010
Table 1. The Hoehn and Yahr scale and symptoms.
Table 1. The Hoehn and Yahr scale and symptoms.
StageHoehn and Yahr Scale
1Unilateral involvement only usually with minimal or no functional disability.
2Bilateral or midline involvement without impairment of balance.
3Bilateral disease: mild to the moderate disability with impaired postural reflexes; physically independent.
4Severely disabling disease; still able to walk or stand unassisted.
5Confinement to bed or wheelchair unless aided.
Table 2. Demographic information of the dataset.
Table 2. Demographic information of the dataset.
ClassSubjectsSexAgeSlice Images
HC63 F, 3 M48 ± 14.730
HYS I2213 F, 9 M68 ± 16.4110
HYS II2715 F, 12 M69 ± 10.3135
HYS III5336 F, 17 M71 ± 9.8265
HYS IV8747 F, 40 M69 ± 10435
HYS V75 F, 2 M65 ± 11.535
Table 3. The parameter of convolution and pooling layer in AlexNet.
Table 3. The parameter of convolution and pooling layer in AlexNet.
ItemConv1Pool1Conv2Pool2Conv3Conv4Conv5Pool5
Filter size11 × 113 × 35 × 53 × 33 × 33 × 33 × 33 × 3
Stride42121112
Padding00201110
Table 4. Pre-trained models comparison. Experimental parameters: Batch size = 10; Epoch = 1
Table 4. Pre-trained models comparison. Experimental parameters: Batch size = 10; Epoch = 1
NameInput SizeLayersFile SizeTraining Time(s)
AlexNet227 × 22725227 MB25.4
GoogLeNet224 × 22414427 MB62.7
ResNet50224 × 22411496 MB138.4
ResNet101224 × 224347167 MB326.1
VGG19224 × 22447535 MB162.1
DenseNet201224 × 22470977 MB880.7
Table 5. The pre-trained model performance in four categories on the difference image format.
Table 5. The pre-trained model performance in four categories on the difference image format.
Category = 4 (Healthy, Early, Mid, and Late PD)
NetImage
Format
BatchsizeMaxEpochsAccuracyRecallPrecisionF-ScoreKappa
AlexNetGray80400.8250.7530.8740.8090.725
Color162810.6220.5040.7010.5870.397
GoogLeNetGray1470.6870.6730.7280.7000.518
Color34170.5860.4920.6720.5680.368
VGG19Gray20100.8190.7580.8700.8100.720
Color24120.8070.8080.8380.8230.707
ResNet50Gray840.7390.7290.7100.7190.607
Color840.8270.8370.7490.7910.743
ResNet101Gray630.7670.6910.6680.6790.656
Color1260.8310.8240.8570.8400.744
DenseNetGray1890.8070.7220.8430.7780.704
Color1680.8550.8210.9030.8600.724
Table 6. The pre-trained models performance in six categories on the difference image format.
Table 6. The pre-trained models performance in six categories on the difference image format.
Category = 6 (Healthy and Parkinson Disease Stages I to V.)
NetImage
Format
BatchsizeMaxEpochsAccuracyRecallPrecisionF-ScoreKappa
AlexNetGray96480.7740.7420.8530.7940.679
Color120600.5210.3490.4910.4080.282
GoogLeNetGray34170.6170.5320.6950.6030.439
Color28140.5560.3780.4780.4220.362
VGG19Gray24120.7780.6120.6650.6370.683
Color22110.7540.7460.7470.7470.669
ResNet50Gray20100.6810.560.7680.6470.537
Color1890.7540.6610.7580.7060.651
ResNet101Gray1680.7220.6020.6450.6230.603
Color1890.7700.6990.7770.7360.673
DenseNet201Gray1680.7620.6700.7390.7030.661
Color840.7780.6960.8140.7500.680
Table 7. Comparison with related works of PD classification.
Table 7. Comparison with related works of PD classification.
Author (year) [reference]CategorySample
Size
Method/# FeatureClassifierAccuracy
R. Prashanth et al. (2016) [37]Normal208 Machine learning/34SVM0.97
PD427
Abdelbasset Brahim et al. (2017) [38]Normal111Voxels as Features approach and Principal Component AnalysisSVM0.88
PD158
Ehsan Adeli et al. (2017) [39]Normal169Kernel-based FeatureSVM0.95
PD369
Mosarrat Rumman et al. (2018) [40]Normal100ROI detection and area calculationANN0.94
PD100
Presented Method: Deep CNN4 stageHealthy = 6
early = 245
mid = 265
late = 470
Popular Pre-trained models in CNNAlex Net
(grayscale)
0.83
DenseNet
(color)
0.85
6 stageHealthy =6
HYS1= 110
HYS2 = 135
HYS3 = 265
HYS4 = 435
HYS5 = 35
Popular Pre-trained models in CNNVGG19
(grayscale)
0.78
DenseNet
(color)
0.78
Table 8. Two sample sizes for PD stage classification demonstrated in VGG19 (hardware environment: NVIDIA GeForce GTX 1060 6GB).
Table 8. Two sample sizes for PD stage classification demonstrated in VGG19 (hardware environment: NVIDIA GeForce GTX 1060 6GB).
ImagesBatch SizeEpochTraining Time (s)Accuracy
3271010602.870.55
32720102942.460.56
32730103905.140.54
3274010GPU out of memory
6721010236.360.78
67220104537.950.77
67230108318.340.63
6724010GPU out of memory
Sample Availability: Samples of the study images all proved by the Medical Ethics Committee of E-DA Hospital approved this clinical study (EMRP-100-054(RIII)). If necessary, it can provide the original images data.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hsu, S.-Y.; Yeh, L.-R.; Chen, T.-B.; Du, W.-C.; Huang, Y.-H.; Twan, W.-H.; Lin, M.-C.; Hsu, Y.-H.; Wu, Y.-C.; Chen, H.-Y. Classification of the Multiple Stages of Parkinson’s Disease by a Deep Convolution Neural Network Based on 99mTc-TRODAT-1 SPECT Images. Molecules 2020, 25, 4792. https://doi.org/10.3390/molecules25204792

AMA Style

Hsu S-Y, Yeh L-R, Chen T-B, Du W-C, Huang Y-H, Twan W-H, Lin M-C, Hsu Y-H, Wu Y-C, Chen H-Y. Classification of the Multiple Stages of Parkinson’s Disease by a Deep Convolution Neural Network Based on 99mTc-TRODAT-1 SPECT Images. Molecules. 2020; 25(20):4792. https://doi.org/10.3390/molecules25204792

Chicago/Turabian Style

Hsu, Shih-Yen, Li-Ren Yeh, Tai-Been Chen, Wei-Chang Du, Yung-Hui Huang, Wen-Hung Twan, Ming-Chia Lin, Yun-Hsuan Hsu, Yi-Chen Wu, and Huei-Yung Chen. 2020. "Classification of the Multiple Stages of Parkinson’s Disease by a Deep Convolution Neural Network Based on 99mTc-TRODAT-1 SPECT Images" Molecules 25, no. 20: 4792. https://doi.org/10.3390/molecules25204792

Article Metrics

Back to TopTop