Next Article in Journal
AI-Powered Synthesis of Structured Multimodal Breast Ultrasound Reports Integrating Radiologist Annotations and Deep Learning Analysis
Next Article in Special Issue
A Longitudinal Model for a Dynamic Risk Score to Predict Delayed Cerebral Ischemia after Subarachnoid Hemorrhage
Previous Article in Journal
Effectiveness of Machine Learning in Predicting Orthodontic Tooth Extractions: A Multi-Institutional Study
Previous Article in Special Issue
Adaptive Detection in Real-Time Gait Analysis through the Dynamic Gait Event Identifier
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Transcranial Sonography-Based Classification of Parkinson’s Disease Using a Novel Dual-Channel CNXV2-DANet

1
Department of Biomedical Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China
2
Research Institute of Smart Ageing, The Hong Kong Polytechnic University, Hong Kong SAR, China
3
Department of Ultrasound, Beijing Tiantan Hospital, Capital Medical University, Beijing 100070, China
*
Authors to whom correspondence should be addressed.
Bioengineering 2024, 11(9), 889; https://doi.org/10.3390/bioengineering11090889
Submission received: 19 July 2024 / Revised: 29 August 2024 / Accepted: 29 August 2024 / Published: 31 August 2024

Abstract

:
Transcranial sonography (TCS) has been introduced to assess hyper-echogenicity in the substantia nigra of the midbrain for Parkinson’s disease (PD); however, its subjective and resource-demanding nature has impeded its widespread application. An AI-empowered TCS-based PD classification tool is greatly demanding, yet relevant research is severely scarce. Therefore, we proposed a novel dual-channel CNXV2-DANet for TCS-based PD classification using a large cohort. A total of 1176 TCS images from 588 subjects were retrospectively enrolled from Beijing Tiantan Hospital, encompassing both the left and right side of the midbrain for each subject. The entire dataset was divided into a training/validation/testing set at a ratio of 70%/15%/15%. Development of the proposed CNXV2-DANet was performed on the training set with comparisons between the single-channel and dual-channel input settings; model evaluation was conducted on the independent testing set. The proposed dual-channel CNXV2-DANet was compared against three state-of-the-art networks (ConvNeXtV2, ConvNeXt, Swin Transformer). The results demonstrated that both CNXV2-DANet and ConvNeXt V2 performed more superiorly under dual-channel inputs than the single-channel input. The dual-channel CNXV2-DANet outperformed the single-channel, achieving superior average metrics for accuracy (0.839 ± 0.028), precision (0.849 ± 0.014), recall (0.845 ± 0.043), F1-score (0.820 ± 0.038), and AUC (0.906 ± 0.013) compared with the single channel metrics for accuracy (0.784 ± 0.037), precision (0.817 ± 0.090), recall (0.748 ± 0.093), F1-score (0.773 ± 0.037), and AUC (0.861 ± 0.047). Furthermore, the dual-channel CNXV2-DANet outperformed all other networks (all p-values < 0.001). These findings suggest that the proposed dual-channel CNXV2-DANet may provide the community with an AI-empowered TCS-based tool for PD assessment.

1. Introduction

Parkinson’s disease (PD) is a rapidly growing, chronic, progressive, and devastating neurodegenerative disorder that has affected over 9.4 million individuals worldwide as of 2020 [1], and it has been recognized as one of the rapidly advancing neurological disorders [2,3]. Age is the primary risk factor for developing PD, and it is most commonly seen in individuals over the age of 60, with men being more susceptible than women at a prevalence ratio of approximately 3:2 [2,4]. Moreover, various modifiable environmental factors—such as pesticides and water pollutants—as well as lifestyle choices like smoking, coffee consumption, exercise, and head trauma, have all been implicated in the pathogenesis of PD across different populations [5,6]. Identifying the early stages of PD can be challenging, often leading to a significant delay between the emergence of noticeable symptoms and an official diagnosis (averaging around 10 years). Early indicators encompass a spectrum ranging from constipation to a diminished sense of smell, asymmetric shoulder pain, and even symptoms of depression [3]. PD can incur a range of debilitating motor symptoms including tremors at rest, rigidity, akinesia (or bradykinesia), and postural instability [7], as well as non-motor symptoms including autonomic dysfunction, neuropsychiatric symptoms, sensory symptoms, sleep disturbance, etc. [8]. These symptoms have posed grievous consequences to patients, their families and caregivers, and societies. Currently, studies are exploring the diagnosis of PD through various methods including voice, language, movement, and empirical mode decomposition [9,10,11]. However, PD is still considered an incurable condition [12,13]. Treatment aims primarily at slowing or stopping the progression of the disease. Conventional management strategies often initiate with dopamine replacement therapies, such as daily oral administration of carbidopa/levodopa or dopamine agonists [14,15]. Furthermore, regular exercise can enhance neuroprotection, thereby mitigating the advancement of the disease [16,17,18].
A growing body of literature has proven that the release of dendritic dopamine in the substantia nigra (SN) within the midbrain structure is crucial for controlling neuronal activity and behavior [19], and that iron ion accumulation in the SN is closely related to the development of PD and Parkinsonism [20]. With regard to this, the detection of iron ion accumulation in the SN of the midbrain plays a pivotal role in enabling early PD diagnosis [21].
The array of imaging modalities that have been deployed to detect PD [22] includes magnetic resonance imaging (MRI), positron emission tomography (PET), single-photon emission computerized tomography (SPECT), and transcranial sonography (TCS). MRI is radiation-free, has high image resolution, and is capable of quantifying iron ion levels by using specific image acquisition sequences, but it is of low accessibility and affordability to the general public, particularly in developing or under-developed countries, rendering it unfit for the population-wide screening of PD [23]. PET and SPECT imaging are capable of assessing striatal dopamine terminal dysfunction that supports the diagnosis of PD [24]. But they are costly, have low accessibility, and induce radiation hazards in patients. In contrast, TCS presents several distinct advantages over MRI, PET, and SPECT. This modality delivers a non-invasive, radiation-free, real-time, and rapid imaging approach, characterized by its remarkable accessibility, cost-effectiveness, and outstanding patient compliance. The population usually is financially demanding, so TCS is an affordable choice [25,26,27]. In particular, TCS has demonstrated its efficacy in establishing a correlation between the hyper-echogenicity region in the SN and the incidence of PD [28,29]. Recognizing the significance of TCS in PD management, numerous international guidelines have been extensively documented [30,31,32]. TCS imaging can reveal hyper-echogenicity regions within the SN in the midbrain, and regions characterized by an enlarged echogenic area in the SN, measuring between 0.20 and 0.25 cm2, occur in approximately 90% of PD. These imaging diagnosis methods can be considered as TCS-based PD diagnosis [27].
In clinical practice, the assessment of the SN region within the midbrain using TCS holds significant importance in the imaging-based diagnosis of PD and serves as a critical component in organizing the whole diagnostic workflow for PD [27,33]. However, the conventional practice of TCS-based diagnosis is entirely reliant on manual procedures, which presents two major drawbacks. First, it is highly time-consuming, which not only imposes considerable clinical burdens, especially in view of the growing demands due to the rapidly expanding aging population, but also necessitates experienced clinicians who are trained in the hyper-echogenicity regions of the SN in the midbrain to achieve accurate assessment results. This challenge could be overwhelming in resource-demanding regions, including developing and underdeveloped countries. Second, the conventional practice lies in the inherent subjectivity of manual assessment processes, which tends to introduce biases into the downstream grading and assessment of PD. Therefore, in the current era of artificial intelligence (AI) in medicine, deep learning-based PD diagnostic algorithms not only alleviate the burden on doctors but also reduce errors caused by subjectivity. Thus, there is an urgent need for an automated, efficient, and objective technique to perform TCS-based PD assessment, which is the first motivation of the research.
In this regard, several methods based on deep learning, especially convolutional neural networks (CNNs), have been applied to explore image-based examinations of PD using different modalities [34,35,36,37]. In 2019, Sivaranjini et al. developed a pre-trained AlexNet model via transfer learning on T2-weighted MR images to classify PD against healthy controls (HCs) using a dataset of 182 subjects, which achieved an accuracy of 88.9% [38]. In the same year, Manzanera et al. proposed an optimized approach employing Conv3D, batch normalization, and non-linearity within a 3D CNN framework. Their method was implemented on a dataset of 310 Fluorodeoxyglucose PET (FDG-PET) scans to classify PD patients against HCs, achieving an accuracy rate of 86.0% [39]. In 2020, Chakraborty et al. proposed a CNN architecture based on 3T T1-weighted MRI scans of 406 subjects acquired from the Parkinson’s Progression Markers Initiative (PPMI) database and achieved an overall accuracy of about 95.3% [37]. Zhao et al. proposed an architecture of a convolutional neural network and a greedy algorithm based on diffusion MRI acquired from a 3T scanner and achieved an accuracy of 80.7% using a PD dataset containing 432 observations [40]. Vyas et al. tested a 2D CNN model and a 3D CNN model and acquired 88.9% accuracy using 318 MRI scans [41]. Notably, one study employed a TCS dataset. In 2020, Shen et al. proposed an improved Deep Polynomial Network (DPN) algorithm, incorporating empirical kernel mapping (EKM), and evaluated it on a TCS dataset of 153 samples. Their method achieved an accuracy of approximately 86.95% [42]. Over the past decade, the aforementioned studies have emphasized the significance of leveraging deep learning techniques in the realm of automated PD diagnosis in patients.
Nevertheless, the contemporary era of AI has placed a growing emphasis on recruiting larger sample sizes for effective learning and including an independent test for model generalizability assessment. The only relevant AI research on TCS-based PD classification in the literature was conducted by Shen et al. However, the authors acknowledged that the sample size was limited because of the practical difficulty of obtaining large datasets in the field of neuroimaging science; therefore, the reported results may still suffer from model overfitting, and hence, their developed models may not be capable of providing sufficient values from clinical implementation perspectives. Moreover, their study did not include an independent testing set for assessing model generalizability, probably because of the insufficient sample size. Further studies using large sample sizes are warranted, which is the second motivation of the present research.
To address these challenges, this study aimed to develop and evaluate a novel deep neural network called dual-channel CNXV2-DANet using a large cohort of 1176 TCS images. The key contributions of the present work are three-fold. First, we employed a large, up-to-date TCS dataset to allow for more effective learning during model development and provide an independent testing set for assessing model generalizability using a naïve dataset. Second, for the first time, we integrated the attention mechanism module from the DANet network into the ConvNeXt V2 architecture, for the sake of enhancing the spatial attention of the network to place greater emphasis on the midbrain and SN regions during model development. Third, we investigated the impacts of single-channel and dual-channel inputs on the network’s learning effectiveness using both left and right sides of TCS images from each of the enrolled subjects, which has not been explored in the current body of literature. The present study is novel in nature, and the findings are expected to offer the community insights into AI-empowered TCS-based workflow for PD assessment, potentially expanding the capacity of TCS imaging for automated population-wide PD screening in community settings towards the smart-aging era.

2. Materials and Methods

2.1. Data

2.1.1. Data Acquisition

A total of 1176 TCS images from 588 subjects (each subject had a pair of TCS images; there were 307 TCS-based PD patients and 281 TCS-based non-PD patients) were retrospectively enrolled from Beijing Tiantan Hospital. This study was approved by the Ethics Board of Beiing Tiantan Hospital, Capita Medical University (No. KT2022-015-04). Informed consent was obtained from all the participants. Ethical approval was obtained from the Human Subject Ethics Sub-committee (HSESC) of the Hong Kong Polytechnic University (HSEARS20231102004). The original TCS images were obtained using a Canon Aplio i900 i-series ultrasound system (Canon Medical Systems Corporation, Otawara, Tochigi, Japan) equipped with an i6CX1 convex array transducer (center frequency = 2.6 MHz). An experienced physician manipulated the ultrasound probe over the left and right temporal windows of each subject, capturing an optimal ultrasound imaging frame of both sides that depicted the representative butterfly-shaped midbrain structures. All the TCS assessments were performed by two highly experienced physicians with over 10 years of experience, utilizing a TCS-based diagnostic approach. Patients diagnosed with PD using the TCS-based approach from either one side of the TCS images were annotated as TCS-based positive, while those not diagnosed with PD were annotated as TCS-based negative.

2.1.2. Image Pre-Processing

Prior to downstream analyses, a series of pre-processing procedures was conducted on the original TCS images of both the left and right sides of the temporal windows. As each subject contained a pair of TCS images (one for the left side and one for the right side), we need to fuse the two images and input them into the model. We developed two pre-processing pipelines, i.e., the single-channel method and the dual-channel method, and analyzed the performance of the studied neural networks. Figure 1 illustrates a schematic diagram for these two methods.
In the single-channel approach, all original TCS images with a resolution of 1280 × 960 pixels were cropped to a size of 192 × 384 pixels, focusing on the midbrain regions as the region of interest (ROI) for both the left (see Figure 1, showing the single-channel method, original image—left) and right (see Figure 1, showing the single-channel method, original image—right) sides of the images. These cropped images were then directly concatenated to create the single-channel input image of 384 × 384 pixels.
For the dual-channel method, the left (see Figure 1, showing the dual-channel method, original image—left) and right (see Figure 1, showing the dual-channel method, original image—right) sides of the original TCS images for each subject were cropped to 384 × 384 pixels as the ROI. These images were then integrated into two separate channels as the dual-channel input image. Both the single-channel and dual-channel methods effectively localized the midbrain region within the ROI while minimizing irrelevant information from other structures, thereby facilitating enhanced learning during subsequent model development.

2.2. Deep Learning Neural Networks

In this study, we proposed a neural network called CNXV2-DANet, which leverages the superiority of ConvNext v2 [43] while integrating the attention mechanism module from DANet [44] for the development of the auto-classification model. The attention mechanism model from DANet was added for the sake of promoting the network’s attention toward the midbrain and hyper-echogenicity regions of the SN during model development. The details of the networks are presented below.

2.2.1. ConvNext v2

ConvNext V2 [43] is a convolutional neural network (CNN) architecture that builds on ConvNeXt proposed by Facebook AI Research (FAIR), incorporating the advantages of ResNet [45] and EfficientNet [46]. ConvNext V2 introduces the following innovative components: a fully convolutional masked autoencoder (FCMAE) and a global response normalization (GRN). The FCMAE framework is a CNN-based self-supervised learning method. Its concept involves randomly masking certain regions on input images and then allowing the model to reconstruct the masked portions. This approach compels the model to learn both the global and local features of the images, thereby enhancing its generalization capability. The GRN layer is a novel CNN layer that normalizes feature maps across channels with the purpose of enhancing feature competition among channels by ensuring normalization.

2.2.2. DANet

DANet is a dual attention network designed to capture global feature dependencies in both spatial and channel dimensions, thereby enhancing the model’s attention weights for crucial locations or channels and improving the model’s performance [44]. Figure 2 visualizes the structure of DANet.
In the position attention module, feature map A undergoes initial processing through 3 convolutional layers to generate 3 feature maps (B, C, and D), which are subsequently reshaped. The transposed form of reshaped feature map B is then multiplied by reshaped feature map C, and the resulting product is subjected to a softmax operation to obtain the spatial attention map (78S). Moreover, reshaped feature map D is multiplied by the transposed S through matrix multiplication, scaled by the factor α, and reshaped back to its original form. Finally, reshaped feature map D is added to A, yielding the ultimate output E. Throughout this process, the weight scaling coefficient α is iteratively learned and adjusted to achieve the optimal value. In the channel attention module, feature map A is reshaped and transposed, and the resulting feature maps are multiplied together. The softmax operation is then applied to calculate channel attention map X. Subsequently, the transposed X is multiplied by reshaped feature map A using matrix multiplication. This product is scaled by the coefficient β and reshaped back to its original form. Finally, the reshaped result is added to A, yielding the final output F. The factor β is iteratively learned to obtain more suitable weights during the training process. To obtain a superior pixel-level prediction of feature representation, the outputs of the two attention modules are combined and aggregated by sum fusion [44].

2.2.3. The Proposed CNXV2-DANet

Figure 3 displays the architecture of CNXV2-DANet, which consists of the basis of ConvNext v2 and the attention mechanism module of DANet. Initially, the network receives the pre-processed images either from the single-channel method or the dual-channel method. Each input undergoes a 2D convolution and Layer Norm operation. Subsequently, the ConvNeXt V2 block is applied in 4 cycles, with a pattern of 3, 3, 27, and 3 convolutional layers, respectively, along with a downsampling step. The DANet module is inserted immediately before the global average pooling layer. Further, a downsampling process is performed, followed by a fully connected layer, generating the classification output.

2.3. Model Development and Comparison

2.3.1. Model Development

The entire dataset (n = 588) with 1176 TCS images was randomly divided into training, validation, and testing sets via a randomization seed, at an approximate ratio of 70% (n = 403), 15% (n = 95), 15% (n = 90), respectively. In the training set, CNXV2-DANet was utilized to develop a classification model under 5-fold cross-validation for the single-channel method and dual-channel method separately. Referring to Figure 1, a single channel input image contains a direct merge of left-side and right-side cropped TCS images as a single input to the network, whereas a dual channel input image contains two separate cropped TCS images as two separated inputs to the network.
For each of the two methods, 5 randomization seeds were deployed to obtain 5 sets of results for the sake of obtaining an averaged performance value for a fair model evaluation across different patient sub-populations. The approach of multiple randomized stratifications is commonly adopted [47,48]. An averaged performance of each model was reported in this study to assess model stability and to improve the validity of the findings by mitigating sampling bias stemming from random data partitioning. The training process was carried out on an in-house service machine equipped with an Nvidia RTX A6000 GPU card. For the CNXV2-DANet model parameter setting on the training dataset, the model was trained for 50 epochs with a learning rate of 0.00001. The cross-entropy loss function and a weight decay of 0.05 were utilized, while a batch size of 8 was used. Our models operated within the PyCharm (version 2021.3.1) integrated development environment (IDE), and all our experiments utilize CUDA version 12.5.40.

2.3.2. Model Comparison

In order to assess the superiority of the proposed CNXV2-DANet model over the existing state-of-the-art networks, ConvNeXt V2, ConvNeXt [49], and Swin Transformer [50] were individually employed for model comparisons. All the networks were trained using the same training dataset and pre-processing methods, incorporating 5 random seeds and 5-fold cross-validation.

2.4. Evaluation Metrics

All the developed models were evaluated on the same testing set; a confusion matrix and several evaluation parameters were applied in this study. For the confusion matrix, four elements were defined as follows: True Positive (TP), False Positive (FP), False Negative (FN), and True Negative (TN). For the evaluation parameters, accuracy, precision, recall, F1-score, and area under the ROC curve (AUC) were calculated [51]. The relevant definitions and equations are described below.
Accuracy measures the percentage of correctly classified samples among all samples. In multi-classification tasks, macro-averaged accuracy and micro-averaged accuracy are often used to assess performance. Accuracy is defined by Equation (1) below:
A c c u r a c y = ( T P + T N ) ( T P + T N + F P + F N )
Precision, also called the positive predictive value, measures the accuracy of positive predictions generated by a model. It calculates the ratio of TP predictions to the total number of predicted positive samples. A higher precision indicates that the model has a lower rate of FPs, making its positive predictions more reliable. This helps minimize incorrect identifications and reduces unnecessary actions or interventions based on FPs. Precision is defined by Equation (2) below:
P r e c i s i o n = T P ( T P + F P )
The recall rate quantifies the proportion of TP samples identified by a model out of all the true samples. A higher recall rate indicates that the model is capable of accurately identifying TP samples, thereby demonstrating superior discriminatory ability. It helps us understand the ability of the model to capture all maize disease cases, ensuring that the model does not miss any potential disease instances. Recall is defined by Equation (3) below:
R e c a l l = T P ( T P + F N )
The F1 score assesses the trade-off between accuracy and recall in a model’s performance. It provides a comprehensive evaluation of how well the model strikes a balance between accurately classifying samples and capturing all relevant instances. A higher F1 score signifies that the model has achieved a better equilibrium between accuracy and recall, indicating its proficiency in correctly identifying diseases and ensuring that no instances are overlooked. The F1-score is defined by Equation (4) below:
F 1 S c o r e = 2 ( P r e c i s i o n R e c a l l ) ( P r e c i s i o n R e c a l l )
AUC is a crucial metric for evaluating the performance of binary classification models. It measures the area under the ROC curve, which ranges from 0 to 1. A higher AUC indicates better model performance. To calculate AUC, the ROC curve was constructed using the predicted probabilities from the model and the true labels, and the area under the curve was computed. In addition, the DeLong method [52] was also employed to conduct a performance comparative analysis between the proposed CNXV2-DANet model and the other models, yielding calculations for covariance and p-values.

3. Results

3.1. Single-Channel vs. Dual-Channel Methods

3.1.1. CNXV2-DANet

Table 1 summarizes the quantitative comparisons of the proposed CNXV2-DANet model between the single-channel and dual-channel pre-processing methods for TCS-based PD classification on the testing dataset. The results of two pre-processing methods were obtained using five randomization seeds under five-fold cross-validation; the averaged evaluation metrics and standard deviation (SD) are reported. Compared with the single-channel method, the dual-channel approach resulted in remarkably higher classification performance, achieving averaged values for accuracy of 0.839 ± 0.028 (single-channel: 0.784 ± 0.037), precision of 0.849 ± 0.014 (single-channel: 0.817 ± 0.090), recall of 0.845 ± 0.043 (single-channel: 0.748 ± 0.093), F1-score of 0.820 ± 0.034 (single-channel: 0.773 ± 0.037), and AUC of 0.906 ± 0.013 (single-channel: 0.861 ± 0.047). Figure 4 shows more details of the results from one of the five randomization seeds.

3.1.2. ConvNeXt V2

An ablation experiment was conducted to remove the attention mechanism modules of DANet from the proposed CNXV2-DANet model. Table 2 summarizes the quantitative comparisons of the performance of ConvNeXt V2 between the single-channel and dual-channel pre-processing methods.
All the results were obtained through five-fold cross-validation using five randomization seeds on the same testing dataset, and the averaged values were recorded. Similar to the proposed CNXV2-DANet model, ConvNeXt V2 exhibited superior performance in the dual-channel method compared with the single-channel approach. The dual-channel approach led to remarkably improved classification performance of ConvNeXt V2, achieving averaged values for accuracy of 0.822 ± 0.073 (single-channel: 0.791 ± 0.041), precision of 0.858 ± 0.104 (single-channel: 0.775 ± 0.059), F1-score of 0.798 ± 0.074 (single-channel: 0.789 ± 0.046), and AUC of 0.896 ± 0.040 (single-channel: 0.856 ± 0.023), despite its slightly lower averaged recall of 0.775 ± 0.069 (single-channel: 0.813 ± 0.099).
It is worth noting that both CNXV2-DANet and ConvNeXt V2 achieved greater classification performance in the dual-channel setting compared with the single-channel setting (Table 1 and Table 2). Furthermore, the proposed CNXV2-DANet model generally outperformed the ConvNeXt V2 model in the averaged metrics for accuracy of 0.839 ± 0.028 (ConvNeXt V2: 0.822 ± 0.073), recall of 0.845 ± 0.043 (ConvNeXt V2: 0.775 ± 0.069), F1-score of 0.820 ± 0.034 (ConvNeXt V2: 0.798 ± 0.074), and AUC of 0.906 ± 0.013 (ConvNeXt V2: 0.896 ± 0.040), despite its slightly lower averaged precision of 0.849 ± 0.014 (ConvNeXt V2: 0.858 ± 0.104).

3.2. CNXV2-DANet vs. State-of-the-Art Networks on the Dual-Channel Setting

Table 3 summarizes the classification performance of the proposed CNXV2-DANet model, in comparison to ConvNeXt V2, ConvNeXt, and Swin Transformer. Among all the comparison networks, the proposed CNXV2-DANet model generally achieved the best classification performance, yielding averaged values for accuracy of 0.839 ± 0.028, precision of 0.849 ± 0.014, recall of 0.845 ± 0.043, F1-score of 0.820 ± 0.034, and AUC of 0.906 ± 0.013. This is followed by the ConvNeXt V2 model, while both ConvNeXt and Swin Transformer were the most under-performing networks, incurring tremendously inferior classification performance than the proposed CNXV2-DANet model. Figure 5 displays a comparison of the performance of the studied models on the testing dataset under the dual-channel setting, including the averaged metrics of accuracy, precision, recall, and F1-score with the corresponding standard deviations across the five randomizations (Figure 5a) and ROC curves under one of the randomization seeds (Figure 5b). This figure shows that the proposed CNXV2-DANet model achieved the highest AUC value of 0.917, followed by ConvNext v2 with 0.865, ConvNext with 0.855, and Swin Transformer with 0.842. Furthermore, the p-values between CNXV2-DANet and ConvNeXt V2, ConvNeXt, and Swin Transformer are all less than 0.001, indicating that CNXV2-DANet’s test performance significantly differs from the other three models.

4. Discussion

TCS-based PD diagnosis assumes a pivotal and indispensable role as the initial and fundamental step toward achieving early PD diagnosis with assessment methods. However, the existing assessment process is labor-intensive, time-consuming, and prone to subjectivity. This highlights the urgent need for an automated, efficient, and objective deep learning-based method to meet the growing demand of the community amid the rapidly growing aging population worldwide. Despite the numerous advantages of TCS, such as its non-invasive, radiation-free, rapid, low-cost, and highly accessible nature, high patient compliance, and ease of operation, there is a severe lack of relevant studies focusing on AI-driven TCS-based diagnosis. In light of this, we proposed CNXV2-DANet combined with the dual-channel approach for automated TCS-based PD classification. Our results indicated that the dual-channel approach yielded a great classification performance compared with the single-channel method, and the proposed CNXV2-DANet model outperformed the other comparison networks in a wide range of evaluation aspects. Thus, CNXV2-DANet potentially provides the community with a favorable and automated method for TCS-based PD assessment in the future.
Aligned with the present study, previous studies [28,42,53] conducted over the past decade have consistently emphasized the significance of TCS-based PD diagnosis, which has contributed valuable insights for the research community. Nonetheless, there is an increasing need to leverage the prosperous era of AI and large datasets for streamlining TCS-based PD diagnosis. In recent years, there has been a paucity of studies employing TCS imaging in PD classification-related studies, as shown in Table 4. Furthermore, the dataset we employed was comparatively larger. In 2020, Shen et al. proposed the deep polynomial network (DPN), the deep neural mapping large margin distribution machine (DNMLDM) algorithm, and the dropout and pruning empirical-kernel mapping-deep polynomial network (D-P-EKN-DPN) algorithm [42] for classification of PD with TCS images. Their method achieved an accuracy of 86.95%, sensitivity of 85.77%, specificity of 87.16%, positive predictive value (PPV) of 84.81%, negative predictive value (NPV) of 86.82%, and F1-score of 0.86. Their model was trained on a small cohort of fewer than 200 samples, while our study used a large cohort of 588 patients with 1176 TCS images. Their reported results are based on a single random train/test split, which may introduce sampling bias. Additionally, because of the limited sample size, an independent dataset was not available for validation. In contrast, our study adopted a larger cohort of data and conducted five randomization samplings for train/test partitioning to obtain an averaged performance, and the models were evaluated on an independent dataset, thereby offering a greater degree of model inference for naïve datasets in the real world. As shown in Table 1, by leveraging the dual-channel mode and employing the proposed CNXV2-DANet model, we achieved an average accuracy of 83.9% ± 2.8%, precision of 84.9% ± 1.4%, recall of 84.5% ± 4.3%, F1-score of 82.0% ± 3.4%, and AUC of 0.906 ± 0.013, compared with other comparing deep learning methods.
In terms of the impacts of single-channel and dual-channel settings, the performance of CNXV2-DANet and ConvNeXt V2 were investigated and analyzed. The findings revealed that the performance of the dual-channel configuration remarkably outperformed that of the single-channel configuration, by CNXV2-DANet, with accuracy (83.9% ± 2.8% vs. 78.4% ± 3.7%), precision (84.9% ± 1.4% vs. 81.7% ± 9.0%), recall (84.5% ± 4.3% vs. 74.8% ± 9.3%), F1-score (82.0% ± 3.8% vs. 77.3% ± 3.7%), and AUC (0.906 ± 0.013 vs. 0.861 ± 0.047) referred in Table 1, and by ConvNeXt V2, with accuracy (82.2% 7.3% vs. 79.1% ± 4.1%), precision (85.8% ± 10.4% vs. 77.5% ± 5.9%), F1-score (79.8% ± 7.4% vs. 78.9% ± 4.6%), and AUC (0.896 ± 0.040 vs. 0.856 ± 0.023), as shown in Table 2. We speculated that the superior performance of the dual-channel in TCS-based PD diagnosis could be attributed to the enhanced information and richer image information input provided by the dual-channel, which enriches the model’s understanding despite the potential presence of irrelevant or ineffective information. In theory, it is easy to understand why excessive redundancy in input data would have detrimental impacts on model performance. However, excessively compressed data, such as those in the single-channel method, can also adversely affect the model’s classification accuracy. Therefore, it is essential to consider a comprehensive approach that encompasses the selection of appropriate and efficient preprocessing methods.
Furthermore, the proposed CNXV2-DANet model introduces a novel enhancement by incorporating the DANet module with spatial and dual-attention mechanisms. The performance of the CNXV2-DANet model was compared with ConvNeXt V2 through an ablation experiment. The results consistently demonstrated that the proposed CNXV2-DANet model outperformed the ConvNeXt V2 model in both dual-channel and single-channel methods. On the grounds that the hyper-echogenicity in the SN region serves as a prominent marker for potential PD [20], the spatial attention mechanism from DANet was employed to enhance the model’s focus on these critical regions locally within the SN region of the midbrain. We speculated that this may be one of the possible reasons explaining the superiority of the proposed CNXV2-DANet model over the ConvNeXt V2, because the separated channels embedded in DANet may facilitate the complementary learning of the model between the left-side midbrain image and the right-side midbrain image of subjects who may exhibit hyper-echogenicity on their left-side, right-side, or both sides of their SN in midbrain; this capability is absent without the attention module of DANet.
Lastly, the performance of the proposed CNXV2-DANet model was further compared with several state-of-the-art CNN networks, including ConvNeXt V2, ConvNeXt [49], and Swin Transformer [50], under the dual-channel setting. As shown in Table 3, the results demonstrated that CNXV2-DANet achieved remarkably better overall performance, yielding averaged values for accuracy of 83.9%, recall of 84.5%, F1-score of 82.0%, and AUC, compared with other comparing networks. However, it is worth noting that ConvNeXt V2 exhibited a higher precision of 85.8%, surpassing CNXV2-DANet’s precision of 84.9%. Considering the research objective of this study, which aims to achieve effective screening of PD, the recall value may hold greater importance in screening PD patients for early diagnosis. Furthermore, we compared the results of CNXV2-DANet and ConvNeXt V2 as part of the results derived from the ablation experiments.
Despite the encouraging results, the present study has several shortcomings that require further investigation in the future. First, the results of this study were generated with a dataset collected from a single institution and the same scanner vendor, which may limit the generalizability of the benchmarked model in real-world clinical settings. A multi-center study with TCS data acquired from different vendors is warranted in the future in order to further validate the findings of this work. Second, although the sample size of 588 (1176 images) in this study was considerably larger than those in previous studies, where the sample sizes ranged from 40 to 130, further investigations using a larger cohort are preferred in the context of deep learning. Third, because of the retrospective nature of this study, although this study recruited highly experienced physicians (>10 years) in ultrasound imaging for generating the reference annotations, there may exist intra- or inter-rater variabilities in the dataset. Therefore, addressing this issue—for example, by expanding the dataset, incorporating multi-center approaches, involving more experienced physicians, and integrating multimodal analyses with a broader range of patient data—in a prospective study is essential for promoting the widespread adoption of the developed model in clinical practice.

5. Conclusions

The proposed CNXV2-DANet model, which is configurated based on ConvNeXt V2 and the DANet module, outperformed several existing state-of-the-art networks including ConvNeXt V2, ConvNeXt, and Swin Transformer. Furthermore, the dual-channel approach yielded a remarkably greater PD classification performance than the single-channel method. The findings of this study demonstrated that the proposed CNXV2-DANet model combined with the dual-channel setting has the potential to provide the community with an automated and effective tool for early PD screening using non-invasive, radiation-free, low-cost TCS scanning. Upon resolving the abovementioned shortcomings, the findings of this study may provide valuable insights for clinical practitioners to implement effective and efficient TCS-based PD screening in the smart-aging era, ultimately benefiting countless sufferers worldwide in the long run.

Author Contributions

Conceptualization, H.K., S.-k.L., Y.S., S.L. and Y.-p.Z.; methodology, H.K. and X.W.; software, H.K. and X.W.; validation, H.K.; formal analysis, H.K.; investigation, S.-k.L. and H.K.; resources, W.Z. and Y.-p.Z.; data curation, Y.S., X.S., F.L., C.H. and W.Z.; writing—original draft preparation, H.K.; writing—review and editing, S.-k.L. and Y.-p.Z.; visualization, H.K.; supervision, S.-k.L., W.Z. and Y.-p.Z.; project administration, W.Z. and Y.-p.Z.; funding acquisition, Y.-p.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Research Institute for Smart Ageing of the Hong Kong Polytechnic University [1-CD5B, 1-YWE0] and the Innovation and Technology Fund of Hong Kong [MRP/022/18X].

Institutional Review Board Statement

Ethical approval for this study was obtained from the Human Subject Ethics Sub-committee (HSESC) of the Hong Kong Polytechnic University (HSEARS20231102004).

Informed Consent Statement

This study was approved by the Ethics Board of Beiing Tiantan Hospital, Capita Medical University (No. KT2022-015-04). Informed consent was obtained from all the participants.

Data Availability Statement

The datasets analyzed during the current study are not publicly available due to patients privacy protection purposes, but are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Maserejian, N.; Vinikoor-Imler, L.; Dilley, A. Estimation of the 2020 Global Population of Parkinson’s Disease (PD). Mov. Disord. 2020, 35, S79–S80. [Google Scholar]
  2. Feigin, V.L.; Nichols, E.; Alam, T.; Bannick, M.S.; Beghi, E.; Blake, N.; Culpepper, W.J.; Dorsey, E.R.; Elbaz, A.; Ellenbogen, R.G.; et al. Global, regional, and national burden of neurological disorders, 1990-2016: A systematic analysis for the Global Burden of Disease Study 2016. Lancet Neurol. 2019, 18, 459–480. [Google Scholar] [CrossRef] [PubMed]
  3. Bloem, B.R.; Okun, M.S.; Klein, C. Parkinson’s disease. Lancet 2021, 397, 2284–2303. [Google Scholar] [CrossRef] [PubMed]
  4. Tolosa, E.; Garrido, A.; Scholz, S.W.; Poewe, W. Challenges in the diagnosis of Parkinson’s disease. Lancet Neurol. 2021, 20, 385–397. [Google Scholar] [CrossRef] [PubMed]
  5. Poewe, W.; Seppi, K.; Tanner, C.M.; Halliday, G.M.; Brundin, P.; Volkmann, J.; Schrag, A.E.; Lang, A.E. Parkinson disease. Nat. Rev. Dis. Primers 2017, 3, 17013. [Google Scholar] [CrossRef]
  6. Kalia, L.V.; Lang, A.E. Parkinson’s disease. Lancet 2015, 386, 896–912. [Google Scholar] [CrossRef]
  7. Jankovic, J. Parkinson’s disease: Clinical features and diagnosis. J. Neurol. Neurosurg. Psychiatry 2008, 79, 368–376. [Google Scholar] [CrossRef]
  8. DeMaagd, G.; Philip, A. Parkinson’s Disease and Its Management: Part 1: Disease Entity, Risk Factors, Pathophysiology, Clinical Presentation, and Diagnosis. Pharm. Ther. 2015, 40, 504–532. [Google Scholar]
  9. Safi, K.; Aly, W.H.F.; AlAkkoumi, M.; Kanj, H.; Ghedira, M.; Hutin, E. EMD-Based Method for Supervised Classification of Parkinson’s Disease Patients Using Balance Control Data. Bioengineering 2022, 9, 283. [Google Scholar] [CrossRef]
  10. Mondol, S.I.M.M.R.; Kim, R.; Lee, S.M. Hybrid Machine Learning Framework for Multistage Parkinson’s Disease Classification Using Acoustic Features of Sustained Korean Vowels. Bioengineering 2023, 10, 984. [Google Scholar] [CrossRef]
  11. Ibarra, E.J.; Arias-Londoño, J.D.; Zañartu, M.; Godino-Llorente, J.I. Towards a Corpus (and Language)-Independent Screening of Parkinson’s Disease from Voice and Speech through Domain Adaptation. Bioengineering 2023, 10, 1316. [Google Scholar] [CrossRef] [PubMed]
  12. Connolly, B.S.; Lang, A.E. Pharmacological Treatment of Parkinson Disease A Review. JAMA J. Am. Med. Assoc. 2014, 311, 1670–1683. [Google Scholar] [CrossRef]
  13. Armstrong, M.J.; Okun, M.S. Diagnosis and Treatment of Parkinson Disease A Review. JAMA J. Am. Med. Assoc. 2020, 323, 548–560. [Google Scholar] [CrossRef]
  14. Hall, M.F.E.; Church, F.C. Integrative Medicine and Health Therapy for Parkinson Disease. Top. Geriatr. Rehabil. 2020, 36, 176–186. [Google Scholar] [CrossRef]
  15. Kim, S.N.; Wang, X.; Park, H.J. Editorial: Integrative Approach to Parkinson’s Disease. Front. Aging Neurosci. 2019, 11, 339. [Google Scholar] [CrossRef] [PubMed]
  16. Church, F.C. Treatment Options for Motor and Non-Motor Symptoms of Parkinson’s Disease. Biomolecules 2021, 11, 612. [Google Scholar] [CrossRef]
  17. Petzinger, G.M.; Fisher, B.E.; McEwen, S.; Beeler, J.A.; Walsh, J.P.; Jakowec, M.W. Exercise-enhanced neuroplasticity targeting motor and cognitive circuitry in Parkinson’s disease. Lancet Neurol. 2013, 12, 716–726. [Google Scholar] [CrossRef] [PubMed]
  18. Palasz, E.; Niewiadomski, W.; Gasiorowska, A.; Wysocka, A.; Stepniewska, A.; Niewiadomska, G. Exercise-Induced Neuroprotection and Recovery of Motor Function in Animal Models of Parkinson’s Disease. Front. Neurol. 2019, 10, 1143. [Google Scholar] [CrossRef] [PubMed]
  19. Crocker, A.D. The regulation of motor control: An evaluation of the role of dopamine receptors in the substantia nigra. Rev. Neurosci. 1997, 8, 55–76. [Google Scholar] [CrossRef]
  20. Brar, S.; Henderson, D.; Schenck, J.; Zimmerman, E.A. Iron accumulation in the substantia nigra of patients with Alzheimer disease and parkinsonism. Arch. Neurol. 2009, 66, 371–374. [Google Scholar] [CrossRef]
  21. de Bie, R.M.A.; Clarke, C.E.; Espay, A.J.; Fox, S.H.; Lang, A.E. Initiation of pharmacological therapy in Parkinson’s disease: When, why, and how. Lancet Neurol. 2020, 19, 452–461. [Google Scholar] [CrossRef] [PubMed]
  22. Grosset, D.G.; Macphee, G.J.A.; Nairn, M.; Grp, G.D. Guidelines Diagnosis and pharmacological management of Parkinson’s disease: Summary of SIGN guidelines. BMJ Brit Med. J. 2010, 340, b5614. [Google Scholar] [CrossRef] [PubMed]
  23. Heim, B.; Krismer, F.; De Marzi, R.; Seppi, K. Magnetic resonance imaging for the diagnosis of Parkinson’s disease. J. Neural Transm. 2017, 124, 915–964. [Google Scholar] [CrossRef]
  24. Brooks, D.J. Imaging approaches to Parkinson disease. J. Nucl. Med. 2010, 51, 596–609. [Google Scholar] [CrossRef]
  25. Bae, Y.J.; Kim, J.M.; Sohn, C.H.; Choi, J.H.; Choi, B.S.; Song, Y.S.; Nam, Y.; Cho, S.J.; Jeon, B.; Kim, J.H. Imaging the Substantia Nigra in Parkinson Disease and Other Parkinsonian Syndromes. Radiology 2021, 300, 260–278. [Google Scholar] [CrossRef]
  26. Behnke, S.; Berg, D.; Naumann, M.; Becker, G. Differentiation of Parkinson’s disease and atypical parkinsonian syndromes by transcranial ultrasound. J. Neurol. Neurosur Ps. 2005, 76, 423–425. [Google Scholar] [CrossRef] [PubMed]
  27. Tao, A.Y.; Chen, G.Z.; Deng, Y.B.; Xu, R.F. Accuracy of Transcranial Sonography of the Substantia Nigra for Detection of Parkinson’s Disease: A Systematic Review and Meta-Analysis. Ultrasound Med. Biol. 2019, 45, 628–641. [Google Scholar] [CrossRef]
  28. Mei, Y.L.; Yang, J.; Wu, Z.R.; Yang, Y.; Xu, Y.M. Transcranial Sonography of the Substantia Nigra for the Differential Diagnosis of Parkinson’s Disease and Other Movement Disorders: A Meta-Analysis. Park. Dis. 2021, 2021, 8891874. [Google Scholar] [CrossRef]
  29. Berg, D.; Seppi, K.; Behnke, S.; Liepelt, I.; Schweitzer, K.; Stockner, H.; Wollenweber, F.; Gaenslen, A.; Mahlknecht, P.; Spiegel, J.; et al. Enlarged substantia nigra hyperechogenicity and risk for Parkinson disease: A 37-month 3-center study of 1847 older persons. Arch. Neurol. 2011, 68, 932–937. [Google Scholar] [CrossRef]
  30. Skoloudík, D.; Fadrná, T.; Bártova, P.; Langová, K.; Ressner, P.; Zapletalová, O.; Hlustík, P.; Herzig, R.; Kannovsky, P. Reproducibility of sonographic measurement of the substantia nigra. Ultrasound Med. Biol. 2007, 33, 1347–1352. [Google Scholar] [CrossRef]
  31. Berg, D.; Godau, J.; Walter, U. Transcranial sonography in movement disorders. Lancet Neurol. 2008, 7, 1044–1055. [Google Scholar] [CrossRef]
  32. Berardelli, A.; Wenning, G.K.; Antonini, A.; Berg, D.; Bloem, B.R.; Bonifati, V.; Brooks, D.; Burn, D.J.; Colosimo, C.; Fanciulli, A.; et al. EFNS/MDS-ES recommendations for the diagnosis of Parkinson’s disease. Eur. J. Neurol. 2013, 20, 16–34. [Google Scholar] [CrossRef]
  33. Basukala, D.; Mukundan, R.; Lim, A.; Hurrell, M.A.; Keenan, R.J.; Dalrymple-Alford, J.C.; Anderson, T.J.; Myall, D.J.; Melzer, T.R. Automated segmentation of substantia nigra and red nucleus using quantitative susceptibility mapping images: Application to Parkinson’s disease. Comput. Electr. Eng. 2021, 91, 107091. [Google Scholar] [CrossRef]
  34. Choi, H.; Ha, S.; Im, H.J.; Paek, S.H.; Lee, D.S. Refining diagnosis of Parkinson’s disease with deep learning-based interpretation of dopamine transporter imaging. Neuroimage Clin. 2017, 16, 586–594. [Google Scholar] [CrossRef]
  35. Vásquez-Correa, J.C.; Arias-Vergara, T.; Orozco-Arroyave, J.R.; Eskofier, B.; Klucken, J.; Nöth, E. Multimodal Assessment of Parkinson’s Disease: A Deep Learning Approach. IEEE J. Biomed. Health 2019, 23, 1618–1630. [Google Scholar] [CrossRef]
  36. Piccardo, A.; Cappuccio, R.; Bottoni, G.; Cecchin, D.; Mazzella, L.; Cirone, A.; Righi, S.; Ugolini, M.; Bianchi, P.; Bertolaccini, P.; et al. The role of the deep convolutional neural network as an aid to interpreting brain [F]DOPA PET/CT in the diagnosis of Parkinson’s disease. Eur. Radiol. 2021, 31, 7003–7011. [Google Scholar] [CrossRef] [PubMed]
  37. Chakraborty, S.; Aich, S.; Kim, H.C. Detection of Parkinson’s Disease from 3T T1 Weighted MRI Scans Using 3D Convolutional Neural Network. Diagnostics 2020, 10, 402. [Google Scholar] [CrossRef] [PubMed]
  38. Sivaranjini, S.; Sujatha, C.M. Deep learning based diagnosis of Parkinson’s disease using convolutional neural network. Multimed. Tools Appl. 2020, 79, 15467–15479. [Google Scholar] [CrossRef]
  39. Manzanera, O.M.; Meles, S.K.; Leenders, K.L.; Renken, R.J.; Pagani, M.; Arnaldi, D.; Nobili, F.; Obeso, J.; Oroz, M.R.; Morbelli, S.; et al. Scaled Subprofile Modeling and Convolutional Neural Networks for the Identification of Parkinson’s Disease in 3D Nuclear Imaging Data. Int. J. Neural Syst. 2019, 29, 1950010. [Google Scholar] [CrossRef]
  40. Zhao, H.L.; Tsai, C.C.; Zhou, M.Y.; Liu, Y.P.; Chen, Y.L.; Huang, F.; Lin, Y.C.; Wang, J.J. Deep learning based diagnosis of Parkinson’s Disease using diffusion magnetic resonance imaging. Brain Imaging Behav. 2022, 16, 1749–1760. [Google Scholar] [CrossRef]
  41. Vyas, T.; Yadav, R.; Solanki, C.; Darji, R.; Desai, S.; Tanwar, S. Deep learning-based scheme to diagnose Parkinson’s disease. Expert. Syst. 2022, 39, e12739. [Google Scholar] [CrossRef]
  42. Shen, L.; Shi, J.; Dong, Y.; Ying, S.H.; Peng, Y.X.; Chen, L.; Zhang, Q.; An, H.D.; Zhang, Y.C. An Improved Deep Polynomial Network Algorithm for Transcranial Sonography-Based Diagnosis of Parkinson’s Disease. Cogn. Comput. 2020, 12, 553–562. [Google Scholar] [CrossRef]
  43. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16133–16142. [Google Scholar]
  44. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  46. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  47. Zhang, J.; Lam, S.K.; Teng, X.Z.; Ma, Z.R.; Han, X.Y.; Zhang, Y.P.; Cheung, A.L.Y.; Chau, T.C.; Ng, S.C.Y.; Lee, F.K.H.; et al. Radiomic feature repeatability and its impact on prognostic model generalizability: A multi-institutional study on nasopharyngeal carcinoma patients. Radiother. Oncol. 2023, 183, 109578. [Google Scholar] [CrossRef] [PubMed]
  48. Lam, S.K.; Zhang, Y.P.; Zhang, J.; Li, B.; Sun, J.C.; Liu, C.Y.T.; Chou, P.H.; Teng, X.Z.; Ma, Z.R.; Ni, R.Y.; et al. Multi-Organ Omics-Based Prediction for Adaptive Radiation Therapy Eligibility in Nasopharyngeal Carcinoma Patients Undergoing Concurrent Chemoradiotherapy. Front. Oncol. 2022, 11, 792024. [Google Scholar] [CrossRef]
  49. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 11976–11986. [Google Scholar]
  50. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  51. Vujovic, Z.D. Classification Model Evaluation Metrics. Int. J. Adv. Comput. Sc. 2021, 12, 599–606. [Google Scholar] [CrossRef]
  52. Delong, E.R.; Delong, D.M.; Clarkepearson, D.I. Comparing the Areas under 2 or More Correlated Receiver Operating Characteristic Curves—A Nonparametric Approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef]
  53. Baranauskas, M.; Jurkonis, R.; Lukosevicius, A.; Matijosaitis, V.; Gleizniene, R.; Rastenyte, D. Diagnostic Ability of Radiofrequency Ultrasound in Parkinson’s Disease Compared to Conventional Transcranial Sonography and Magnetic Resonance Imaging. Diagnostics 2020, 10, 778. [Google Scholar] [CrossRef]
Figure 1. A schematic diagram of the pre-processing approaches for the single-channel input images and dual-channel input images.
Figure 1. A schematic diagram of the pre-processing approaches for the single-channel input images and dual-channel input images.
Bioengineering 11 00889 g001
Figure 2. The DANet structure.
Figure 2. The DANet structure.
Bioengineering 11 00889 g002
Figure 3. The architecture of CNXV2-DANet.
Figure 3. The architecture of CNXV2-DANet.
Bioengineering 11 00889 g003
Figure 4. Classification results of CNXV2-DANet and ConvNeXt V2 on the testing set for single-channel and dual-channel methods: (a) confusion matrix and (b) ROC curve.
Figure 4. Classification results of CNXV2-DANet and ConvNeXt V2 on the testing set for single-channel and dual-channel methods: (a) confusion matrix and (b) ROC curve.
Bioengineering 11 00889 g004
Figure 5. Comparison of the performance of the four studied models on the testing dataset (a) in terms of the averaged metrics of accuracy, precision, recall, and F1-score with the corresponding standard deviations across the 5 randomizations, and (b) in terms of ROC curves under one of the randomization seeds.
Figure 5. Comparison of the performance of the four studied models on the testing dataset (a) in terms of the averaged metrics of accuracy, precision, recall, and F1-score with the corresponding standard deviations across the 5 randomizations, and (b) in terms of ROC curves under one of the randomization seeds.
Bioengineering 11 00889 g005
Table 1. Performance of CNXV2-DANet in both single-channel and dual-channel settings.
Table 1. Performance of CNXV2-DANet in both single-channel and dual-channel settings.
Method AccuracyPrecisionRecallF1-ScoreAUC
Single-channelSeed10.7670.8710.6140.7200.884
Seed20.7890.7720.7910.7820.868
Seed30.7330.6790.8640.7600.784
Seed40.8000.8650.7110.7800.861
Seed50.8330.8970.7600.8230.909
Average0.7840.8170.7480.7730.861
SD0.0370.0900.0930.0370.047
Dual-channelSeed10.8670.8330.9090.8700.924
Seed20.8610.8630.8620.7850.898
Seed30.8440.8410.8410.8410.898
Seed40.8000.8650.8110.7800.894
Seed50.8230.8410.8040.8220.917
Average0.8390.8490.8450.8200.906
SD0.0280.0140.0430.0380.013
Table 2. Performance of ConvNeXt V2 in both single-channel and dual-channel settings.
Table 2. Performance of ConvNeXt V2 in both single-channel and dual-channel settings.
Method AccuracyPrecisionRecallF1-ScoreAUC
Single-channelSeed10.7440.7630.6740.7160.852
Seed20.8440.7870.9020.8410.884
Seed30.7560.6900.9090.7840.838
Seed40.8000.7830.8180.8000.876
Seed50.8110.8540.7610.8050.830
Average0.7910.7750.8130.7890.856
SD0.0410.0590.0990.0460.023
Dual-channelSeed10.9110.9720.8370.9000.946
Seed20.8110.8750.6830.7670.919
Seed30.8560.8780.8180.8470.903
Seed40.7110.6880. 7500.7170.847
Seed50.8220.8750.7610.7610.865
Average0.8220.8580.7750.7980.896
SD0.0730.1040.0690.0740.040
Table 3. Quantitative performance of the proposed CNXV2-DANet model and three state-of-the-art networks (ConvNeXt V2, ConvNeXt, and Swin Transformer).
Table 3. Quantitative performance of the proposed CNXV2-DANet model and three state-of-the-art networks (ConvNeXt V2, ConvNeXt, and Swin Transformer).
ModelAccuracy (SD)Precision (SD)Recall (SD)F1-Score (SD)AUC (SD)
CNXV2-DANet
(proposed)
0.839 (0.028)0.849 (0.014)0.845 (0.043)0.820 (0.038)0.906 (0.013)
ConvNeXt V20.822 (0.073)0.858 (0.104)0.775(0.069)0.798 (0.074)0.896 (0.040)
ConvNeXt0.733 (0.045)0.710 (0.060)0.794 (0.117)0.744 (0.051)0.842 (0.016)
Swin Transformer0.753 (0.022)0.734 (0.057)0.799 (0.075)0.760 (0.028)0.834 (0.030)
Table 4. PD classification performance comparison with the related-studies in recent years.
Table 4. PD classification performance comparison with the related-studies in recent years.
ResearcherYearDatasetModalityMethodAccuracy (%)With an Independent
Test Set
Sivaranjini et al. [38]2019182MRIAlexNet +
transfer learning
88.9No
Manzanera et al. [39]2019310PETCNN86.0Yes
Chakraborty et al. [37]2020406MRICNN95.3No
Shen et al. [42]2020153TCSDeep polynomial
network
86.9No
Zhao et al. [40]2022432MRI3D CNN80.7Yes
Our study2024588TCSCNXV2-DANet83.9Yes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kang, H.; Wang, X.; Sun, Y.; Li, S.; Sun, X.; Li, F.; Hou, C.; Lam, S.-k.; Zhang, W.; Zheng, Y.-p. Automatic Transcranial Sonography-Based Classification of Parkinson’s Disease Using a Novel Dual-Channel CNXV2-DANet. Bioengineering 2024, 11, 889. https://doi.org/10.3390/bioengineering11090889

AMA Style

Kang H, Wang X, Sun Y, Li S, Sun X, Li F, Hou C, Lam S-k, Zhang W, Zheng Y-p. Automatic Transcranial Sonography-Based Classification of Parkinson’s Disease Using a Novel Dual-Channel CNXV2-DANet. Bioengineering. 2024; 11(9):889. https://doi.org/10.3390/bioengineering11090889

Chicago/Turabian Style

Kang, Hongyu, Xinyi Wang, Yu Sun, Shuai Li, Xin Sun, Fangxian Li, Chao Hou, Sai-kit Lam, Wei Zhang, and Yong-ping Zheng. 2024. "Automatic Transcranial Sonography-Based Classification of Parkinson’s Disease Using a Novel Dual-Channel CNXV2-DANet" Bioengineering 11, no. 9: 889. https://doi.org/10.3390/bioengineering11090889

APA Style

Kang, H., Wang, X., Sun, Y., Li, S., Sun, X., Li, F., Hou, C., Lam, S. -k., Zhang, W., & Zheng, Y. -p. (2024). Automatic Transcranial Sonography-Based Classification of Parkinson’s Disease Using a Novel Dual-Channel CNXV2-DANet. Bioengineering, 11(9), 889. https://doi.org/10.3390/bioengineering11090889

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop