Next Article in Journal
Automotive 2.1 μm Full-Depth Deep Trench Isolation CMOS Image Sensor with a 120 dB Single-Exposure Dynamic Range
Next Article in Special Issue
Mobile Crowdsensing in Ecological Momentary Assessment mHealth Studies: A Systematic Review and Analysis
Previous Article in Journal
A Lightweight and Efficient Method of Structural Damage Detection Using Stochastic Configuration Network
Previous Article in Special Issue
User Perspectives of Geriatric German Patients on Smart Sensor Technology in Healthcare
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Clinically Informed Automated Assessment of Finger Tapping Videos in Parkinson’s Disease

1
Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
2
Pacific Parkinson Research Centre, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
3
Department of Neurology, Faculty of Medicine, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(22), 9149; https://doi.org/10.3390/s23229149
Submission received: 19 September 2023 / Revised: 30 October 2023 / Accepted: 10 November 2023 / Published: 13 November 2023
(This article belongs to the Special Issue Intelligent Sensors for Healthcare and Patient Monitoring)

Abstract

:
The utilization of Artificial Intelligence (AI) for assessing motor performance in Parkinson’s Disease (PD) offers substantial potential, particularly if the results can be integrated into clinical decision-making processes. However, the precise quantification of PD symptoms remains a persistent challenge. The current standard Unified Parkinson’s Disease Rating Scale (UPDRS) and its variations serve as the primary clinical tools for evaluating motor symptoms in PD, but are time-intensive and prone to inter-rater variability. Recent work has applied data-driven machine learning techniques to analyze videos of PD patients performing motor tasks, such as finger tapping, a UPDRS task to assess bradykinesia. However, these methods often use abstract features that are not closely related to clinical experience. In this paper, we introduce a customized machine learning approach for the automated scoring of UPDRS bradykinesia using single-view RGB videos of finger tapping, based on the extraction of detailed features that rigorously conform to the established UPDRS guidelines. We applied the method to 75 videos from 50 PD patients collected in both a laboratory and a realistic clinic environment. The classification performance agreed well with expert assessors, and the features selected by the Decision Tree aligned with clinical knowledge. Our proposed framework was designed to remain relevant amid ongoing patient recruitment and technological progress. The proposed approach incorporates features that closely resonate with clinical reasoning and shows promise for clinical implementation in the foreseeable future.

1. Introduction

Parkinson’s Disease (PD) is the most-common neurodegenerative movement disorder characterized by various motor symptoms, including resting tremor, bradykinesia, muscular rigidity, and postural instability [1]. Among these symptoms, bradykinesia, the slowness of movement often associated with a reduced movement amplitude, is a fundamental feature and a key diagnostic criterion for Parkinsonism [2].
The Unified Parkinson’s Disease Rating Scale (UPDRS) is recognized as the definitive standard for assessing the severity of Parkinson’s Disease (PD). This scale has undergone a partial revision via the Movement Disorder Society (MDS), resulting in the MDS-UPDRS [3]. An essential element of the MDS-UPDRS is the motor examination (Part III), encompassing 18 motor tasks performed by the patient and evaluated by a rater, who scores each task based on the observed level of impairment.
A significant task in this scale, Item 4, is finger tapping, which is used to gauge bradykinesia. In this test, patients are instructed to tap their index finger to their thumb as quickly and widely as possible, completing ten taps. The rater observes and evaluates several aspects of this action, such as the speed and amplitude, any noticeable decrease in speed or amplitude, and the presence of hesitations or stops. Each side of the body is evaluated separately, and scores are assigned on a scale from 0 (normal) to 4 (severe), as outlined in Table 1.
While the transition from the UPDRS to the MDS-UPDRS addressed several uncertainties and shortcomings of the original version, it did not completely overcome all challenges. Notably, the subjective nature of this manually administered clinical scale persists, resulting in potential scoring discrepancies by the same (intra-rater) or different (inter-rater) evaluators. Moreover, human assessments of this nature can be both time-intensive and costly.
Advancements in machine learning technologies have greatly improved the objective and automated assessment of Parkinson’s Disease (PD) [4,5,6,7]. Notably, wearable sensor technology has been utilized to measure bradykinesia during finger tapping [8,9], with several of these sensors now available for clinical use [10,11]. The use of machine learning in the assessment of Parkinsonism holds promise for clinical care, potentially providing a more-efficient, extensive, and remote monitoring system. Additionally, machine learning approaches could greatly benefit clinical trials by offering a quantitative, objective digital biomarker for Parkinsonism.
However, concerns regarding costs and potential issues with patient compliance regarding wearable sensors have led to growing interest in video-based analysis. This method offers a non-contact, unobtrusive alternative for the automated assessment of motor symptoms in PD [12,13]. That said, there are technical challenges associated with the remote interpretation of video recordings [5]. A significant limitation of existing video-based assessments of PD is their focus on only a subset of the clinically relevant features detailed in the UPDRS instructions, often resulting in limited clinical interpretation of the models. Given that UPDRS is grounded in decades of clinical expertise and encompasses a wide range of bradykinesia symptoms in Parkinsonism, both the comprehensiveness and explainability of any model are vital. Furthermore, many studies have relied on solely collecting data in a standardized video collection environment, raising concerns about the sustainability and adaptability of these models. Addressing these gaps, our study aimed to develop a machine-learning model for the automatic prediction of UPDRS bradykinesia scores from finger tapping videos. We focused on enhancing the robustness of our model by extracting features that are in strict accordance with the clinical guidelines provided in the UPDRS, employing a white-box model for better explainability. This approach also involves incorporating a variety of video collection environments to ensure broader applicability and resilience of the model in diverse settings.

2. Related Work

In this section, we examine the existing literature on the automated assessment of Parkinson’s Disease (PD) severity using Machine Learning (ML) techniques. The subsequent subsections offer a comprehensive overview of cutting-edge methodologies and their respective contributions to this field.

2.1. Advances in Automated Assessment of Parkinson’s Disease with Machine Learning Approach

The majority of ML-based automated assessment of PD approaches utilize body-worn inertial sensors, such as accelerometers and gyroscopes. These sensors include independent devices integrated with short-range communication capabilities [8], as well as sensors embedded in smartphones or smartwatches that facilitate data collection using the patients’ own mobile devices [14]. AI technology using sensor signals has been used to investigate a wide range of PD symptoms, both motor and non-motor, including bradykinesia, tremor, rigidity, gait disturbances, freezing of gait, falls, sweating, and sleep disturbances [8]. Bradykinesia, a key symptom for diagnosing PD, is the most extensively investigated symptom [8]. Various studies exploring bradykinesia, resting tremor, or rigidity in the upper limbs have employed commercially available sensors placed on the arms, wrists, or fingertips, yielding diverse outcome measures [8,15,16]. The use of sensor technology for detailed gait analysis, particularly for measuring parameters like stride length, velocity, and cadence, has been emphasized due to the debilitating nature of gait and balance impairments in PD [17,18]. Additionally, focused research has been conducted on specific PD motor symptoms such as dysarthria and micrographia [19,20,21]. However, there are significant limitations associated with wearable-sensor-based monitoring. First, the need for patients to wear these devices can be intrusive, potentially leading to reduced compliance, particularly in the elderly population, primarily impacted by PD. Additionally, the diverse levels of digital literacy among older adults pose further challenges. Consequently, there have been suggestions for alternative approaches to wearable inertial sensors. Notable among these alternatives are techniques employing electromagnetic signals [22], with video-based assessment also gaining prominence as a major competitor to wearable sensors.

2.2. The Current Landscape of Video-Based Assessment for Parkinson’s Disease

Capitalizing on the remarkable advancements in computer vision technology, a multitude of studies have focused on analyzing the motor symptoms of Parkinson’s Disease (PD) using video data. While the precise methodologies vary across these studies, they typically adhere to a four-step process: (i) recording symptoms with a camera, (ii) estimating human poses, (iii) extracting pertinent kinematic features of the motor symptom, and (iv) implementing machine learning models for analysis.
In the realm of video-based Parkinson’s Disease (PD) assessment, studies generally fall into two categories based on their initial video acquisition approach: those employing advanced video recording technologies and those using standard consumer-grade RGB cameras. Advanced technologies include RGB–depth cameras like Microsoft Kinect, which, by providing 3D data indicating the distance between the captured subject and the camera, are used to assess the bradykinesia of the arms or gait impairment in PD [23,24,25].
Examples also encompass innovations like the Leap Motion controller, a virtual reality tool for upper limb rehabilitation in PD patients through serious games [26]. The first-generation system utilized three LEDs and two infrared cameras, capturing data at a resolution of 640 × 240 at 120 frames per second, designed to function optimally within a 60 cm range. This system is now being integrated into automatic PD classification research [27,28,29]. However, a large portion of research still relies on readily available consumer cameras, such as standard camcorders and smartphone cameras, for daily clinical practice [4,30,31,32]. Regarding the categorization of these studies, they can be either marker-based Motion Capture systems (MoCaps) or markerless, determined by whether or not they use body-attached markers. Infrared MoCaps, while the gold standard for clinical gait analysis, require extensive space and equipment, limiting their accessibility. Recent studies, therefore, are exploring the viability of markerless systems as a feasible alternative to MoCaps [33,34].
For the human pose estimation step, various studies incorporate specialized or generic 3D or 2D pose estimation algorithms. Popular open-source libraries like OpenPose [35], Google MediaPipe [36], and AlphaPose [37] are frequently employed for systems designed to assess PD motor symptoms. Here, we chose MediaPipe for its efficiency, effectiveness in diverse environments, and regular updates by the GoogleAI team.
In terms of kinematic feature extraction and their application in machine learning models, there is a range of utilized features and models. Park et al. [32] extracted kinematic features related to the amplitude, velocity, and decremental response of bradykinesia from 110 videos of finger tapping and utilized a Support Vector Machine (SVM) classifier to show good reliability with clinical ratings. Vignoud et al. [38] proposed a method using seven parameters extracted from finger tapping and hand movement videos representing speed, amplitude, fatigue, and periodicity to predict the appropriate UDPRS score with linear regression and Decision Tree models. Liu et al. [39] incorporated a more-novel network architecture, a Global Temporal-difference Shift Network (GTSN), to estimate the MDS-UPDRS tremor scores from video recordings. A Convolutional Neural Network (CNN) examining location, velocity, and inter-joint relationship information derived from video recordings has also been proposed [40].
It should be noted that, while the Unified Parkinson’s Disease Rating Scale (UPDRS) advises evaluators to consider factors such as amplitude, velocity, decrement, and pauses or hesitations, numerous studies have focused on a limited subset of these attributes. Recognizing this, recent research has increasingly prioritized the extraction of more-comprehensive features and the interpretability of the resulting models. Morinan et al. [4,30] introduced a Random Forest algorithm for classifying bradykinesia, utilizing 11 kinetic features derived from video recordings of 1156 MDS-UPDRS assessments across five medical centers. Similarly, Islam et al. [31] selected clinically pertinent features from an initial set of 53, explaining their model’s performance by employing the Shapley Additive Explanations method, alongside various regression models. Within the extensive array of pertinent research, we have distilled and summarized the four most-current and -relevant studies that align closely with our investigation in Table 2.
In light of the trends identified in previous research, our analysis utilized standard RGB cameras in two different recording environments, combined with MediaPipe for human pose estimation. We fine-tuned our methodology for extracting kinematic features to better correspond with the clinical characteristics of bradykinesia. Furthermore, we opted for a Decision Tree model to improve the interpretability of our findings.

3. Method

Recognizing that data collection is a persistent and long-term pursuit, with our data repository steadily growing, we present a framework, illustrated in Figure 1, to maintain the lifecycle of the proposed system. Our framework consists of three sequential and independent cycles: a data cycle, a model learning cycle, and an algorithm deployment cycle.

3.1. Data Cycle

The data cycle consists of three parts; data collection, data preprocessing, and data annotation.
Data collection: Patient-generated videos can be sourced from diverse platforms, comprising self-recorded, home-based footage and professionally captured videos within laboratory settings by healthcare practitioners. This compilation may include a range of data forms: structured and unstructured data (encompassing text, images, and videos), as well as external data acquired via Application Programming Interfaces (APIs).
Data preprocessing: Following data collection, the preprocessing phase is undertaken to transform the data into a format suitable for further analysis and model training. This phase includes handling missing values, eliminating outliers, normalizing or scaling features, and ensuring data format uniformity. We employed a range of methods to process the initial video data, as illustrated in Figure 2.
  • Upscaling: Upscaling is employed to increase the resolution or dimensions of images or videos. This enhancement improves the quality of low-resolution data, thereby strengthening the model’s ability to extract significant features [41,42].
  • Removing motion blur: Motion blur occurs in images or videos as a result of camera wobble or object motion [43]. When collecting multi-source data from various devices and environments, it is essential to remove motion blur to regain the sharpness and clarity of the data.
  • Denoising: Noise may be introduced during the image- or video-acquisition process, leading to a degradation in quality and potentially hindering the capacity to extract meaningful features. A denoising step is often employed to reduce noise from the data while retaining the details.
  • Frame interpolation: This technique generates additional frames in videos with low frame rates [44]. This process estimates the motion between existing frames and synthesizes new ones, thereby smoothing the motion and enhancing the temporal resolution.
These measures are designed to enhance data quality and facilitate the extraction of meaningful information for more-effective model training.
Data annotation: In this step, domain experts, typically clinicians, are involved in providing UPDRS scores or labels for the patient’s data, which serve as a reference for model training and evaluation [45]. It is imperative to establish precise and uniform guidelines for the scoring criteria in the annotation process. As shown in Table 1, the UPDRS documentation offers detailed instructions on how clinicians should allocate scores to patients’ data. Our study adopted a dual-annotator system. The primary annotator, an experienced Movement Disorder Specialist, annotated all the videos. In parallel, a secondary annotator, also skilled in movement disorders, was tasked with validation, acting as a quality control to guarantee the accuracy of the annotations. These annotators were blind to each other’s ratings throughout the process. To ascertain the annotations’ quality and reliability, we evaluated the inter-annotator agreement. It is essential to manage any discrepancies and disagreements among clinicians during the data-annotation process. This control is necessary to maintain the reliability and consistency of the annotated data.
In practice, the data cycle is iterative to allow the data repository to grow. As the data volume increases, this repetitive method facilitates continuous improvements and modifications to the models, leading to enhanced performance over time [45].

3.2. Model Learning Cycle

The model learning cycle consists of three key stages: hand pose estimation, feature extraction, and machine learning.
Initially, the 3D hand pose is extracted using the most-recent version of MediaPipe, updated in February 2023, to generate time series data of crucial joint positions [36]. MediaPipe was selected over other deep learning libraries for hand pose estimation due to its efficient performance in various recording contexts and its relatively lightweight architecture. The raw time series data of key joint positions are represented as K = { K j ( t ) } , where j means the j-th key point, j = 1 , 2 , . . , 20 , and t denotes the image frame index (for t = 1 , 2 , . . . , L ). We calculated the time series of the thumb–index finger distance, denoted as I with length L, which corresponds to the distance between points K 4 and K 8 in the finger tapping task. In the second stage, we extracted 15 specific features for each video, as detailed in Table 3. These features encompass two demographic elements and 13 kinematic features, which were extracted through signal processing of the time series data. These kinematic features are aligned with the guidelines of the UPDRS instructions.
Lastly, based on these features, we trained a machine learning model.

3.2.1. Time Series Preprocessing

For the time series input I , p = [ p 1 , p 2 , . . . , p n , . . . , p N ] represents the peaks, p n represents the n-th peak, and A n represents the amplitude of the corresponding peak p n . Given a hand model, as shown in Figure 3, K j = ( k x , j , k y , j , k z , j ) represents the position of the j-th keypoint. The palm size at time t is estimated using K 0 and K 9 as:
P a l m t = ( k x , 0 k x , 9 ) 2 + ( k y , 0 k y , 9 ) 2 + ( k z , 0 k z , 9 ) 2 | t
Then, normalization is performed to make P a l m t more suitable for subjectwise comparison, as below:
I n o r m = I P a l m t
Following this, we applied Empirical Mode Decomposition (EMD) to decompose I n o r m into a series of Intrinsic Mode Functions (IMFs) and a residual component, as depicted in Figure 4. Each IMF captures a distinct oscillatory mode characterized by variable frequencies and amplitudes. In our approach, we selected the first three IMFs as the primary components of the signal, striking a balance between simplifying the data and preserving vital information. However, even after EMD-based decomposition, the signal may contain noise or unwanted artifacts, such as those arising from finger tremors. To address this, we employed a denoising and smoothing process to eliminate such disturbances.
After preprocessing and cleaning the time series data, the subsequent step involved the detection of peaks or specific events of interest within the signal. The peak-detection module was applied to find local maxima or minima, which served as key points in the data.

3.2.2. Feature Extraction

Feature vector: demographics. In our model, we incorporated two demographic factors: age and sex. This inclusion was based on the observation that older individuals may exhibit signs of bradykinesia related to aging, independent of Parkinson’s Disease (PD), and that both age and sex can influence the progression of PD [1,46,47]. While the Unified Parkinson’s Disease Rating Scale (UPDRS) instructs raters to evaluate based on observable symptoms, there is a possibility that human annotators may subconsciously consider these demographic characteristics. Therefore, we chose to include age and sex as features in our model. Furthermore, given our goal of assessing the model’s explainability, incorporating these demographic factors may offer a more-comprehensive understanding that might not be immediately evident through the UPDRS guidelines alone.
Feature vector: amplitude. Based on the time series data, we extracted 5 amplitude-related features. First, according to Equations (3) and (4), we extracted F a m p a v g and F a m p v a r as:
F a m p a v g = n = 1 N A n N
F a m p v a r = n = 1 N ( A n F a m p a v g ) 2 N
Then, based on the peaks P, we applied piecewise linear regression to model the distinct patterns of different regions of the data, as shown in Figure 5:
y = α 0 + F a m p α 1 x + F a m p α 2 ( x F a m p b p ) I ( x > F a m p b p ) + ϵ
where α 0 is the intercept, F a m p α 1 and F a m p α 2 represent the slope coefficients of two segments, respectively, and ϵ represents the error term or residual, accounting for the deviations of the actual data points from the fitted model. I ( · ) is an indicator function that takes a value of 1 if x is greater than a specific breakpoint F a m p b p and 0 otherwise.
Feature vector: velocity. We extracted 5 velocity-related features. Here, we have
V n = A n + A n 1 Δ t n
where Δ t n = t n t n 1 is the time interval between the n-th and ( n 1 ) -th peak. Then, we have
F v e l a v g = n = 1 N 1 V n N 1
F v e l v a r = n = 1 N 1 ( V n F v e l a v g ) 2 N 1
We also extracted another 3 features, F v e l α 1 , F v e l α 2 , and F v e l b p , from the piecewise linear regression fitting of the velocity time series, to estimate the velocity trend.
y = α 0 + F v e l α 1 x + F v e l α 2 ( x F v e l b p ) I ( x > F v e l b p ) + ϵ
Feature vector: halt and hesitation. In PD, “halts or hesitations” are characterized by brief instances of freezing or a delayed ability to initiate movements, often seen during repetitive actions. While the concept of the freezing of gait, essentially halts and hesitations during walking, is well-documented in PD [48], there is a scarcity of detailed definitions and quantitative assessments of “halts and hesitations” in upper limb movements during repetitive tasks.
Through extensive discussions to clarify “halts” and “hesitation” within our study, we defined “halts” as moments of complete stoppage, characterized by zero amplitude. In contrast, “hesitation” was identified as a noticeable reduction in velocity, registering values above zero, but significantly below a predefined empirical threshold. This distinction was quantitatively established in Equation (11), where the “hesitation” threshold was set at 20% of the average amplitude observed immediately before and after the occurrence. More specifically, we define the “halt and hesitation” in finger tapping tasks as follows: given a time series of the distance between the keypoints K 4 and K 8 , p = [ p 1 , p 2 , . . . , p n , . . . , p N ] represents the detected peaks. A curve fitting was applied to p , aiming to estimate the trend of the peaks. Assuming p n to be the peak when halt and hesitation happened (the red arrow in Figure 6), p ^ n represents the predicted peak based on the fitted model. The residual between the observed value and the predicted value is denoted by | p ^ n p n | . Then, the feature “halt and hesitation” could be defined as:
F h h 1 if   | p ^ n p n |   θ 0 if   | p ^ n p n |   < θ .
As demonstrated in Figure 6, in this situation, the patient may experience a sudden interruption or drop in his/her ability to perform repetitive finger tapping movements. The threshold θ is defined as:
θ = α · A n + 1 + A n 1 2
where α is a parameter controlling the detection threshold. In our framework, we empirically set α to 0.2.

3.2.3. Machine Learning

We chose the Decision Tree approach as our primary classification method to improve the interpretability and quantification of assessments tailored to our specific task. Decision Trees offer clear interpretability and the ability to visually represent the model, facilitating an uncomplicated explanation of the decision-making process. Inherently proficient in managing categorical data, Decision Tree models are particularly suitable for modeling ordinal UPDRS scores. Additionally, the inherent transparency and interpretability of Decision Tree models correspond well with the explicit criteria set forth by the UPDRS. This alignment ensures that our model’s predictions are not only easily understandable, but also hold clinical relevance.
The Pearson correlation coefficients between the 15 features and the UPDRS scores are presented in Figure 7. Redundant or overlapping features can lead to inefficient splits and diminished predictive capability in the model. However, as we identified a high correlation solely between the frequency and velocity of tapping, we decided to maintain all 15 original features in our analysis. This approach was chosen to ensure that the unique contribution of each feature is preserved.

3.3. Algorithm Deployment Cycle

Model experimentation: This initial phase involved the training and evaluation of various models using different algorithms, hyperparameters, and subsets of data. The purpose of model experimentation is to identify the most-promising model based on its performance across various evaluation metrics.
Model updating: After selecting a suitable model, further refinement and improvement are often necessary. This updating process includes retraining the chosen model with additional data or altered hyperparameters to boost its performance. The goal of this iterative procedure is to enhance the model’s accuracy, robustness, and ability to generalize.
Deployment: The deployment phase involves making the refined and updated model accessible in a production setting. This stage encompasses establishing the required infrastructure to host and operate the model, which may involve servers, cloud services, or containerization platforms. Additionally, it includes the development of the necessary APIs or interfaces to enable the user or system access to the model for predictions or analyses.
The deployment cycle in MLOps begins with model experimentation, where a variety of models are trained and assessed. Following the selection of the most-effective model, it enters the updating stage for further improvement. The enhanced model is then deployed in a production environment, ready for practical application.
It is crucial to recognize that this deployment cycle is often iterative and ongoing. As new data are introduced, models require updates and redeployment to meet shifting demands. Continuous monitoring, testing, and feedback are integral to the deployment cycle, ensuring the deployed model maintains its performance, reliability, and efficacy in real-world settings.

4. Experiments

4.1. Participants and Dataset

RGB, single-view videos were collected in two settings with different camera and supervision setups: in-lab and in-clinic. In the laboratory setting, patients were invited to perform the finger tapping task as outlined in the MDS-UPDRS guidelines. This task was carried out in a spacious laboratory room. A research assistant was present to guide the patients through the task, which was captured by eight cameras (Lorex 4K Bullet IP Camera/E841CA-E from Lorex Technology Inc., Toronto, ON, Canada ) mounted on the walls. The raw data from these cameras had a resolution of 3840 × 2160 px and a frame rate of 15/s. Of the eight camera angles, our primary focus was on the single camera view that provided the clearest, most-unobstructed perspective of the hand movements. While not a central aspect of this study, an inertial sensor was also attached to the patient’s hand during recording, visible as a black band on the palm in Figure 1. In the clinic setting, patients were recruited to perform various tasks from the MDS-UPDRS, including finger tapping, in a designated room within the waiting area of the Movement Disorders Clinic. These tasks were recorded using an AI camera, OAK-D (Luxonis, Littleton, CO, USA), installed on an iMac 24” desktop (Apple, Cupertino, CA, USA). In this setting, patients received little or no assistance from a research assistant in using the app for recording. The raw RGB data collected in the clinic had a resolution of 1920 × 1080 px and a frame rate of 60 fps.
A total of 65 patients diagnosed with PD were recruited to do the UDPRS assessment. The clinical diagnosis for each patient was confirmed based on the United Kingdom Parkinson’s Disease Society Brain Bank clinical diagnostic criteria. To ensure the quality of the data, specialists manually inspected the videos, applying several criteria to evaluate their suitability. Videos were excluded if they exhibited poor visibility of motion, such as the hands frequently moving out of the camera frame, a significant distance between the hands and the camera, inconsistent execution of tasks, or frequent interruptions (verbal or physical) either by the patients or bystanders. After the exclusion process based on these quality criteria, 75 videos from 50 patients were deemed suitable for assessing the effectiveness of the proposed method. This dataset comprised 20 subjects recorded in a laboratory setting and 30 subjects in a clinic environment, as detailed in Table 4. For the purposes of our study, 60 videos were designated for the training set and 15 videos for the testing set. The selection of testing subjects was randomized, ensuring a balanced representation of different categories. The Research Ethics Board of the University of British Columbia approved this study, and all participants provided written consent. All individuals were evaluated using the MDS-UPDRS, with the entire evaluation process captured on video. Due to regulations safeguarding patient privacy, these videos are not permitted to be made publicly available.

4.2. Evaluation Matrix

To evaluate the performance of our model, we employed well-known metrics such as accuracy, precision, recall, and the F1-score. These metrics are defined as follows:
A c c u r a c y = T P + T N T P + F N + F P + T N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where T P , T N , F P , and F N represent “true positive”, “true negative”, “false positive”, and “false negative”, respectively, and the F score (F1 when β = 1 and F2 when β = 2) is defined as:
F β = 1 + β 2 · Precision · Recall β 2 · Precision + Recall
In our evaluation matrix for accuracy, we considered two scenarios:
  • y p r e d = y g t , which is reported as “Accuracy(t1)”, represents that the prediction y p r e d is equal to the ground truth y g t .
  • y p r e d [ y g t 1 , y g t , y g t + 1 ] , reported as “Accuracy(t2)”, represents that the prediction is within a range of the ground truth, which could also be called “acceptable” accuracy. This situation takes the inter-annotator agreement into consideration, and we will further analyze this situation in the result part.

4.3. Method for Comparison

To assess the performance of our proposed method, we compared it against two recent studies: AEMPD-FCN [49] and FTTST [23]. It is important to note that there might be significant differences in data quality, volume, and time series length between our study and those in the referenced work, which could potentially influence the comparative results. Additionally, we explored the effectiveness of different classifiers by including results from three other methods. These methods all employ the same feature-extraction framework, but differ in their classification approaches, such as Support Vector Machine (SVM) and Random Forest (RF). Specifically, b a s e l i n e - S P is the baseline model that follows the UDPRS criteria precisely. To some extent, this criterion is overly strict for some data points situated at the boundaries. For example, a slight decrease in amplitude or velocity could be classified as decrementing features. F15- S V M - G A uses the 15 features as the input vector to the SVM classifier, where the Genetic Algorithm is used to search for the best parameters. F15- R F uses Random Forest for classification, and in F15- D T , the D T represents a Decision Tree.

5. Results and Discussion

5.1. Labeling UPDRS Finger Tapping Videos

In our study, we calculated the level of agreement between annotators, as shown in Table 5. “Complete”: This term refers to the percentage of instances where both annotators assigned an identical score to a specific piece of data. “Acceptable”: This category covers cases where the scores given by the two annotators differed by either 0 or 1. It represents a more-lenient agreement threshold. “None”: This indicates the percentage of cases where the discrepancy in scoring between the two annotators was more than 1. The data presented in Table 5 highlights the inherent difficulty in achieving complete consensus among human annotators, even when they are experts in their field, as has been previously observed [50]. Acknowledging this challenge, we report two types of accuracy metrics in our experiments, namely Accuracy(t1) and Accuracy(t2), as defined in Section 4.2. These metrics are intended to evaluate the efficacy of our proposed method, taking into account the varying levels of agreement among annotators.

5.2. Classification of UPDRS Finger Tapping Score

First, the main classification task was performed on the finger tapping features using the clinical ratings. The proposed method achieved a high accuracy rate, an Accuracy(t1) of 80% and an Accuracy(t2) of 93.3%. Table 6 lists the specific accuracy(t1 and t2), precision, recall, and F-score of the final prediction results on five UPDRS scores. The structure of the Decision Tree and the confusion matrix of the classification results are shown in Figure 8 and Figure 9, respectively.
Clinical interpretation: Overall, the Decision Tree our model produced (as shown in (Figure 8)) aligns well with the MDS-UPDRS rating guidelines. These guidelines instruct raters to evaluate finger tapping based on: (i) amplitude, (ii) velocity, (iii) decrement in amplitude or velocity during repetitive tapping, and (iv) the presence of halts or hesitations [3]. The root node in our Decision Tree (Node 0) utilizes the “mean amplitude” feature, dividing the data into two paths and indicating amplitude as a pivotal factor in the finger tapping assessment. In the second layer, the nodes “vel_alpha2” and “vel_alpha1” reflect the velocity decrement, suggesting that greater decrements often correspond with lower UPDRS scores. This points to the relevance of the “decrement in amplitude or velocity” criterion in cases of mild bradykinesia, while in more-severe cases, the velocity notably reduces from the beginning of the task. Further nodes incorporate amplitude, velocity, and decrement aspects, aligning conceptually with the MDS-UPDRS guidelines, highlighting the model’s ability to offer more-quantifiable data and reduce intra-rater variability compared to human ratings.
Additional clinical insights emerged from the Decision Tree. Notably, “halts and hesitations” had minimal impact on the model. This could be due to the rarity and lesser correlation of upper limb freezing with PD severity compared to gait freezing. Our model’s interpretation suggests a non-linear relationship between these halts and overall bradykinesia severity [51]. It is important to recognize the limitations posed by the relatively small video dataset used for the AI training, potentially under-representing upper limb freezing episodes Furthermore, our study’s definition of halts and hesitations, based purely on significant amplitude reduction without a time duration criterion, may not fully align with clinical descriptions of a halt as a prolonged pause. Future studies, especially with larger datasets, might need a more-refined definition, possibly incorporating time-based thresholds for a better characterization of these phenomena. Lastly, while we considered demographic variables like age and gender, their absence in the Decision Tree underscores that UPDRS ratings focus on motor performance irrespective of demographic factors.

6. Limitations and Future Work

Our study marks a step forward in automated Parkinson’s Disease (PD) symptom assessment, yet it is important to recognize its limitations. Firstly, the limited number of patient videos used for training the model, though yielding promising accuracy, may not fully represent the broad spectrum of PD symptoms, including pronounced halts or hesitation phenomena. A more-extensive dataset will be critical for enhancing the model’s generalizability and robustness. Secondly, the study’s focus was solely on the finger tapping task, an important, but not complete aspect of bradykinesia evaluation. Other key tasks in assessing bradykinesia, such as hand movements and leg agility, were not included. Furthermore, PD involves a variety of motor symptoms beyond bradykinesia. Future research will aim to extend our video-based approach to cover a broader range of motor tasks from the UPDRS, enabling a more-comprehensive assessment of PD motor symptoms. Thirdly, the study’s setting was restricted to controlled clinical environments, which may not fully replicate the varied conditions in which the model might be used. Factors like changing lighting, patient positioning, and camera quality or angles, crucial for video quality, were not thoroughly examined. Additionally, our approach of manually excluding low-quality videos needs to be replaced by automated algorithms in future work. These algorithms should be capable of distinguishing between low- and high-quality videos based on predefined quality thresholds. Finally, while not directly related to our scientific methodology, the ethical aspects of patient privacy and data security are crucial. As we move towards real-world applications of this technology, a multidisciplinary approach involving legal, ethical, clinical, engineering, and patient stakeholders is essential. This collaborative effort is key to addressing the complex ethical landscape, ensuring the protection of patient privacy and data security.

7. Conclusions

The proposed innovative vision-based classification model offers a promising solution to automate UPDRS scoring accurately during the finger tapping test, a crucial aspect of motor evaluation in Parkinson’s Disease (PD). Our approach adheres to the UPDRS guidelines by extracting the relevant features and quantifying them, enabling a thorough analysis. The proposed method was designed to adapt to future changes in patient demographics, disease subtypes, and technological advancements. We are actively working to expand our data collection using a variety of devices and in different settings. This effort is crucial to enhance the model’s applicability and effectiveness across diverse contexts and technological platforms.
Our comprehensive experiments demonstrated the effectiveness and reliability of the proposed method, producing robust results with datasets obtained both in laboratory and clinical settings.

Author Contributions

Conceptualization and methodology, T.Y., M.J.M., and Z.J.W.; experiments, T.Y.; PD data collection, annotation, and curation, K.W.P. and M.J.M.; writing—original draft, T.Y. and K.W.P.; writing—review and editing, T.Y., K.W.P., Z.J.W., and M.J.M.; validation and visualization, T.Y. and K.W.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) and the Canadian Institutes of Health Research (CIHR). This research was partially supported by a Collaborative Health Research Project (CHRP) grant (MJM), and the John Nichol Chair in Parkinson’s Research.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the UBC Clinical Research Ethics Board of the Department of Medicine (H18-03548, 4/5/2019).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are not allowed to be made public according to the relevant regulations protecting patient privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balestrino, R.; Schapira, A. Parkinson disease. Eur. J. Neurol. 2020, 27, 27–42. [Google Scholar] [CrossRef] [PubMed]
  2. Fish, J. Unified Parkinson’s Disease Rating Scale. In Encyclopedia of Clinical Neuropsychology; Kreutzer, J.S., DeLuca, J., Caplan, B., Eds.; Springer: New York, NY, USA, 2011; pp. 2576–2577. [Google Scholar] [CrossRef]
  3. Goetz, C.G.; Tilley, B.C.; Shaftman, S.R.; Stebbins, G.T.; Fahn, S.; Martinez-Martin, P.; Poewe, W.; Sampaio, C.; Stern, M.B.; Dodel, R.; et al. Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS): Scale presentation and clinimetric testing results. Mov. Disord. Off. J. Mov. Disord. Soc. 2008, 23, 2129–2170. [Google Scholar] [CrossRef]
  4. Morinan, G.; Dushin, Y.; Sarapata, G.; Rupprechter, S.; Peng, Y.; Girges, C.; Salazar, M.; Milabo, C.; Sibley, K.; Foltynie, T.; et al. Computer vision quantification of whole-body Parkinsonian bradykinesia using a large multi-site population. NPJ Park. Dis. 2023, 9, 10. [Google Scholar] [CrossRef]
  5. Park, S.H.; Seo, N.Y.; Hwang, S.M.; Park, H.Y.; Jung, Y.J. Quantifying Finger-tapping-test Scores using a Three-dimensional Motion Analysis Program: A Preliminary Study. J. Magn. 2022, 27, 411–420. [Google Scholar] [CrossRef]
  6. Lu, M.; Poston, K.; Pfefferbaum, A.; Sullivan, E.V.; Fei-Fei, L.; Pohl, K.M.; Niebles, J.C.; Adeli, E. Vision-based estimation of MDS-UPDRS gait scores for assessing Parkinson’s disease motor severity. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020. Proceedings, Part III 23. pp. 637–647. [Google Scholar]
  7. Ayaz, Z.; Naz, S.; Khan, N.H.; Razzak, I.; Imran, M. Automated methods for diagnosis of Parkinson’s disease and predicting severity level. Neural Comput. Appl. 2023, 35, 14499–14534. [Google Scholar] [CrossRef]
  8. Mughal, H.; Javed, A.R.; Rizwan, M.; Almadhor, A.S.; Kryvinska, N. Parkinson’s disease management via wearable sensors: A systematic review. IEEE Access 2022, 10, 35219–35237. [Google Scholar] [CrossRef]
  9. Balakrishnan, A.; Medikonda, J.; Namboothiri, P.K.; Natarajan, M. Role of wearable sensors with machine learning approaches in gait analysis for Parkinson’s disease assessment: A review. Eng. Sci. 2022, 19, 5–19. [Google Scholar] [CrossRef]
  10. Joshi, R.; Bronstein, J.M.; Keener, A.; Alcazar, J.; Yang, D.D.; Joshi, M.; Hermanowicz, N. PKG movement recording system use shows promise in routine clinical care of patients with Parkinson’s disease. Front. Neurol. 2019, 10, 1027. [Google Scholar] [CrossRef] [PubMed]
  11. Hadley, A.J.; Riley, D.E.; Heldman, D.A. Real-world evidence for a smartwatch-based parkinson’s motor assessment app for patients undergoing therapy changes. Digit. Biomarkers 2021, 5, 206–215. [Google Scholar] [CrossRef] [PubMed]
  12. Schneider, R.B.; Myers, T.L.; Tarolli, C.G.; Amodeo, K.; Adams, J.L.; Jensen-Roberts, S.; Dorsey, E.R. Remote administration of the MDS-UPDRS in the time of COVID-19 and beyond. J. Park. Dis. 2020, 10, 1379–1382. [Google Scholar] [CrossRef]
  13. Park, K.W.; Wu, H.J.; Yu, T.; Mahal, R.; Mirian, M.S.; McKeown, M.J. Potential Pitfalls of Remote and Automated Video Assessments of Movements Disorders. Mov. Disord. 2023, 38, 504–506. [Google Scholar] [CrossRef] [PubMed]
  14. Tripathi, S.; Malhotra, A.; Qazi, M.; Chou, J.; Wang, F.; Barkan, S.; Hellmers, N.; Henchcliffe, C.; Sarva, H. Clinical Review of Smartphone Applications in Parkinson’s Disease. Neurologist 2022, 27, 183–193. [Google Scholar] [CrossRef]
  15. Jeon, H.; Lee, W.; Park, H.; Lee, H.J.; Kim, S.K.; Kim, H.B.; Jeon, B.; Park, K.S. Automatic classification of tremor severity in Parkinson’s disease using a wearable device. Sensors 2017, 17, 2067. [Google Scholar] [CrossRef]
  16. Mahadevan, N.; Demanuele, C.; Zhang, H.; Volfson, D.; Ho, B.; Erb, M.K.; Patel, S. Development of digital biomarkers for resting tremor and bradykinesia using a wrist-worn wearable device. NPJ Digit. Med. 2020, 3, 5. [Google Scholar] [CrossRef]
  17. Borzì, L.; Sigcha, L.; Rodríguez-Martín, D.; Olmo, G. Real-time detection of freezing of gait in Parkinson’s disease using multi-head convolutional neural networks and a single inertial sensor. Artif. Intell. Med. 2023, 135, 102459. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, P.; Yu, N.; Yang, Y.; Yu, Y.; Sun, X.; Yu, H.; Han, J.; Wu, J. Quantitative assessment of gait characteristics in patients with Parkinson’s disease using 2D video. Park. Relat. Disord. 2022, 101, 49–56. [Google Scholar] [CrossRef] [PubMed]
  19. Talitckii, A.; Kovalenko, E.; Shcherbak, A.; Anikina, A.; Bril, E.; Zimniakova, O.; Semenov, M.; Dylov, D.V.; Somov, A. Comparative study of wearable sensors, video, and handwriting to detect Parkinson’s disease. IEEE Trans. Instrum. Meas. 2022, 71, 2509910. [Google Scholar] [CrossRef]
  20. Ngo, Q.C.; Motin, M.A.; Pah, N.D.; Drotár, P.; Kempster, P.; Kumar, D. Computerized analysis of speech and voice for Parkinson’s disease: A systematic review. Comput. Methods Programs Biomed. 2022, 226, 107133. [Google Scholar] [CrossRef]
  21. Aouraghe, I.; Khaissidi, G.; Mrabti, M. A literature review of online handwriting analysis to detect Parkinson’s disease at an early stage. Multimed. Tools Appl. 2023, 82, 11923–11948. [Google Scholar] [CrossRef]
  22. Yang, Y.; Yuan, Y.; Zhang, G.; Wang, H.; Chen, Y.C.; Liu, Y.; Tarolli, C.G.; Crepeau, D.; Bukartyk, J.; Junna, M.R.; et al. Artificial intelligence-enabled detection and assessment of Parkinson’s disease using nocturnal breathing signals. Nat. Med. 2022, 28, 2207–2215. [Google Scholar] [CrossRef]
  23. Guo, Z.; Zeng, W.; Yu, T.; Xu, Y.; Xiao, Y.; Cao, X.; Cao, Z. Vision-based finger tapping test in patients with Parkinson’s disease via spatial-temporal 3D hand pose estimation. IEEE J. Biomed. Health Inform. 2022, 26, 3848–3859. [Google Scholar] [CrossRef]
  24. Ospina, B.M.; Chaparro, J.A.V.; Paredes, J.D.A.; Pino, Y.J.C.; Navarro, A.; Orozco, J.L. Objective arm swing analysis in early-stage Parkinson’s disease using an RGB-D camera (Kinect®). J. Park. Dis. 2018, 8, 563–570. [Google Scholar] [CrossRef] [PubMed]
  25. Muñoz Ospina, B.; Valderrama Chaparro, J.A.; Arango Paredes, J.D.; Castaño Pino, Y.J.; Navarro, A.; Orozco, J.L. Age matters: Objective gait assessment in early parkinson’s disease using an RGB-D camera. Park. Dis. 2019, 2019, 5050182. [Google Scholar] [CrossRef] [PubMed]
  26. Fernández-González, P.; Carratalá-Tejada, M.; Monge-Pereira, E.; Collado-Vázquez, S.; Sánchez-Herrera Baeza, P.; Cuesta-Gómez, A.; Oña-Simbaña, E.D.; Jardón-Huete, A.; Molina-Rueda, F.; Balaguer-Bernaldo de Quirós, C.; et al. Leap motion controlled video game-based therapy for upper limb rehabilitation in patients with Parkinson’s disease: A feasibility study. J. Neuroeng. Rehabil. 2019, 16, 133. [Google Scholar] [CrossRef]
  27. Butt, A.H.; Rovini, E.; Dolciotti, C.; De Petris, G.; Bongioanni, P.; Carboncini, M.; Cavallo, F. Objective and automatic classification of Parkinson disease with Leap Motion controller. Biomed. Eng. Online 2018, 17, 1–21. [Google Scholar] [CrossRef]
  28. Khwaounjoo, P.; Singh, G.; Grenfell, S.; Özsoy, B.; MacAskill, M.R.; Anderson, T.J.; Çakmak, Y.O. Non-Contact Hand Movement Analysis for Optimal Configuration of Smart Sensors to Capture Parkinson’s Disease Hand Tremor. Sensors 2022, 22, 4613. [Google Scholar] [CrossRef]
  29. Moshkova, A.; Samorodov, A.; Voinova, N.; Volkov, A.; Ivanova, E.; Fedotova, E. Parkinson’s disease detection by using machine learning algorithms and hand movement signal from LeapMotion sensor. In Proceedings of the 2020 26th Conference of Open Innovations Association (FRUCT), Yaroslavl, Russia, 20–24 April 2020; pp. 321–327. [Google Scholar]
  30. Sarapata, G.; Dushin, Y.; Morinan, G.; Ong, J.; Budhdeo, S.; Kainz, B.; O’Keeffe, J. Video-based activity recognition for automated motor assessment of Parkinson’s disease. IEEE J. Biomed. Health Inform. 2023, 10, 5032–5041. [Google Scholar] [CrossRef]
  31. Islam, M.S.; Rahman, W.; Abdelkader, A.; Yang, P.T.; Lee, S.; Adams, J.L.; Schneider, R.B.; Dorsey, E.; Hoque, E. Using AI to Measure Parkinson’s Disease Severity at Home. arXiv 2023, arXiv:2303.17573. [Google Scholar]
  32. Park, K.W.; Lee, E.J.; Lee, J.S.; Jeong, J.; Choi, N.; Jo, S.; Jung, M.; Do, J.Y.; Kang, D.W.; Lee, J.G.; et al. Machine learning–based automatic rating for cardinal symptoms of Parkinson disease. Neurology 2021, 96, e1761–e1769. [Google Scholar] [CrossRef] [PubMed]
  33. Scott, B.; Seyres, M.; Philp, F.; Chadwick, E.K.; Blana, D. Healthcare applications of single camera markerless motion capture: A scoping review. PeerJ 2022, 10, e13517. [Google Scholar] [CrossRef] [PubMed]
  34. Moro, M.; Marchesi, G.; Hesse, F.; Odone, F.; Casadio, M. Markerless vs. marker-based gait analysis: A proof of concept study. Sensors 2022, 22, 2011. [Google Scholar] [CrossRef]
  35. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  36. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.L.; Yong, M.G.; Lee, J.; et al. MediaPipe: A framework for building perception pipelines. arXiv 2004, arXiv:1906.08172. [Google Scholar]
  37. Fang, H.S.; Li, J.; Tang, H.; Xu, C.; Zhu, H.; Xiu, Y.; Li, Y.L.; Lu, C. Alphapose: Whole-body regional multi-person pose estimation and tracking in real-time. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7157–7173. [Google Scholar] [CrossRef] [PubMed]
  38. Vignoud, G.; Desjardins, C.; Salardaine, Q.; Mongin, M.; Garcin, B.; Venance, L.; Degos, B. Video-Based Automated Assessment of Movement Parameters Consistent with MDS-UPDRS III in Parkinson’s Disease. J. Park. Dis. 2022, 12, 2211–2222. [Google Scholar] [CrossRef]
  39. Liu, W.; Lin, X.; Chen, X.; Wang, Q.; Wang, X.; Yang, B.; Cai, N.; Chen, R.; Chen, G.; Lin, Y. Vision-based estimation of MDS-UPDRS scores for quantifying Parkinson’s disease tremor severity. Med. Image Anal. 2023, 85, 102754. [Google Scholar] [CrossRef] [PubMed]
  40. Li, H.; Shao, X.; Zhang, C.; Qian, X. Automated assessment of Parkinsonian finger tapping tests through a vision-based fine-grained classification model. Neurocomputing 2021, 441, 260–271. [Google Scholar] [CrossRef]
  41. Liang, J.; Lugmayr, A.; Zhang, K.; Danelljan, M.; Van Gool, L.; Timofte, R. Hierarchical conditional flow: A unified framework for image super-resolution and image rescaling. arXiv 2021, arXiv:2108.05301. [Google Scholar]
  42. Pan, Z.; Li, B.; He, D.; Yao, M.; Wu, W.; Lin, T.; Li, X.; Ding, E. Towards bidirectional arbitrary image rescaling: Joint optimization and cycle idempotence. arXiv 2022, arXiv:2203.00911. [Google Scholar]
  43. Li, J.; Wang, L.; Xu, B. Research on Motion Blur Image of Infrared Target Deblurring Based on Wavelet Transform. J. Phys. Conf. Ser. 2020, 1570, 012080. [Google Scholar] [CrossRef]
  44. Shi, Z.; Xu, X.; Liu, X.; Chen, J.; Yang, M.H. Video frame interpolation transformer. arXiv 2022, arXiv:2111.13817. [Google Scholar]
  45. Mazurowski, M.A.; Dong, H.; Gu, H.; Yang, J.; Konz, N.; Zhang, Y. Segment anything model for medical image analysis: An experimental study. Med. Image Anal. 2023, 89, 102918. [Google Scholar] [CrossRef]
  46. Ho, P.W.L.; Leung, C.T.; Liu, H.; Pang, S.Y.Y.; Lam, C.S.C.; Xian, J.; Li, L.; Kung, M.H.W.; Ramsden, D.B.; Ho, S.L. Age-dependent accumulation of oligomeric SNCA/α-synuclein from impaired degradation in mutant LRRK2 knockin mouse model of Parkinson disease: Role for therapeutic activation of chaperone-mediated autophagy (CMA). Autophagy 2020, 16, 347–370. [Google Scholar] [CrossRef]
  47. Kang, K.W.; Choi, S.M.; Kim, B.C. Gender differences in motor and non-motor symptoms in early Parkinson disease. Medicine 2022, 101, e28643. [Google Scholar] [CrossRef] [PubMed]
  48. Nutt, J.G.; Bloem, B.R.; Giladi, N.; Hallett, M.; Horak, F.B.; Nieuwboer, A. Freezing of gait: Moving forward on a mysterious clinical phenomenon. Lancet Neurol. 2011, 10, 734–744. [Google Scholar] [CrossRef] [PubMed]
  49. Li, Z.; Lu, K.; Cai, M.; Liu, X.; Wang, Y.; Yang, J. An Automatic Evaluation Method for Parkinson’s Dyskinesia Using Finger Tapping Video for Small Samples. J. Med. Biol. Eng. 2022, 42, 351–363. [Google Scholar] [CrossRef]
  50. Williams, S.; Wong, D.; Alty, J.E.; Relton, S.D. Parkinsonian hand or clinician’s eye? Finger tap Bradykinesia interrater reliability for 21 movement disorder experts. J. Park. Dis. 2023, 13, 525–536. [Google Scholar] [CrossRef]
  51. Nieuwboer, A.; Vercruysse, S.; Feys, P.; Levin, O.; Spildooren, J.; Swinnen, S. Upper limb movement interruptions are correlated to freezing of gait in Parkinson’s disease. Eur. J. Neurosci. 2009, 29, 1422–1430. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The main structure of the proposed framework. The pipeline consists of three distinct cycles: data cycle, model learning cycle, and algorithm deployment cycle.
Figure 1. The main structure of the proposed framework. The pipeline consists of three distinct cycles: data cycle, model learning cycle, and algorithm deployment cycle.
Sensors 23 09149 g001
Figure 2. Data preprocessing flowchart.
Figure 2. Data preprocessing flowchart.
Sensors 23 09149 g002
Figure 3. Illustration of the hand keypoints.
Figure 3. Illustration of the hand keypoints.
Sensors 23 09149 g003
Figure 4. Depiction of the time series data preprocessing pipeline. The input time series, denoted as I , corresponds to the distance between K 4 and K 8 during the finger tapping task. The X−axis indicates the frame index within an image sequence, reflecting the duration of the finger tapping action. The Y−axis represents the normalized amplitude of finger movement, adjusted according to the palm size of each subject.
Figure 4. Depiction of the time series data preprocessing pipeline. The input time series, denoted as I , corresponds to the distance between K 4 and K 8 during the finger tapping task. The X−axis indicates the frame index within an image sequence, reflecting the duration of the finger tapping action. The Y−axis represents the normalized amplitude of finger movement, adjusted according to the palm size of each subject.
Sensors 23 09149 g004
Figure 5. Extraction of F a m p α 1 , F a m p α 2 , and F a m p b p .
Figure 5. Extraction of F a m p α 1 , F a m p α 2 , and F a m p b p .
Sensors 23 09149 g005
Figure 6. Illustration of halt and hesitation. The red row indicates the precise moment when the halt and hesitation occur.
Figure 6. Illustration of halt and hesitation. The red row indicates the precise moment when the halt and hesitation occur.
Sensors 23 09149 g006
Figure 7. Correlations between the 15 extracted features and UPDRS scores.
Figure 7. Correlations between the 15 extracted features and UPDRS scores.
Sensors 23 09149 g007
Figure 8. Structure of the proposed Decision Tree.
Figure 8. Structure of the proposed Decision Tree.
Sensors 23 09149 g008
Figure 9. Confusion matrix of the predicted score labels and the true labels.
Figure 9. Confusion matrix of the predicted score labels and the true labels.
Sensors 23 09149 g009
Table 1. MDS-UDPRS criteria for the finger tapping task.
Table 1. MDS-UDPRS criteria for the finger tapping task.
UPDRS ScoreSymptoms
0: NormalNo problems.
1: SlightAny of the following:
a. one or two interruptions or hesitations;
b. slight slowing;
c. the amplitude decrements near the end.
2: MildAny of the following:
a. 3 to 5 interruptions during tapping;
b. mild slowing;
c. the amplitude decrements midway in the 10-tap sequence.
3: ModerateAny of the following:
a. more than 5 interruptions or at least one longer arrest (freeze) in ongoing movement;
b. moderate slowing;
c. the amplitude decrements starting after the 1st tap.
4: SevereCannot or can only barely perform the task because of slowing, interruptions, or decrements.
Table 2. Overview of selected studies on video-based bradykinesia assessment in Parkinson’s disease: “Input” specifies the type of video format utilized; “Pose Estimation” indicates the software library employed for human pose analysis; “Extracted Features” pertains to the kinematic attributes obtained from the pose estimation data over time; “Method” details the specific machine learning approach used for categorizing the data; “Result” summarizes the outcomes of the classification process; “Tasks” outlines the specific UPDRS activities evaluated in each study.
Table 2. Overview of selected studies on video-based bradykinesia assessment in Parkinson’s disease: “Input” specifies the type of video format utilized; “Pose Estimation” indicates the software library employed for human pose analysis; “Extracted Features” pertains to the kinematic attributes obtained from the pose estimation data over time; “Method” details the specific machine learning approach used for categorizing the data; “Result” summarizes the outcomes of the classification process; “Tasks” outlines the specific UPDRS activities evaluated in each study.
Authors, Year [Ref]InputPose EstimationExtracted FeaturesMethodResultsTasks
Z. Guo et al., 2022 [23]RGB–depthA2JAmplitude, velocitySVMSeverity levelFinger tapping
Morinan et al., 2023 [4]RGBOpenpose11 features representing speed, amplitude, hesitations, and decrementRandom ForestBinary/ordinal classificationFinger tapping
Sarapata et al., 2023 [30]RGBOpenposeRegion of interest extraction of motor tasks, followed by kinetic feature extraction for each taskRandom ForestBinary/ordinal classificationFinger tapping, hand movement, pronation/supination, toe tapping, leg agility, rising from a chair, gait
Islam et al., 2023 [31]RGBMediaPipe53 features representing speed, amplitude, hesitations, and decrementLightGBMRegressionFinger tapping
Proposed methodRGBMediaPipe15 features representing demographics, amplitude, velocity, halt and hesitations, and decrementDecision TreeSeverity levelFinger tapping
Table 3. Fifteen features extracted from the finger tapping time series.
Table 3. Fifteen features extracted from the finger tapping time series.
FeaturesDescription
F a g e Age of the patients
F g e n Gender of the patients
F a m p v a r Variance of the amplitude
F a m p a v g Mean of the amplitude
F a m p b p Breakpoint of the amplitude
F a m p α 1 Slope 1 of the amplitude
F a m p α 2 Slope 2 of the amplitude
F v e l v a r Variance of the velocity
F v e l a v g Mean velocity of the finger movement
F v e l b p Breakpoint of the velocity
F v e l α 1 Slope 1 of the velocity
F v e l α 2 Slope 2 of the velocity
F f r e q Frequency of finger tapping
F h h Number of halts and hesitations
F p Number of peaks of the time series
Table 4. Baseline characteristics of the participants (n = 50) and the videos (n = 75).
Table 4. Baseline characteristics of the participants (n = 50) and the videos (n = 75).
Baseline Characteristics
Age, years
    Median71
    Mean (SD)69.65
Sex
    Female/Male, n/n (%/%)19/31(38/62)
Need assistance for walking, n (%)2 (4)
Dominant side of Parkinsonism
    Left23
    Right27
UPDRS Score
    0/1/2/3/4, n8/21/18/17/11
Table 5. Inter-annotator agreement analysis in the proposed task.
Table 5. Inter-annotator agreement analysis in the proposed task.
CompleteAcceptableNone
52%97%3%
Table 6. Performance results of UPDRS classification using different classifiers.
Table 6. Performance results of UPDRS classification using different classifiers.
MethodEvaluation Metric
Accuracy(t1)Accuracy(t2)PrecisionRecallF1
AEMPD-FCN [49]0.2670.600.180.190.173
FTTST [23]0.4670.800.500.600.513
b a s e l i n e - S P 0.4670.7330.4800.600.509
F 15 - S V M - G A 0.5330.8670.5530.5900.556
F15- R F 0.5330.8670.5330.5520.540
F 15 - D T 0.800.9330.8430.800.797
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, T.; Park, K.W.; McKeown, M.J.; Wang, Z.J. Clinically Informed Automated Assessment of Finger Tapping Videos in Parkinson’s Disease. Sensors 2023, 23, 9149. https://doi.org/10.3390/s23229149

AMA Style

Yu T, Park KW, McKeown MJ, Wang ZJ. Clinically Informed Automated Assessment of Finger Tapping Videos in Parkinson’s Disease. Sensors. 2023; 23(22):9149. https://doi.org/10.3390/s23229149

Chicago/Turabian Style

Yu, Tianze, Kye Won Park, Martin J. McKeown, and Z. Jane Wang. 2023. "Clinically Informed Automated Assessment of Finger Tapping Videos in Parkinson’s Disease" Sensors 23, no. 22: 9149. https://doi.org/10.3390/s23229149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop