Next Article in Journal
Improved Quasi-Z-Source High Step-Up DC–DC Converter Based on Voltage-Doubler Topology
Next Article in Special Issue
Movement Recognition through Inductive Wireless Links: Investigation of Different Fabrication Techniques
Previous Article in Journal
Building a COVID-Safe Navigation App Using a Meta-Model Based Context Server
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Inertial Sensor-Based Activity Recognition in Neurological Populations

1
Department of Computer and Information Sciences, Northumbria University, Newcastle upon Tyne NE1 8ST, UK
2
Department of Electrical and Electronics Engineering, Karamanoglu Mehmetbey University, Karaman 70100, Turkey
3
Department of Sport, Exercise and Rehabilitation, Northumbria University, Newcastle upon Tyne NE1 8ST, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9891; https://doi.org/10.3390/s22249891
Submission received: 24 November 2022 / Revised: 14 December 2022 / Accepted: 14 December 2022 / Published: 15 December 2022
(This article belongs to the Special Issue Intelligent Wearable Sensor-Based Gait and Movement Analysis)

Abstract

:
Inertial sensor-based human activity recognition (HAR) has a range of healthcare applications as it can indicate the overall health status or functional capabilities of people with impaired mobility. Typically, artificial intelligence models achieve high recognition accuracies when trained with rich and diverse inertial datasets. However, obtaining such datasets may not be feasible in neurological populations due to, e.g., impaired patient mobility to perform many daily activities. This study proposes a novel framework to overcome the challenge of creating rich and diverse datasets for HAR in neurological populations. The framework produces images from numerical inertial time-series data (initial state) and then artificially augments the number of produced images (enhanced state) to achieve a larger dataset. Here, we used convolutional neural network (CNN) architectures by utilizing image input. In addition, CNN enables transfer learning which enables limited datasets to benefit from models that are trained with big data. Initially, two benchmarked public datasets were used to verify the framework. Afterward, the approach was tested in limited local datasets of healthy subjects (HS), Parkinson’s disease (PD) population, and stroke survivors (SS) to further investigate validity. The experimental results show that when data augmentation is applied, recognition accuracies have been increased in HS, SS, and PD by 25.6%, 21.4%, and 5.8%, respectively, compared to the no data augmentation state. In addition, data augmentation contributes to better detection of stair ascent and stair descent by 39.1% and 18.0%, respectively, in limited local datasets. Findings also suggest that CNN architectures that have a small number of deep layers can achieve high accuracy. The implication of this study has the potential to reduce the burden on participants and researchers where limited datasets are accrued.

1. Introduction

Human activity recognition (HAR, also termed activity pattern/classification) investigates objective detection of daily activities such as level walking or stair ascent [1,2,3]. HAR in neurological populations to identify periods of activity is important as it enables clinicians to better understand patients’ functional abilities, which may inform treatment or prognosis [4]. More broadly, HAR has previously been adopted in healthcare applications such as mobility and fall detection in older adults [5], adolescents with cerebral palsy [6], and stroke survivors (SS) [7] to better understand the quality of life-related outcomes.
Camera and radar-based technologies are utilized in HAR applications but are limited due to high cost, privacy issues, and computational requirements [1,8,9]. Alternatively, low-cost and lightweight wearable inertial measurement units (IMUs: accelerometer and gyroscopes) enable researchers to cost-effectively quantify longitudinal mobility data in controlled and/or free-living environments (e.g., home) [3]. Wearable IMUs [1,8,10,11] (occasionally integrated with other sensing modalities, e.g., magnetometer [12], electrocardiograph [13,14], and electromyography [15]) with contemporary classification architectures [10,11] provide highly accurate HAR. Typically, accelerometers are the dominant inertial sensor for HAR, but the inclusion of a gyroscope increases recognition accuracies by providing data from rotational activities of the trunk or legs, such as during turning, stair ascent, and descent [3]. Often, labeling wearable-based HAR data in clinics is performed manually because of the controlled conditions [16], i.e., clearly defined (scripted) periods of walking with time stamps. Moving beyond the lab creates challenges that greatly impact the practicality and use of manual segmentation, such as vast amounts of unlabeled data [11]. Consequently, artificial intelligence (AI) such as deep learning (DL) based approaches have become key to automatically identifying daily/habitual activities [17,18], fall detection [19], and negating time-consuming manual segmentation and data labeling.
Performances of automated HAR approach depend on the complexity level of the recognition process and the predictive capacity of the AI recognition models adopted since each individual tends to perform activities in different ways due to, e.g., habits, personal preferences, age, and health [3]. Studies show that wearable IMUs attached to people with neurological conditions generate different acceleration and angular velocity signals than healthy controls [20], and having such diverse data cause intra-class variations which impact model performance [21,22]. Previously, models that were trained with data belonging to healthy participants demonstrated significant drops in HAR accuracy when classifying activities of Stroke survivors (SS) [23] and people with Parkinson’s disease (PD) [24]. In order to overcome such limitations, previous studies suggested generating HAR models for each user profile [3], e.g., training a model specifically for SS and those with PD. However, accurate HAR of daily activities requires a diverse and balanced dataset [8]. Previous studies reported that existing public datasets have either a limited number of activities and participants or include data belonging to limited user profiles or limited and unbalanced data of neurological populations [3,8]. This can be attributed to: participants having difficulty performing certain daily activities due to poor mobility, challenges of data collection in healthcare due to privacy issues [10] and/or; establishment of multidisciplinary teams to aid patient/participant recruitment that are well characterized i.e., with clinical notes/records.
In this paper, we propose a methodology to investigate how limited data can be better utilized to achieve accurate HAR/mobility classification in limited healthy, PD, and SS population-specific models. To the authors’ best knowledge, this is the first study that aims to solve low HAR performances in limited datasets of neurological populations. To achieve our goal, we propose numerical-to-image conversion as the fundamental component within our proposed methodology. The use of data augmentation complements our framework by providing solutions to the limited dataset and overfitting problems. Finally, using transfer learning enable applications with small data to benefit from models that are more experienced and trained with big data. An investigation of the proposed method’s performance was initially performed on two public datasets. Results were compared to the reference studies with and without data augmentation operations in the same datasets. Then, several pilot studies tested our numerical-to-image conversion approach along with a data augmentation technique on limited local datasets belonging to healthy, PD, and SS participants. Therefore, the contributions of this study are:
  • Developing a novel framework that converts inertial sensor time-series data into images (activity images).
  • Adopting established data augmentation techniques in image processing to artificially increase limited datasets for the purpose of better HAR in neurological populations (where access to data may be difficult).
  • Verifying the proposed approach in public datasets and conducting experimental pilot studies for a single sensor-based HAR on limited HS, PD, and SS datasets.

2. Related Works

Machine learning (ML) algorithms, such as Support Vector Machine (SVM) or Decision Tree (DT), rely on manual feature extraction and selection that greatly impact HAR accuracy. Prior works have shown that designing hand-crafted features in a specific application requires human-based domain knowledge [25], and heuristically-defined features may perform well in recognizing one activity but not others [26]. Furthermore, hand-crafted features may not be sensitive to targeted cohorts and environments [27], i.e., models developed with a set of features in a lab lose accuracy when applied in freeliving (beyond the lab) due to the diversity of user’s habitual behavior and complexity of activities and environments. Equally, human expertise may not always select the best features, which can decrease accuracy and make it necessary to apply additional feature selection methods to reduce dimensionality [3]. The use of ensemble classifiers has been recommended to increase classification accuracy [28,29], but studies utilized complex methods that were computationally inefficient. In order to optimize performance, IMU-based HAR approaches have generally converged on DL [8]. DL algorithms are capable of generating complex and high-level features that well represent raw data and do not require expert knowledge for feature extraction and selection [3,30]. DL methods are considered state-of-the-art in computational processing [31] and have provided very accurate classification approaches [2,22].
Common DL approaches include Convolutional Neural Networks (CNN), which are able to learn multiple layers of feature hierarchies to provide high accuracy for the recognition of repetitive activities with a long duration [8]. Compared to other AI methods, CNNs have a local dependency, an ability to identify the correlation between close signals and scale invariance with an ability to work with different frequencies in time series data [2]. CNN models have been used with other AI methods, such as Long-short-term memory (LSTM) recurrent neural networks, to capture time dependencies on features extracted by convolution operations. This kind of combined architecture outperformed other studies that used the same HAR dataset [32]. Additionally, spectrogram-based feature extraction methods using Short-Time Fourier transform (STFT) from raw IMU data have been proposed through data augmentation with down sampling and shuffling techniques before classification with LSTM [33].
In both ML and DL models, the variety and size of data have the utmost importance in minimizing overfitting. Failing to provide a diverse and large data set will cause training and validation errors. Data augmentation is a powerful method to solve training, validation errors, overfitting [34,35], and data sparsity problems. Previously, a two-stage end-to-end CNN model was proposed along with an augmentation technique to enhance datasets by inserting data points via linear interpolation [36]. The results of the proposed methodology outperformed previous studies in terms of classifying activities in a dataset of healthy participants. Another study used two different time series data augmentation techniques to investigate the impact on accuracy and reported that the use of data augmentation significantly enhances recognition accuracy in three public datasets of healthy participants [37]. Alternatively, the Generative Adversarial Network (GAN) framework [38] was adopted to generate more data samples. Although GAN could improve the performance of classifiers with limited labeled data, weaknesses such as lack of explicit representation of the generator’s distribution and the need for model synchronization were reported [10]. Synthetic Minority Over-sampling technique (SMOTE) is another technique that uses oversampling to generate more data samples [39] and achieves better classifier performances in ML classifiers (such as Naive Bayes) but has not been fully investigated in DL classifiers and HAR of neurological populations.
Interpretation of numeric IMU data as images has been implemented in very few HAR studies. In [5], IMU data was stacked row by row into an array (called a signal image) before a 2D Discrete Fourier transform (DFT) was applied to generate activity images which were then input to a CNN. Elsewhere, frequency (activity) images were created from the raw IMU signals by applying STFT [22] and Fast Fourier Transform (FFT) [40] before being used as input to a CNN. However, the referenced studies performed HAR using activity images (spectrum) rather than a direct representation of numerical sensor values. Although these studies produced accurate HAR, the images (spectrum) used do not fully represent raw sensor data. Using raw sensor data to create images where pixel brightness increases/decreases with the numerical value of the IMU is a novel and potentially more accurate alternative as it better represents raw (sample level) IMU data. Previously, images that were created with this approach provided very promising classification results of the survival status of the patient using a clinical record dataset [41].

Inertial Sensor-Based HAR in Neurologic Populations

The use of inertial sensors in HAR eliminates immediate privacy and security concerns and offers pragmatic data collection possibilities via various technologies such as commercially available devices, smartphones, and smartwatches. Despite providing unique opportunities, inertial sensor-based HAR also poses many challenges, such as accurately recognizing the activity type from an unknown environment using an inertial signal [1]. Unlike camera-based HAR systems, inertial sensor-based HAR requires additional mechanisms such as video recording or scripted data collection protocol to label the data before training. Another challenge posed by inertial sensor-based HAR is the requirement of wearing multiple sensors. Although multiple inertial sensors-based HAR has provided highly accurate activity classification [22], wearing multiple devices may cause discomfort while increasing computation and project costs. Accordingly, most studies utilize a single waist-mounted sensor [42].
Several publicly available benchmark datasets have been generated using a single sensor configuration to enable researchers to develop highly accurate HAR models [43,44]. However, those datasets were produced from healthy people only [2]. The lack of HAR benchmarking datasets for neurological populations force researchers to create local (project-specific) datasets. The creation of a local dataset that has diverse and sufficient data is challenging due to several reasons [10]. For example, researchers interested in HAR within neurological disorders may struggle with patient recruitment (due to a lack of clinical partners) or ensure the longevity of recording to obtain sufficient data due to a lack of patient adherence. Additionally, data may be skewed as those with functional limitations may generally perform light activities only, such as level ground walking rather than stair ascent/descent or walking over uneven terrain due to fear of falling. These real-life implications result in datasets of SS [27,45], PD [24], and people with spinal cord injury [46] that may not be rich and diverse enough to achieve very high HAR accuracies on new data.
Accurate HAR in neurological populations requires diverse data from multiple participants with a broad range of ages, fitness levels, disease duration, mood, and health conditions to ensure inter-subject and intrasubject variability have minimal impact on recognition accuracy [47]. For example, people with different stroke types (e.g., ischemic, hemorrhagic) and post-stroke recovery durations may show different levels of impaired mobility during stair ascending/descending. Increasing the size of the dataset may also contribute to minimizing the impact of subject variability in classification models.
In this study, we hypothesize that converting numerical sensor data into activity images and implementing data augmentation techniques can alleviate diversity and data balance issues, thereby increasing the performance of DL methods by utilizing well-established techniques in image processing [48]. The use of image data for training and testing models makes CNN models a viable choice because CNN models not only extract high-level features from images but also present more compactly and robustly what the image essentially represents.

3. Methodology

The proposed methodology developed for better HAR of people with neurological conditions is presented in Figure 1. Three limited local datasets and two independent benchmarking public datasets were used to verify the proposed methodology. In order to replicate the pragmatic problems in this domain, the local dataset has a limited number of participants, data sparsity, and class imbalance. In the proposed methodology, numerical inertial sensor data were first normalized and then converted into images (initial state). Then, established image augmentation techniques were adopted to artificially increase the number of images (enhanced state). Finally, generated images were fed into different CNN architectures. All steps are further detailed in this section.

3.1. Data Normalization and Numerical to Image Conversion (Initial State)

Raw accelerometer and gyroscope signals experience different lower and upper limits because of configuration (e.g., an accelerometer typically can collect data in the range of ±16 m/s2, whereas gyroscopes can sense up to ±2000°/s). Normalizing features with different upper and lower limits is a commonly used pre-process in AI, as extreme differences between different features may have a negative impact on learning abilities [49]. In the normalization step, a feature scaling-based normalization method is preferred due to its convenience. Here, raw IMU data ( x ) is normalized ( x ^ ) considering max value ( x m a x ) and min value ( x m i n ), as depicted in Figure 1c. As a result of normalization, the value in matrices ranges between 0 and 1 for both accelerometer and angular velocity, Equation (1)
x ^ = x x m i n x m a x x m i n
After normalization, data were divided into sub-segments (windows), considering each sub-segment should contain sufficient characteristics that allow HAR to be successfully performed. A previous study [50] investigated windows size impact on HAR application and reported that the ideal size for fixed windows ranges between 2 s and 5 s considering a frequency of 20 Hertz (Hz) to 50 Hz. Therefore, each activity was divided into consecutive segments of fixed-length (≈2.5 s windows), considering that at least two strides are needed to recognize walking and stair ambulation. IMUs typically sense tri-axial acceleration ( a x , a y , a z ) and tri-axial angular velocity ( w x , w y , w z ) in the t moment (Equation (2)). Generally, popular CNN models are not suitable to use 1D datasets and require 2/3D images to feed input layers [51]. Therefore, many previous studies [32,52,53] extract IMU data features with 1D convolution layers and then evaluate those features with recurrent neural network-based methods. Here, we convert numerical IMU data to images to go beyond that limit, as shown in Figure 1d.
IMU t = [ a x t , a y t , a z t , w x t , w y t , w z t ]
Equation (3) represents 2D data (also can be referred to as an image) created by vertical placement of accelerometer and gyroscope values recorded in 2.50 s window/250 sample and 2.56 s windows/128 sample for the local dataset and UCI HAR dataset, respectively. In the WISDM dataset, only accelerometer values were placed in a 2.50 s window/50 sample. Unlike previous studies [5,22,40], this study ensures that each numerical IMU value corresponds to a specific pixel in an image. The normalized values in the matrices were multiplied by 255 to produce grey images with pixels ranging from 0 to 255. As a result, images whose brightness increases/decreases with the numerical value of the IMU are produced. However, image dimensions are not suitable to feed the input layer of CNN models since each CNN model’s input layer accepts images with a size of 224 × 224 [51]. Therefore, resizing is applied by stretching the row length to obtain a square matrix from these images Figure 1e.
[ IMU acc x IMU acc y IMU acc z IMU gyro x IMU gyro y IMU gyro z ] = [ a x t , a x t + 1 , a x t + 2 , , a x t + 126 / 248   , a x t + 127 / 249 a y t , a y t + 1 , a y t + 2 , , a y t + 126 / 248   , a y t + 127 / 249 a z t , a z t + 1 , a z t + 2 , , a z t + 126 / 248   , a z t + 127 / 249 w x t , w x t + 1 , w x t + 2 , , w x t + 126 / 248   , w x t + 127 / 249 w y t , w y t + 1 , w y t + 2 , , w y t + 126 / 248   , w y t + 127 / 249 w z t , w z t + 1 , w z t + 2 , , w z t + 126 / 248   , w z t + 127 / 249 ]

3.2. Data Augmentation

Table 1 presents the number of occurrences along with class distribution in limited local datasets. In order to alleviate the problems related to small dataset size and prevent overfitting; data augmentation was applied to increase the number of generated images using established image processing techniques. In this sense, four different image position augmentation techniques (reflection, rotation, scale, and translation) were applied to each image to ensure data diversity and robust training, see Figure 1f. Reflection, also known as symmetry, is an image pre-processing operation that can occur in horizontal or vertical access. Rotation, scaling, and translation is other pre-processing operations that deal with spinning, resizing, and moving (right, left, up, and down) in given upper and lower limits, respectively. The lower and upper limit values of rotation, translation (pixel), and scale are ±30°, ±10°, and 0.9–1.1, respectively, since these values have proved to be efficient [41]. Consequently, the size of the original dataset in the initial state was enhanced by adding 8 times more artificial data (4 different techniques with lower and upper limits). In this context, the number of occurrences for each class in the local datasets is increased, Table 2.

3.3. HAR via CNN

Benchmarking analysis of various deep learning models was previously studied, and performance indices such as accuracy, model complexity, memory usage, computing power, and interference times were evaluated [51,54]. We determined our priority performance indices as high accuracy rate, minimal computing power, and short prediction time to achieve an effective HAR framework. Therefore, we chose four optimal pre-trained networks GoogleNet [55], ResNet18 [56], ResNet50 [56], and MobileNet-v2 [57,58], in the Pareto frontier as these architectures satisfy our requirements. Each CNN architecture used in this study differs from the others in layer, size, and parameters and is often preferred in benchmarking studies to evaluate CNN performances [59,60], Table 3. MATLAB® (2021, MathWorks, Inc., Natick, MA, USA) software on a laptop with Intel Core i7-7700HG CPU (2.80 GHz), 16 GB RAM, NVIDIA GeForce GTX 1050 4 GB was used to perform CNN training and testing.
A residual network (ResNet) [56] was developed to improve unexpected low performances of deeper network architectures by adding a skip connection (shortcut) to convey information between layers and avoid the vanishing gradient problem [60]. There are different ResNet variants (18-layer, 34-layer, 50-layer, 101-layer, 152-layer) proposed considering the number of layers and output sizes. ResNet18 and ResNet50 were implemented here. MobileNet was employed as it has low computation and fast operation by using depth-wise separable convolutions to reduce the number of parameters and computation time. Specifically, MobileNet-v2 [58] was implemented, which has 54-layers, distinguishing it from MobileNet in using inverted residual blocks with bottleneck properties. GoogleNet [55] is 22-layer deep (excluding pooling) model designed with computational efficiency and practicality. It uses the inception module to extract features more effectively using various filter sizes. And the computational load is reduced with a 1 × 1 convolution of the depth of the network. Minor adjustments, such as the use of fine-tuning networks were made to the existing architecture for the four-class classification problem in this study. In this context, a fully connected layer with four outputs and a classification layer was added to the existing structure, see Figure 1g.

4. Datasets

4.1. Local Datasets

Ten HS (28.4 ± 7.0 years, 79.2 ± 14.4 kg, 176.8 ± 8.4 cm, 8 Male, M: 2 Female, F), five people with PD (61.5± 3.43 years, 82.9 ± 10.3 kg, 175.8 ± 4.6 cm, 5M) and three SS (72.3 ± 3.1 years, 78.5 ± 12.1 kg, 176 ± 8.2 cm, 3M) were recruited, as illustrated in Figure 1a. Each participant was instructed to stand for 2-min (eyes open and comfortable standing), then walk over the level ground for 2-min around a 20-m circuit at their self-selected walking speed inside the lab. Afterward, participants ascended and descended stairs (15 steps) outside of the lab (in a generic university campus stairwell).
Assessment and instrumentation were carried out by a physiotherapist and a trained researcher, respectively. Ethical consent was granted by the Northumbria University Research Ethics Committee (REF: 21603). All participants gave informed written consent before participating in this study. Testing took place inside and outside of a gait laboratory/lab, Coach Lane Campus, Northumbria University, Newcastle, upon Tyne.
Each participant wore a Shimmer3 IMU device (5.1 cm × 3.4 cm × 1.4 cm, 23.6 g) on the 5th lumbar vertebrae (L5), as shown in Figure 1b. IMU signals (tri-axial accelerometer and tri-axial gyroscope) were recorded at a sampling frequency of 100 Hz and configured with 16-bit resolution (±8 g, ±500°/s). IMU data were transferred to a workstation (Windows 10) from the IMU device via proprietary software (Consensys, Shimmer). Labeling of activities in a continuous data stream was conducted via a wearable camera for PD and SS, whereas a scripted experimental protocol was used for HS. All participants performed the same protocol. Inertial data streams for each activity were segmented into 2.5 s (250 sample points) windows with 50% overlap using a sliding window.

4.2. UCI-HAR and WISDM Independent Benchmarking Datasets

UCI-HAR dataset [44] was preferred to test the development methodology as it was created using the same data collection protocol as the local dataset. UCI-HAR dataset has an accelerometer and gyroscope recording of 30 HS (19–48 years), collected by a device attached at waist level. The dataset was randomly portioned into training and testing. Data were recorded at a sampling frequency of 50 Hz and segmented to fixed-width sliding windows of 2.56 s (128 sample points) with 50% overlap. The WISDM dataset was created from 36 HS under controlled laboratory conditions. The dataset has tri-axial accelerometer readings only recorded at 20 Hz. Accelerometer recordings were segmented to fixed-width sliding windows of 2.50 s with 50% overlap.
Table 4 presents activity classes along with class distributions in the benchmarking datasets. Skewed class distributions are present in the public datasets. This typically limits the learning/training process by causing class overlapping, small sample size, or small disjuncts [61]. In addition, models trained with imbalanced datasets are often biased towards the majority class, and therefore there is a greater misclassification rate for the minority class occurrences such as sitting and standing in the WISDM dataset [62]. Furthermore, the most common evaluation metric, accuracy, treats all classes as equally important, which makes it inefficient [3]. To alleviate the limitations of imbalanced public datasets, we utilized 500 occurrences from each class for training in public datasets. In total, 3000 occurrences were utilized for each dataset, and the train/test split ratio, along with occurrence numbers, are presented in Section 5.

5. Analytical Procedures

This section presents the results of the classification models in the initial state (after numerical to image conversion) and enhanced state (after data augmentation). In local datasets, 80% of the data (occurrence/images) were used for training and 20% for testing. In UCI-HAR and WISDM public datasets, a total of 3000 occurrences (500 for each class) were utilized for each dataset, where 80% (2400 occurrence/images) were used for training, and 20% (600 occurrence/images) were used for testing. Five quantitative metrics are used to evaluate the performance of each model, Equations (4)–(8). Accuracy is the most common metric and gives a general representation of model performance, but it can be inefficient when used in unbalanced datasets. Accordingly, sensitivity and specificity were also calculated as additional evaluation matrices to evaluate classes separately. F1-measure deals with a score resulting from the combination of precision and recall value, where TP: true positive, TN: true negative, FP: false positive, and FN: false negative. In addition, Matthew’s correlation coefficient (MCC) was included as it includes TN, unlike F1-measure. Total execution time was also calculated for enhanced states of all models.
Accuracy = TP + TN TP + FP + TN + FN × 100
Sensitivity = TP TP + FN
Specicificity = TN TN + FP
F 1 score = 2 × TP 2 × TP + FP + FN
Matthew s   correlation   coefficient ( MCC ) = ( TP × TN ) ( FP × FN ) ( TP + FN ) × ( TN + FP ) × ( TP + FP ) × ( TN + FN )

6. Results

6.1. UCI-HAR Datasets

Table 5 presents the results of performance metrics for initial and enhanced states in UCI-HAR. In the initial state, ResNet18 architecture slightly outperformed its counterparts in all performance metrics. Moreover, the data augmentation operation provided slight improvements in the performance metrics of each architecture, whereas the largest improvement was observed in GoogleNet. In the enhanced state, ResNet50 architecture provided slightly higher performances compared to other CNN architectures and reached 97% accuracy. However, comparing execution time reveals that GoogleNet classifies HAR activities faster than its counterparts. Table 6 presents the ResNet50 confusion matrix of the UCI-HAR dataset in the initial and enhanced states as it outperforms other architectures in terms of all performance metrics except execution time. Here, notable improvements are observed after data augmentation, especially in static activities (sitting, standing, and laying).

6.2. WISDM Datasets

Table 7 presents the classification results of the four CNN architectures using the WISDM dataset in initial and enhanced states. In the initial state, ResNet50 architecture classified HAR activities better than ResNet18 and MobileNet-v2, whereas GoogleNet showed a notably poorer performance. However, this is not valid for specificity metrics which experienced similar values in all architectures. After data augmentation is implemented, significant improvements are observed in all architectures. ResNet18 reached 95.8% accuracy with the shortest training time, whereas ResNet50 and MobileNet-v2 provided slightly lower accuracies but in a much longer time (≥130 min). Although GoogleNet is improved in its enhanced state, it is still the poorest in activity recognition compared to other architectures. Table 8 presents the confusion matrix for the best-enhanced state (ResNet18). Comparing activity recognition performances for each class in the initial and enhanced state reveals that the largest improvements are obtained in the accurate recognition of static activities (sitting and standing).

6.3. Local Datasets (HS Model)

Table 9 shows the initial and enhanced state results of HAR in the local dataset created from HS. In the initial state, MobileNet-v2 architecture outperforms its counterparts in terms of each performance metric, whereas GoogleNet architecture performs poorly in recognition of HAR activities. Significant improvements are observed in the enhanced state where ResNet50 reaches the highest accuracy with 100%, especially GoogleNet accuracy is more than doubled in the enhanced state. Table 10 presents the confusion matrix created from ResNet50 architecture, which experienced misclassification in recognition of stair activities in the initial state. After data augmentation, ResNet50 architecture better-adopted stair classes and corrected the misclassifications.

6.4. Local Datasets (PD Model)

Table 11 presents the initial and enhanced results of HAR in those with PD. In the initial state, all CNN architectures experience comparable results where ResNet18 and ResNet50 outperform other architectures. Later in the enhanced state, notable improvements were observed in all architectures, but MobileNet-v2 achieved the highest performance. Table 12 presents a confusion matrix belonging to the classification result of MobileNet-v2, where misclassification in stair descent and walking activities were improved in the enhanced state.

6.5. Local Datasets (SS Model)

Table 13 shows performances from initial and enhanced states in the local SS dataset. In the initial state, ResNet18, ResNet50, and MobileNet-v2 experience accuracies just above 70%, whereas GoogleNet shows the poorest performance with 65.7% accuracy. In the enhanced state, all architectures except GoogleNet experience significant improvements and reach over 95% accuracy. On the other hand, GoogleNet also experiences improvements but with a small margin compared to its counterparts. Table 14 presents the confusion matrix of ResNet50 from initial and enhanced states. In the SS group, stair ascent occurrences were mostly misclassified, whereas stair descent and walking activities suffered from low recognition. In the enhanced state, notable improvements were observed, especially in stair activities.

7. Discussion

The computational performance of the framework was deemed acceptable for data preparation (normalization, generally having low computational cost). Specifically, normalization of each segmented IMU window took approx. 5.4 milliseconds which was then converted into the activity image within approx. 2.1 milliseconds are, resulting in total data preparation for each occurrence of about 7.5 milliseconds. However, model training was prolonged and is discussed in Section 7.3, Limitations. Here, we first verify the proposed approach in benchmarking datasets and compare it with reference studies, Section 7.1. This tests whether the proposed numerical-to-image conversion approach is a valid and reliable approach in independent datasets. Results suggest that the proposed framework can classify activity classes in both benchmarking datasets with high accuracy, especially after data augmentation. The pre-trained networks used in this study can achieve better or comparable classification accuracies against reference studies even when the networks are trained with a portion of the original datasets.
After promising results are obtained in benchmarking datasets, we provide an evaluation regarding the pilot studies (in HS, PD, and SS), which test the proposed approach (numerical to image conversion and data augmentation) on limited local datasets. In addition, we present an analysis regarding why some CNN architectures perform better than others and recommend the necessary properties a pre-trained network needs to achieve sufficient learning.

7.1. Verification of the Results in Public Datasets

Table 15 compares the proposed framework against several reference studies with and without data augmentation in the same public datasets. Overall, numerical-to-image conversion, along with data augmentation, significantly improves the performance of CNN architectures in HAR. This study utilized 500 occurrences/instances for each class to provide unbiased evaluation metrics, as detailed in 4.2. Therefore, our findings should be considered in this context.

7.1.1. UCI-HAR Dataset

Comparing our initial results with a reference study [37] initial results in the same dataset reveals that the proposed numerical-to-image conversion approach is an effective method. Here, ResNet18 architecture reaches 93.3 % accuracy, which is superior to 80% accuracy [37]. In the enhanced state of the UCI-HAR dataset, the methodology proposed here provides similar or better results compared to the reference studies, Table 15. Comparing the training times with a reference study [37] that uses an exponential smoothing augmentation technique reveals that our approach reaches 97.0% accuracy in 166 min training duration, whereas the reference study reaches 97.9% accuracy in 210 min. This suggests that the proposed framework can provide comparable accuracies with smaller training data with shorter durations. The difference in the training times could be attributed to the preferred data augmentation technique. For example, the exponential smoothing approach assigns exponentially decreasing weights for older observations. However, our framework uses raw numerical data to produce activity images that are independent of the numerical values in the data stream. Producing images (e.g., activity images or spectrogram) directly from raw sensor data was proved to be effective in HAR [5,22,40].

7.1.2. WISDM Dataset

In the initial state, our numerical-to-image conversion technique with ResNet50 reaches 86% accuracy, which is superior to 83.4% in [37] and comparable to 86.4% in [36]. In the enhanced state, our accuracy reaches 95.8% with ResNet18 architecture, which is comparable to 95.7% in [36] but poorer than 97.1% in [37]. Comparing the training time with a reference study [37] reveals that our proposed framework reaches comparable accuracies with smaller training data and shorter training duration.

7.2. Verification in Local Datasets

We tested the proposed approach (initial state and enhanced state) on local datasets of HS, PD, and SS groups. In the initial state, in terms of accuracy, CNN architectures provide higher performances in the PD dataset compared to HS and SS. This could be associated with the fact that the PD dataset is more balanced than SS and larger than both HS and SS. In addition, majority classes (walking and standing) are better recognized than minority classes (ascent and descent) in the PD dataset. When the sizes of the datasets were artificially increased with data augmentation techniques in the enhanced state, improvements were achieved in all CNN architectures. It is important to highlight that data augmentation has no impact on the balance of a dataset because each class is enhanced at the same rate.
Figure 2 presents the average performances of all CNN architectures from Table 9, Table 11, and Table 13. Sensitivity and specificity values were normalized to 0–100 to present comparable results against accuracy. Comparing initial and enhanced results considering the overall performance of all CNN architectures in the local datasets reveals that the largest improvement in terms of accuracy is observed in HS with 25.6%, followed by SS with 21.4% and PD with 5.8%, as seen in Figure 2. Comparing accuracy, sensitivity, and specificity reveals that data augmentation had the largest improvement in sensitivity at 18.81%, followed by the accuracy at 17.62% and relatively small improvements in specificity at 5.99%. This finding could be associated with the nature of the limited and imbalanced local datasets. In the initial state, the number of true positive (TP) and true negative (TN) in the classification were relatively low. After data augmentation, models experienced better performance in predicting positive classes compared to negative classes. This resulted in a larger increase in TP compared to TN. Consequently, improvements in sensitivity were found to be significantly larger than specificity, Equations (4)–(6).
All four CNN architectures showed a test accuracy exceeding 90% in the enhanced state. ResNet50 outperformed all other architectures in the enhanced state, whereas MobileNet-v2 achieved the best result in the initial state. Although GoogleNet architecture experienced the sharpest enhancement after data augmentation, overall performance in both initial and enhanced states is poorer than its counterparts, as shown in Figure 3. Interpreting these outcomes with the properties of pre-trained CNN architectures (Table 3) could provide useful information regarding the most suitable CNN architecture. Initially, comparing ResNet18 (18 layers) with ResNet50 and MobileNet-v2 (50 and 54 layers) reveals that a higher network layer does not necessarily provide better accuracy because ResNet18 achieved comparable results, aligning with the findings of a previous study that employs the same CNN architectures [60]. This suggests that network size and the number of parameters that a network can learn also have an impact on accuracy. Among the two architectures with the greatest number of deep layers, ResNet50 (larger size and more parameters) provides better classification than MobileNet-v2 (smaller size and fewer parameters) in the enhanced state. Alternatively, MobileNet-v2 (smaller size and fewer parameters) achieves better results than ResNet50 (larger size and more parameters) in the initial state where the dataset is limited and unbalanced. This phenomenon can also be partially observed when two architectures with the lowest number of deep layers are compared. ResNet18 (larger size and more parameters) achieves higher performance than GoogleNet (smaller size and fewer parameters) in the enhanced state. As a result, findings of enhanced state suggest that CNN architectures require approximately 22 deep layers and 7 million parameters (GoogleNet) to classify walking, standing, ascent, and descent activities with more than 90% accuracy. In order to achieve better accuracy, the number of deep layers and/or the number of parameters needs to be increased. The maximum accuracy can be potentially achieved with approximately 50 deep layers and 25.6 million parameters (ResNet50) or approximately 54 deep layers and 3.5 million parameters (MobileNet-v2) because ResNet50 and MobileNet-v2 were found superior in HS, SS, and PD datasets, respectively. On occasions when training time is considered as important as accuracy, ResNet18 architecture could be potentially a more suitable choice because this architecture has fewer deep layers and fewer parameters (fewer computation costs) than ResNet50. However, inconsistencies can occur, as the previous study [51] reports that not all CNN architectures use their parameters with the same level of efficiency.
Our findings revealed that walking and standing are recognized with higher accuracy compared to stair activities, as shown in Figure 4. We also found stair ascent is the activity with the lowest recognition accuracy, aligning with many previous studies that use a single waist device [23,36,66]. Moreover, the figure reveals that data augmentation contributes to better detection of stair ascent and stair descent by 39.1% and 18.0%, respectively. These findings align with a similar study [36] where data augmentation was shown to be effective in recognizing stair activities. Recognition of basic daily life activities in PD and stroke populations with high accuracy has the potential to provide more robust and accurate movement analysis in real life. This framework can be used to accurately classify walking bouts and assist the extraction of clinically important spatiotemporal parameters during walking. Moreover, it can also provide a better picture of the functional capabilities of people with PD and stroke by recognizing stair ambulation activities more accurately.

7.3. Limitation and Future Work

A limitation of the work includes total model training time. Deep learning models are structurally different from traditional machine learning models and involve significantly more training parameters, Table 3. Therefore, deep learning-based CNN models are more complex than traditional machine learning models [67]. This computational complexity can be observed in training times in Table 5 and Table 7. Although the training time reported in this study is shorter than a reference study [37], it still needs improvements.
In this study, the framework was examined within the context of four basic mobility tasks only. In addition, the dataset was created in a semi-controlled environment with a scripted experimental protocol, i.e., all participants walked in the same route while wearing the same device. Future studies will aim to investigate the performances of more complex daily activities in free-living environments (e.g., home). In addition, this framework can be deployed to advanced microcontrollers (Raspberry pi 4- 1.5 GHz) to perform real-time HAR. However, this could still be slower than offline computing as a faster CPU (Core i7-7700HG-2.80 GHz) is used in this study.

8. Conclusions

HAR models typically suffer from low recognition accuracy in neurological populations due to the limitations in data collection. Although highly accurate models have been developed in HAR of healthy people, these models have been found to be limited when recognizing the activities of people with walking impairments. The lack of suitable datasets for those with neurological movement disorders is a major limitation in HAR research. This study proposes a framework to enhance limited HAR datasets, which will have utility in those with a neurological movement disorder. Results showed significant improvements in HAR. The implication of this study can complement future HAR studies where the creation of diverse and balanced data sets may not be feasible. Making maximum use of limited data is important to ensure those with physical impairments may not need to perform difficult dynamic tasks for longer periods to create rich datasets. Therefore, the proposed framework also has the potential to reduce the participant and researcher burden to generate complex and diverse datasets.

Author Contributions

Conceptualization, Y.C., M.F.A., K.S. and A.G. methodology, Y.C., M.F.A. and K.S.; software, Y.C., M.F.A. and K.S.; investigation, Y.C.; data curation, Y.C. and A.G.; writing—original draft preparation, Y.C.; writing—review and editing, M.F.A., K.S., S.S., W.L.W. and A.G.; supervision, K.S., S.S., W.L.W. and A.G.; funding acquisition, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

Yunus Celik receives Ph.D. studentship support from the Turkish Ministry of National Education. Stuart is supported, in part, by the Parkinson’s Foundation (PF-FBS-1898, PF-CRA-2073).

Institutional Review Board Statement

Ethical consent was granted by the Northumbria University Research Ethics Committee (REF: 21603, approved 11 February 2020). All participants gave informed written consent before participating in this study. Testing took place inside and outside of a gait laboratory/lab, Coach Lane Cam-pus, Northumbria University, Newcastle, upon Tyne.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank all those who participated in the study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ramamurthy, S.R.; Roy, N. Recent trends in machine learning for human activity recognition—A survey. Wiley Interdiscip. Rev: Data Min. Knowl. Discov. 2018, 8, e1254. [Google Scholar] [CrossRef]
  2. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  3. Lima, W.S.; Souto, E.; El-Khatib, K.; Jalali, R.; Gama, J. Human activity recognition using inertial sensors in a smartphone: An overview. Sensors 2019, 19, 3213. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Giggins, O.M.; Clay, I.; Walsh, L. Physical activity monitoring in patients with neurological disorders: A review of novel body-worn devices. Digit. Biomark. 2017, 1, 14–42. [Google Scholar] [CrossRef] [PubMed]
  5. Jiang, W.; Yin, Z. Human activity recognition using wearable sensors by deep convolutional neural networks. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 1307–1310. [Google Scholar]
  6. Ahmadi, M.; O’Neil, M.; Fragala-Pinkham, M.; Lennon, N.; Trost, S. Machine learning algorithms for activity recognition in ambulant children and adolescents with cerebral palsy. J. Neuroeng. Rehabil. 2018, 15, 1–9. [Google Scholar] [CrossRef] [Green Version]
  7. Capela, N.; Lemaire, E.; Baddour, N.; Rudolf, M.; Goljar, N.; Burger, H. Evaluation of a smartphone human activity recognition application with able-bodied and stroke participants. J. Neuroeng. Rehabil. 2016, 13, 1–10. [Google Scholar] [CrossRef] [Green Version]
  8. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human activity recognition using inertial, physiological and environmental sensors: A comprehensive survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef]
  9. Hussain, Z.; Sheng, M.; Zhang, W.E. Different approaches for human activity recognition: A survey. arXiv 2019, arXiv:1906.05074. [Google Scholar]
  10. Liu, R.; Ramli, A.A.; Zhang, H.; Datta, E.; Henricson, E.; Liu, X. An Overview of Human Activity Recognition Using Wearable Sensors: Healthcare and Artificial Intelligence. arXiv 2021, arXiv:2103.15990. [Google Scholar]
  11. Rast, F.M.; Labruyère, R. Systematic review on the application of wearable inertial sensors to quantify everyday life motor activity in people with mobility impairments. J. NeuroEngineering Rehabil. 2020, 17, 1–19. [Google Scholar] [CrossRef]
  12. Masum, A.K.M.; Bahadur, E.H.; Shan-A-Alahi, A.; Chowdhury, M.A.U.Z.; Uddin, M.R.; al Noman, A. Human activity recognition using accelerometer, gyroscope and magnetometer sensors: Deep neural network approaches. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–6. [Google Scholar]
  13. Jia, R.; Liu, B. Human daily activity recognition by fusing accelerometer and multi-lead ECG data. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2013), KunMing, China, 5–8 August 2013; pp. 1–4. [Google Scholar]
  14. Liu, J.; Chen, J.; Jiang, H.; Jia, W.; Lin, Q.; Wang, Z. Activity recognition in wearable ECG monitoring aided by accelerometer data. In Proceedings of the 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; pp. 1–4. [Google Scholar]
  15. Cheng, J.; Chen, X.; Shen, M. A framework for daily activity monitoring and fall detection based on surface electromyography and accelerometer signals. IEEE J. Biomed. Health Inform. 2012, 17, 38–45. [Google Scholar] [CrossRef] [PubMed]
  16. Celik, Y.; Stuart, S.; Woo, W.L.; Godfrey, A. Gait analysis in neurological populations: Progression in the use of wearables. Med. Eng. Phys. 2020, 87, 9–29. [Google Scholar] [CrossRef]
  17. Şengül, G.; Ozcelik, E.; Misra, S.; Damaševičius, R.; Maskeliūnas, R. Fusion of smartphone sensor data for classification of daily user activities. Multimed. Tools Appl. 2021, 80, 33527–33546. [Google Scholar] [CrossRef]
  18. Issa, M.E.; Helmi, A.M.; Al-Qaness, M.A.; Dahou, A.; Elaziz, M.A.; Damaševičius, R. Human Activity Recognition Based on Embedded Sensor Data Fusion for the Internet of Healthcare Things. Healthcare 2022, 10, 1084. [Google Scholar] [CrossRef] [PubMed]
  19. Şengül, G.; Karakaya, M.; Misra, S.; Abayomi-Alli, O.O.; Damaševičius, R. Deep learning based fall detection using smartwatches for healthcare applications. Biomed. Signal Process. Control. 2022, 71, 103242. [Google Scholar] [CrossRef]
  20. Trojaniello, D.; Ravaschio, A.; Hausdorff, J.M.; Cereatti, A. Comparative assessment of different methods for the estimation of gait temporal parameters using a single inertial sensor: Application to elderly, post-stroke, Parkinson’s disease and Huntington’s disease subjects. Gait Posture 2015, 42, 310–316. [Google Scholar] [CrossRef] [PubMed]
  21. San, P.P.; Kakar, P.; Li, X.-L.; Krishnaswamy, S.; Yang, J.-B.; Nguyen, M.N. Deep learning for human activity recognition. In Big Data Analytics for Sensor-Network Collected Intelligence; Elsevier: Amsterdam, The Netherlands, 2017; pp. 186–204. [Google Scholar]
  22. Lawal, I.A.; Bano, S. Deep human activity recognition with localisation of wearable sensors. IEEE Access 2020, 8, 155060–155070. [Google Scholar] [CrossRef]
  23. O’Brien, M.K.; Shawen, N.; Mummidisetty, C.K.; Kaur, S.; Bo, X.; Poellabauer, C.; Kording, K.; Jayaraman, A. Activity recognition for persons with stroke using mobile phone technology: Toward improved performance in a home setting. J. Med. Internet Res. 2017, 19, e184. [Google Scholar] [CrossRef] [PubMed]
  24. Albert, M.V.; Toledo, S.; Shapiro, M.; Koerding, K. Using mobile phones for activity recognition in Parkinson’s patients. Front. Neurol. 2012, 3, 158. [Google Scholar] [CrossRef] [Green Version]
  25. Zeng, M.; Nguyen, L.T.; Yu, B.; Mengshoel, O.J.; Zhu, J.; Wu, P.; Zhang, J. Convolutional neural networks for human activity recognition using mobile sensors. In Proceedings of the 6th International Conference on Mobile Computing, Applications and Services, Austin, TX, USA, 6–7 November 2014; pp. 197–205. [Google Scholar]
  26. Huynh, T.; Schiele, B. Analyzing features for activity recognition. In Proceedings of the 2005 Joint Conference on Smart Objects and Ambient Intelligence: Innovative Context-Aware Services: Usages and Technologies; ACM: New York, NY, USA, 2005; pp. 159–163. [Google Scholar]
  27. Capela, N.A.; Lemaire, E.D.; Baddour, N. Feature selection for wearable smartphone-based human activity recognition with able bodied, elderly, and stroke patients. PLoS ONE 2015, 10, e0124414. [Google Scholar] [CrossRef] [Green Version]
  28. Catal, C.; Tufekci, S.; Pirmit, E.; Kocabag, G. On the use of ensemble of classifiers for accelerometer-based activity recognition. Appl. Soft Comput. 2015, 37, 1018–1022. [Google Scholar] [CrossRef]
  29. Shuvo, M.M.H.; Ahmed, N.; Nouduri, K.; Palaniappan, K. A Hybrid Approach for Human Activity Recognition with Support Vector Machine and 1D Convolutional Neural Network. In Proceedings of the 2020 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington DC, DC, USA, 13–15 October 2020; pp. 1–5. [Google Scholar]
  30. Huang, W.; Zhang, L.; Teng, Q.; Song, C.; He, J. The convolutional neural networks training with Channel-Selectivity for human activity recognition based on sensors. IEEE J. Biomed. Health Inform. 2021, 25, 3834–3843. [Google Scholar] [CrossRef] [PubMed]
  31. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  32. Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  33. Eyobu, O.S.; Han, D.S. Feature representation and data augmentation for human activity classification based on wearable IMU sensor data using a deep LSTM neural network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef] [Green Version]
  34. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  35. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
  36. Huang, J.; Lin, S.; Wang, N.; Dai, G.; Xie, Y.; Zhou, J. TSE-CNN: A two-stage end-to-end CNN for human activity recognition. IEEE J. Biomed. Health Inform. 2019, 24, 292–299. [Google Scholar] [CrossRef] [PubMed]
  37. Alawneh, L.; Alsarhan, T.; Al-Zinati, M.; Al-Ayyoub, M.; Jararweh, Y.; Lu, H. Enhancing human activity recognition using deep learning and time series augmented data. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 10565–10580. [Google Scholar] [CrossRef]
  38. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar] [CrossRef]
  39. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  40. Lawal, I.A.; Bano, S. Deep human activity recognition using wearable sensors. In Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments; Association for Computing Machinery: New York, NY, USA, 2019; pp. 45–48. [Google Scholar]
  41. Aslan, M.F.; Sabanci, K.; Durdu, A. A CNN-based novel solution for determining the survival status of heart failure patients with clinical record data: Numeric to image. Biomed. Signal Process. Control. 2021, 68, 102716. [Google Scholar] [CrossRef]
  42. Ortiz, J.L.R. Human Activity Dataset Generation. In Smartphone-Based Human Activity Recognition; Springer Theses; Springer: Cham, Switzerland, 2015. [Google Scholar] [CrossRef]
  43. Zhang, M.; Sawchuk, A.A. USC-HAD: A daily activity dataset for ubiquitous activity recognition using wearable sensors. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing; Association for Computing Machinery: New York, NY, USA, 2012; pp. 1036–1043. [Google Scholar]
  44. Anguita, D.; Ghio, A.; Oneto, L.; Perez, X.P.; Ortiz, J.L.R. A public domain dataset for human activity recognition using smartphones. In Proceedings of the 21th International European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013; 2013; pp. 437–442. [Google Scholar]
  45. Vågeskar, E. Activity Recognition for Stroke Patients. Master’s Thesis, NTNU, Trondheim, Norway, 2017. Available online: http://hdl.handle.net/11250/2468160 (accessed on 23 November 2022).
  46. Albert, M.V.; Azeze, Y.; Courtois, M.; Jayaraman, A. In-lab versus at-home activity recognition in ambulatory subjects with incomplete spinal cord injury. J. Neuroeng. Rehabil. 2017, 14, 1–6. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Jimale, A.O.; Noor, M.H.M. Subject variability in sensor-based activity recognition. J. Ambient. Intell. Humaniz. Comput. 2021, 1–14. [Google Scholar] [CrossRef]
  48. Mikołajczyk, A.; Grochowski, M. Data augmentation for improving deep learning in image classification problem. In Proceedings of the 2018 international interdisciplinary PhD workshop (IIPhDW), Swinoujscie, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar]
  49. Shao, J.; Hu, K.; Wang, C.; Xue, X.; Raj, B. Is normalization indispensable for training deep neural network? Adv. Neural Inf. Process. Syst. 2020, 33, 13434–13444. [Google Scholar]
  50. Banos, O.; Galvez, J.-M.; Damas, M.; Pomares, H.; Rojas, I. Window size impact in human activity recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef] [Green Version]
  51. Bianco, S.; Cadene, R.; Celona, L.; Napoletano, P. Benchmark analysis of representative deep neural network architectures. IEEE Access 2018, 6, 64270–64277. [Google Scholar] [CrossRef]
  52. Mutegeki, R.; Han, D.S. A CNN-LSTM approach to human activity recognition. In Proceedings of the 2020 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan, 19–21 February 2020; pp. 362–366. [Google Scholar]
  53. Tufek, N.; Yalcin, M.; Altintas, M.; Kalaoglu, F.; Li, Y.; Bahadir, S.K. Human action recognition using deep learning methods on limited sensory data. IEEE Sens. J. 2019, 20, 3101–3112. [Google Scholar] [CrossRef] [Green Version]
  54. Canziani, A.; Paszke, A.; Culurciello, E. An analysis of deep neural network models for practical applications. arXiv 2016, arXiv:1605.07678. [Google Scholar]
  55. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  57. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  58. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  59. Ardakani, A.A.; Kanafi, A.R.; Acharya, U.R.; Khadem, N.; Mohammadi, A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks. Comput. Biol. Med. 2020, 121, 103795. [Google Scholar] [CrossRef] [PubMed]
  60. Kim, J.-E.; Nam, N.-E.; Shim, J.-S.; Jung, Y.-H.; Cho, B.-H.; Hwang, J.J. Transfer learning via deep neural networks for implant fixture system classification using periapical radiographs. J. Clin. Med. 2020, 9, 1117. [Google Scholar] [CrossRef] [Green Version]
  61. Galar, M.; Fernandez, A.; Barrenechea, E.; Bustince, H.; Herrera, F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans. Syst. Man Cybern. Part C 2011, 42, 463–484. [Google Scholar] [CrossRef]
  62. López, V.; Fernández, A.; García, S.; Palade, V.; Herrera, F. An insight into classification with imbalanced data: Empirical results and current trends on using data intrinsic characteristics. Inf. Sci. 2013, 250, 113–141. [Google Scholar] [CrossRef]
  63. Yen, C.-T.; Liao, J.-X.; Huang, Y.-K. Human Daily Activity Recognition Performed Using Wearable Inertial Sensors Combined With Deep Learning Algorithms. IEEE Access 2020, 8, 174105–174114. [Google Scholar] [CrossRef]
  64. Li, H.; Trocan, M. Deep learning of smartphone sensor data for personal health assistance. Microelectron. J. 2019, 88, 164–172. [Google Scholar] [CrossRef]
  65. Cho, H.; Yoon, S.M. Divide and conquer-based 1D CNN human activity recognition using test data sharpening. Sensors 2018, 18, 1055. [Google Scholar] [CrossRef] [Green Version]
  66. Lonini, L.; Gupta, A.; Kording, K.; Jayaraman, A. Activity recognition in patients with lower limb impairments: Do we need training data from each patient? In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 3265–3268. [Google Scholar]
  67. Hu, X.; Chu, L.; Pei, J.; Liu, W.; Bian, J. Model complexity of deep learning: A survey. Knowl. Inf. Syst. 2021, 63, 2585–2619. [Google Scholar] [CrossRef]
Figure 1. Data collection protocol and proposed framework: (a) Dataset illustration. The flow of the proposed HAR methodology with data augmentation and CNN architectures: (b) IMU data acquisition, (c) data normalization, (d) numerical to image conversion, (e) resizing, (f) data augmentation, (g) CNN classification.
Figure 1. Data collection protocol and proposed framework: (a) Dataset illustration. The flow of the proposed HAR methodology with data augmentation and CNN architectures: (b) IMU data acquisition, (c) data normalization, (d) numerical to image conversion, (e) resizing, (f) data augmentation, (g) CNN classification.
Sensors 22 09891 g001
Figure 2. Comparison of performance matrices between initial and enhanced states in the local dataset. Sensitivity and specificity values are normalized to 0–100 to provide comparable results with accuracy.
Figure 2. Comparison of performance matrices between initial and enhanced states in the local dataset. Sensitivity and specificity values are normalized to 0–100 to provide comparable results with accuracy.
Sensors 22 09891 g002
Figure 3. Comparison of CNN architectures in terms of accuracy in initial and enhanced status in the local datasets (HS-SS-PD combined).
Figure 3. Comparison of CNN architectures in terms of accuracy in initial and enhanced status in the local datasets (HS-SS-PD combined).
Sensors 22 09891 g003
Figure 4. Recognition accuracy comparison of each activity in the initial result of the local dataset. This graph was derived from the architectures that provide the best performances in enhanced results.
Figure 4. Recognition accuracy comparison of each activity in the initial result of the local dataset. This graph was derived from the architectures that provide the best performances in enhanced results.
Sensors 22 09891 g004
Table 1. Class distributions in local datasets (initial state).
Table 1. Class distributions in local datasets (initial state).
DatasetWalkingAscentDescentStandingTotal
HS50 (25)50 (25)49 (25)50 (25)199 (100)
PD81 (29)64 (23)60 (21)75 (27)280 (100)
SS49 (28)18 (11)31 (18)75 (43)173 (100)
The number of occurrences/images (% class distribution).
Table 2. Number of occurrences after data augmentation (enhanced state) in the local dataset.
Table 2. Number of occurrences after data augmentation (enhanced state) in the local dataset.
DatasetWalkingAscentDescentStandingTotal
HS4504504414501791
PD7295765406752520
SS4411622796751557
The number of occurrences/images (% class distribution).
Table 3. Properties of pre-trained CNN architectures.
Table 3. Properties of pre-trained CNN architectures.
CNN ArchitectureLayer (Depth)Size (Megabyte)Parameters (Millions)Input Image Size
ResNet18184411.7224 × 224
ResNet50509625.6224 × 224
MobileNetv254133.5224 × 224
GoogleNet22277224 × 224
Table 4. Class distributions in benchmarking datasets (initial state).
Table 4. Class distributions in benchmarking datasets (initial state).
Dataset WalkingAscentDescentSittingStandingLayingJoggingTotal
UCI-HAROriginal1226 (17)1073 (15)986 (13)1286 (17)1374 (19)1407 (19)-7352
Utilized 500 (16.6)500 (16.6)500 (16.6)500 (16.6)500 (16.6)500 (16.6)-3000
WISDMOriginal424,400 (38.6)122,869 (11.2)100,427 (9.1)59,939 (5.5)48,395 (4.4)-34,217 (31.2)756,030
Utilized 500 (16.6)500 (16.6)500 (16.6)500 (16.6)500 (16.6)-500 (16.6)3000
Number of occurrences/images (% class distribution).
Table 5. HAR performance metrics in UCI HAR dataset.
Table 5. HAR performance metrics in UCI HAR dataset.
Initial State Enhanced State
DL-CNN
Epochs: 5
Iteration: 3750
Learning rate: 0.001
Batch size: 32
Pre-trained networkAcc. (%)Sens.Spec.F1MCCAcc. (%)Sens.Spec.F1_MCCTraining time (min)
ResNet1893.30.9290.9870.9280.91596.10.9600.9920.9610.95389.26
ResNet5091.80.9140.9840.9110.89797.00.9700.9940.9700.964165.38
MobileNet-v290.70.9030.9820.8990.88396.20.9620.9920.9620.954143.41
GoogleNet81.00.8030.9620.8000.77191.90.9190.9840.9180.90375.55
Acc.: accuracy, Sens.: sensitivity, Spec.: specificity, F1: F1_score.
Table 6. Confusion matrix of UCI HAR- ResNet50 (initial results—left, final results—right).
Table 6. Confusion matrix of UCI HAR- ResNet50 (initial results—left, final results—right).
WalkingAscentDescentSittingStandingLaying WalkingAscentDescentSittingStandingLaying
Walking10600000Walking49401000
Ascent01080000Ascent15473000
Descent01106000Descent10477000
Sitting00089312Sitting0014651217
Standing110146611Standing0001946111
Laying0002476Laying000158467
Table 7. HAR performance metrics in the WISDM dataset.
Table 7. HAR performance metrics in the WISDM dataset.
Initial State Enhanced State
DL-CNN
Epochs: 5
Iteration: 3750
Learning rate: 0.001
Batch size: 32
Pre-trained networkAcc. (%)Sens.Spec.F1MCCAcc. (%)Sens.Spec.F1MCCTraining time
(min)
ResNet1883.50.8320.9670.8280.79995.80.9580.9920.9580.94972.2
ResNet5086.00.8540.9720.8540.82795.40.9530.9910.9530.944163.49
MobileNet-v282.70.8210.9650.8210.78795.40.9530.9910.9530.944129.52
GoogleNet71.50.7190.9430.7180.67889.30.8910.9790.8920.87180.27
Acc.: accuracy, Sens.: sensitivity, Spec.: specificity, F1: F1_score.
Table 8. Confusion matrix of WISDM dataset-ResNet18 (initial results—left, final results—right).
Table 8. Confusion matrix of WISDM dataset-ResNet18 (initial results—left, final results—right).
JoggingWalkingAscentDescentSittingStanding JoggingWalkingAscentDescentSittingStanding
Jogging10020301Jogging48833100
Walking01060200Walking05460500
Ascent34851203Ascent56453824
Descent3487955Descent062045748
Sitting00016032Sitting003146819
Standing01001071Standing004221463
Table 9. HAR performance in local HS dataset.
Table 9. HAR performance in local HS dataset.
Initial State Enhanced State
DL-CNN
Epochs: 5
Iteration: 190
Learning rate: 0.001
Batch size: 32
Pre-trained networkAcc. (%)Sens.Spec.F1MCCAcc. (%)Sens.Spec.F1MCC
ResNet1880.00.8210.9360.8030.75399.70.9970.9990.9970.996
ResNet5082.50.8270.9420.8220.765100.01.0001.0001.0001.000
MobileNet-v285.00.8630.9510.8520.81097.50.9750.9910.9750.967
GoogleNet42.50.3580.7980.3130.22495.30.9530.9840.9520.937
Acc.: accuracy, Sens.: sensitivity, Spec.: specificity, F1: F1_score.
Table 10. Confusion matrix of HS local dataset– ResNet50 (initial results—left, final results—right).
Table 10. Confusion matrix of HS local dataset– ResNet50 (initial results—left, final results—right).
AscentDescentWalkingStanding AscentDescentWalkingStanding
Ascent9210Ascent86000
Descent2500Descent09500
Walking02110Walking00910
Standing0008Standing00087
Table 11. HAR performance metrics in local PD dataset.
Table 11. HAR performance metrics in local PD dataset.
Initial State Enhanced State
DL-CNN
Epochs: 5
Iteration: 190
Learning rate: 0.001
Batch size: 32
Pre-trained networkAcc. (%)Sens.Spec.F1MCCAcc. (%)Sens.Spec.F1MCC
ResNet1894.60.9490.9820.9470.92998.80.9870.9960.9870.983
ResNet5094.60.9400.9810.9450.92899.00.9890.9970.9900.986
MobileNet-v292.90.9360.9760.9310.90899.20.9920.9970.9920.989
GoogleNet89.30.8950.9640.8960.86497.610.9730.9910.9780.975
Acc.: accuracy, Sens.: sensitivity, Spec.: specificity, F1: F1_score.
Table 12. Confusion matrix of PD local dataset– MobileNet-v2 (initial results—left, final results—right).
Table 12. Confusion matrix of PD local dataset– MobileNet-v2 (initial results—left, final results—right).
AscentDescentWalkingStanding AscentDescentWalkingStanding
Ascent15102Ascent121102
Descent11000Descent09900
Walking00130Walking001350
Standing00014Standing001145
Table 13. HAR performance in local SS dataset.
Table 13. HAR performance in local SS dataset.
Initial State Enhanced State
DL-CNN
Epochs: 5
Iteration: 190
Learning rate: 0.001
Batch size: 32
Pre-trained networkAcc. (%)Sens.Spec.F1MCCAcc. (%)Sens.Spec.F1MCC
ResNet1874.30.6900.9170.6430.59196.20.9440.9870.9480.936
ResNet5071.40.6670.9030.6290.55898.10.9680.9930.9730.967
MobileNet-v274.30.6900.9130.6500.59097.40.9600.9920.9600.952
GoogleNet65.70.5000.8740.5630.51679.80.6550.9270.6560.647
Acc.: accuracy, Sens.: sensitivity, Spec.: specificity, F1: F1_score.
Table 14. Confusion matrix of SS local dataset– ResNet50 (initial results—left, final results—right).
Table 14. Confusion matrix of SS local dataset– ResNet50 (initial results—left, final results—right).
AscentDescentWalkingStanding AscentDescentWalkingStanding
Ascent1114Ascent36003
Descent0401Descent15101
Walking00120Walking001200
Standing1208Standing01099
Table 15. Reference studies with benchmarking datasets.
Table 15. Reference studies with benchmarking datasets.
StudyMethodAugmentationAccuracy (%)
UCIWISDM
Alawneh et al. [37]RNNMoving average and the exponential smoothing 97.9–80.0 *97.13–83.4 *
Huang et al. [36]CNNStep detection based novel augmentation technique-
not appropriate for passive activities
-95.7–86.4 *
Yen et al. [63]CNNNA95.99-
Jiang and Yin [5]CNNNA97.59-
Li and Trocan [64]CNNNA95.75-
Cho and Yoon [65]CNNData sharpening97.62-
Proposed frameworkCNNNumerical image conversion + image augmentation97.0–93.3 *95.8–86.0 *
* Represents initial results where available.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Celik, Y.; Aslan, M.F.; Sabanci, K.; Stuart, S.; Woo, W.L.; Godfrey, A. Improving Inertial Sensor-Based Activity Recognition in Neurological Populations. Sensors 2022, 22, 9891. https://doi.org/10.3390/s22249891

AMA Style

Celik Y, Aslan MF, Sabanci K, Stuart S, Woo WL, Godfrey A. Improving Inertial Sensor-Based Activity Recognition in Neurological Populations. Sensors. 2022; 22(24):9891. https://doi.org/10.3390/s22249891

Chicago/Turabian Style

Celik, Yunus, M. Fatih Aslan, Kadir Sabanci, Sam Stuart, Wai Lok Woo, and Alan Godfrey. 2022. "Improving Inertial Sensor-Based Activity Recognition in Neurological Populations" Sensors 22, no. 24: 9891. https://doi.org/10.3390/s22249891

APA Style

Celik, Y., Aslan, M. F., Sabanci, K., Stuart, S., Woo, W. L., & Godfrey, A. (2022). Improving Inertial Sensor-Based Activity Recognition in Neurological Populations. Sensors, 22(24), 9891. https://doi.org/10.3390/s22249891

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop