Next Article in Journal
PSDAAP: Provably Secure Data Authenticated Aggregation Protocols Using Identity-Based Multi-Signature in Marine WSNs
Previous Article in Journal
Surface Profiling and Core Evaluation of Aluminum Honeycomb Sandwich Aircraft Panels Using Multi-Frequency Eddy Current Testing
Previous Article in Special Issue
A Lifetime Maximization Relay Selection Scheme in Wireless Body Area Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Developing Fine-Grained Actigraphies for Rheumatoid Arthritis Patients from a Single Accelerometer Using Machine Learning

1
The Hamlyn Centre, Imperial College London, London SW7 2AZ, UK
2
School of Computer Science and Electronic Engineering, University of Essex, Colchester CO4 3SQ, UK
3
Clinical Innovation & Digital Platforms; Projects, Clinical Platforms & Sciences, GSK, Stevenage SG1 2NY, UK
4
Emerging Platforms, Platform Technology & Science, GSK, Stevenage SG1 2NY, UK
5
Tessella, Altran’s World Class Center for Analytics, Stevenage SG1 3QP, UK
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(9), 2113; https://doi.org/10.3390/s17092113
Submission received: 20 July 2017 / Revised: 24 August 2017 / Accepted: 24 August 2017 / Published: 14 September 2017
(This article belongs to the Special Issue Advances in Body Sensor Networks: Sensors, Systems, and Applications)

Abstract

:
In addition to routine clinical examination, unobtrusive and physical monitoring of Rheumatoid Arthritis (RA) patients provides an important source of information to enable understanding the impact of the disease on quality of life. Besides an increase in sedentary behaviour, pain in RA can negatively impact simple physical activities such as getting out of bed and standing up from a chair. The objective of this work is to develop a method that can generate fine-grained actigraphies to capture the impact of the disease on the daily activities of patients. A processing methodology is presented to automatically tag activity accelerometer data from a cohort of moderate-to-severe RA patients. A study of procesing methods based on machine learning and deep learning is provided. Thirty subjects, 10 RA patients and 20 healthy control subjects, were recruited in the study. A single tri-axial accelerometer was attached to the position of the fifth lumbar vertebra (L5) of each subject with a tag prediction granularity of 3 s. The proposed method is capable of handling unbalanced datasets from tagged data while accounting for long-duration activities such as sitting and lying, as well as short transitions such as sit-to-stand or lying-to-sit. The methodology also includes a novel mechanism for automatically applying a threshold to predictions by their confidence levels, in addition to a logical filter to correct for infeasible sequences of activities. Performance tests showed that the method was able to achieve around 95% accuracy and 81% F-score. The produced actigraphies can be helpful to generate objective RA disease-specific markers of patient mobility in-between clinical site visits.

1. Introduction

Rheumatoid arthritis (RA) is a chronic systemic disease that typically affects adults between the ages of 30 and 60 [1], leading to a life disability in which patients describe increasing pain in several joints over time. This disease has a prevalence of 0.24% that varies between 0.3% and 1% in the developed countries [2,3]. Some studies have also concluded that the risk of mortality of RA patients is approximately 38% higher than the general population, with an increased risk of 55% for women [4]. RA patients are likely to exhibit worse functional status and quality of life than those affected by osteoarthritis [5,6]. A high percentage of RA sufferers (around 50%) are unable to hold a full time job due to the considerable impact of the disease on their daily lives [3]. Research also suggests that the decrease in physical function correlates with a high prevalence of depression and anxiety [7,8]. Thus far, the effect of the disease is assessed based on sparse time-point check-outs and questionnaires during ordinary appointments.
Recent advances in wearable sensing allow continuous patient monitoring in an unobtrusive way, providing a potential source for estimating actigraphies [9,10,11]. The use of additional motion sensors such as gyroscopes, magnetometers and inertial measurement units provides a rich source of kinematic information to obtain effortless actigraphies with high precision and accuracy [12,13]. Clinical studies using several accelerometers have been proposed for studies lasting from one to three days [14]. However, due to battery, life long-term continuous clinical mobility experiments, performed during periods of several weeks up to months, have employed of single inertial sensors with on-board storage memory [15]. Consequently, in the present study, inertial data is also collected from a single tri-axial accelerometer attached to the skin in the position of the fifth lumbar vertebra (L5). The effects of RA can lead to pain in different parts of the body therefore limiting the range of movement, as well as articular sensitivity and function. Hence, certain patients may exhibit distorted mobility patterns, which can make the recognition of a determined activity from a signal segment difficult.
Actigraphies provide a closer look at some relevant factors in RA management such as alteration on patient life-styles and changes on their circadian rhythms [16,17,18,19]. However, these factors cannot be monitored accurately by ordinary appointments that are sparse in time by nature. Manual actigraphies are prone to subjectivity and ambiguity, hence the importance of automated methods for continuous assessment. Fine-grained actigraphies able to predict physical activities from accelerometer signals are key to elucidate novel digital endpoints. In this work a laboratory study is performed on rheumatoid arthritis patients to evaluate the capabilities of the recordings from a single sensor to recognize a set of human physical activities. A set of desirable activity tags are defined, namely sedentary activities (sitting, lying, and standing), acyclic activities (walking), short-timed transitions (sitting-to-standing, standing-to-sitting, lying-to-sitting, sitting-to-lying), and composite activities (lying-to-standing, standing-to-lying). In order to predict these activity tags over the acceleration data from a single sensor, a “hierarchical dichotomy” is proposed and described in detail. Several alternative machine learning methods are used and performance results are compared. Additional functions to filter for unknown activities as well as a logical filter to correct for an unfeasible sequence of transitions are also introduced.

2. Background and Motivation

While questionnaires and self-administered surveys of behaviours, mobility and quality of life have been largely used in clinical research in RA, their susceptibility to error and subjectivity have stimulated the utilisation of wearable sensors [20]. A physical activity assessment with more than a hundred RA patients has been attempted with an accelerometer attached to the waistline of the right hip in line with the right axila [21]. The accelerometer was used to count activity bouts occurring in a ten-minute sequence. Activity bouts can be used as a surrogate measure of the relative amount of activity; however, it does require long periods of recording and the specific activity the patient is performing is undetermined. In [22], a study with a cohort of 98 RA patients used an accelerometer sensor attached to the arm to determine an actigraphy based on the detection of four different levels of activity (sedentary, very light, light and moderate) based on a scale related to energy expenditure. The time that a patient spends performing each activity level is used as a measure of clinical interest. Although activity levels provide an insight into the amount of physical activity that patients perform during the day, a higher level of detail on the actigraphy could provide additional information on the effects that disease manifestation might have on patient behaviours, circadian rhythms and daily activity patterns.
Generating actigraphies from RA patients is challenging as they may exhibit an amalgam of anomalous movement patterns. Issues in mobility have been described in RA patients for different activities such as those arising from a chair or bed as a result of poor gripping and trunk flexion, low range of movement and pain [23,24] or difficulty walking due to weak and painful knees [23], disturbance during sitting or lying due to stiffness and hip pain [25], and difficulty of maintaining a standing upright position [25]. As a difference to healthy subjects who might exhibit a quite similar kinetic and kinematic pattern when performing activities and transitions, disorders in activity patterns can be presented at different levels of intensity with respect to the level of the disease. Most common symptoms are displayed in Figure 1. Moreover, symptoms vary across patients and such heterogenity may accentuate the differences between patterns even further. The key challenge that this paper aims to address is to provide a method that can produce accurate fine-grained actigraphies from RA patients with moderate to severe levels of disease. In this paper, we present a method that have worked effectively in a clinical setting.

3. Materials and Methods

3.1. Experimental Settings

A mobility lab experiment was performed at GSK’s clinical unit in Addenbrookes Hospital (Cambridge, UK), where ten patients with moderate to severe RA and twenty healthy volunteers (HV) were asked to perform a circuit of daily living activities. This consisted of ten tasks denoted as sitting, lying, standing, walking and the transitions sit-to-stand, stand-to-sit, lie-to-sit, sit-to-lie, lie-to-stand and stand-to-lie. The tasks were performed according to a circuit that consisted of the following stations: (1) walking normally at different speeds with turning; (2) standing, sitting on different types of seats then standing; (3) lying, sitting and lying; (4) lying, standing and lying and (5) lying normally for a short period of time. Stations 2 to 4 with transitional tasks were repeated several times. The subjects were asked to perform the activities at their own pace and according to their own habit. They were conducted under normalized settings with conventional furniture such as commonly designed (unarmed) chairs, couches and beds. The sensor was attached with a clinical plaster to the fifth lumbar vertebra, as shown in Figure 2a. The mobility sensor consisted of a wearable three-axis accelerometer logger (Axivity AX3, Axivity Ltd., York, UK) [26]. The accelerometer was configured to sample data at 100 Hz.
Figure 2b shows the typical pattern of inertial signals captured from the accelerometer HV and RA patients where it can be observed that patterns on healthy volunteers are more homogeneous than the ones obtained from the RA patients. These differences in patterns might be produced as a result of a deficient mobility due to the lower range of motions as well as the patient’s trembling and balancing to avoid pain. Additionally, the patient might end up resting in an unconventional position while sitting or lying and might also exhibit a hunched behaviour while standing or walking.

3.2. Inertial Data Pre-Processing

For pre-processing, bias and scale parameters are computed for the calibration of the inertial sensor readings using a Gauss–Newton nonlinear optimization method based on the recordings from nine different static positions [27]. Since we only rely on one accelerometer sensor to develop the actigraphy, besides the sensed g-forces (Acc), three additional measurements of the rotation angles (i.e., roll, pitch, and yaw) are extracted, which we denote with the symbol O. Artifactual high-frequency noise is reduced by applying third-order median filter and the effect of the earth gravitational forces is reduced by low-pass filtering the signal up to 1 Hz and subtracting the resulting filtered signal (gravitational component) from the original inertial data, leaving just body acceleration. An estimation of the body orientation was performed by band-pass filtering the raw acceleration between 0.5 Hz and 15 Hz and then estimating the roll, pitch and yaw orientations by solving the following equations [28]:
O r o l l = arctan A c c x , s i g n ( A c c z ) A c c z + 0.01 A c c x 2 180 / π ,
O p i t c h = arctan A c c x , A c c y 2 + A c c z 2 180 / π ,
O t i l t = arctan A c c z A c c x 2 + A c c Y 2 + A c c z 2 180 / π ,
where s i g n ( A c c z ) has a value of +1 if A c c z is non-negative and −1 if A c c z is negative, in the second term of the roll a 0.01 fraction of the A c c x 2 is added to avoid making this term zero. The orientation computed by Formulae (1)–(3) is not supported by the uses of magnetometers and gyroscopes, hence it can only be considered as an initial estimation. Nevertheless, they can still provide information in addition to the body acceleration, which is helpful for the activity classification. From the body acceleration and estimated body orientation, signal descriptors are computed over a sliding window for a period of three seconds with one-second overlap. The extracted descriptors are applied to both body acceleration and orientation for each of the three signal axes. These descriptors are: mean, variance, root mean square (RMS), number of peaks, difference in magnitude between the maximum peak and the minimum peak, number of through, average distance between peaks and troughs, zero-crossings, coefficient of variation, interquartile range, entropy, cross-correlation between the axes, power of the signal, dominant frequency, power of the signal around the dominant frequency, mean and variance of the signal magnitude area, and mean and variance of the signal magnitude vector. The vector of input sample data and the vectors of activity labels for each sample are denoted in this paper as X = [ x 1 M , x 2 M , x 3 M , , x n M ] and Y = [ y 1 , y 2 , y 3 , , y n ] , where n, M { 1 , , m } is the subset of dimensions of the input vector x , and m is the total number of descriptors. For the sake of simplicity, we will refer to x as the input vector x M in the following sections.

3.3. Subset Selection and Subspace Mapping

The next step is to select a subset of predictors from the ones already computed from Section 3.2 and projecting them into a new subspace that maximizes the separability among the different classes of activities. As the only available information is provided from a single sensor, the data is pre-processed to discover meaningful representations that could ease the recognition. First, a classifier model from the ones described in Section 3.4 is incorporated and a feature selection approach starting from a single descriptor selected at random, appends or removes variables (i.e., descriptors) to and from the dataset that is used to train the classifier according to a predefined criterion based on recognition accuracy. Specifically, we employ a sequential forward floating selection (SFFS) [29], which yields feasible time-complexity while considering a reasonable amount of combinations. The resulting subset of predictors will be used to compute a new optimal set of features used to train a classifier. Two types of mapping learning models are adopted in this work, the first being based on metric learning mapping and the second on deep learning.
Metric learning mapping: This method consists of estimating the hyperparameters of a metric (distance) that collapses the neighborhood of all points from the same class to a single location in the feature space. The neighborhood of a sample is defined by a metric, more specifically a distance function between the points. This procedure is commonly known as metric learning. An optimization step of the hyperparameters for a maximally collapsing metric (MCM), in particular the positive singular matrix of the Mahalanobis distance, is presented in [30]. This method is aimed at minimizing the Kullback–Leibler (KL) divergence between a bi-level distribution of the class labels and a conditional distribution over the points,
m i n A K L p 0 ( j | i ) p ^ ( j | i ) ,
where the distribution p 0 is defined by y i = y j p 0 ( j | i ) = 1 and y i y j p 0 ( j | i ) = 0 , the conditional distribution over the inputs p ^ is defined as e d i j / q i e d i q in which i j and d i j = ( x i x j ) T Σ ( x i x j ) is the Mahalanobis distance, where Σ is the data covariance matrix.
Deep learning features: in this method, a multi-layered neural network is used to learn high-level features from the different hierarchical transformations of the data through the nonlinear processing units, of which each layer is made. These features are optimized via the back-propagation error minimization procedure. Autoencoders can be used to learn a compressed but representative encoding (representation) of the input data. This is achieved by learning an approximating identity function so that the output of the network is similar to the training input. This can be achieved by programming the network to minimize a reconstruction error (e.g., a squared input-output error). The resulting set of reduced features can serve as input for any other machine learning classifier. One disadvantage of auto-encoders with respect to metric learning is that they are not designed to integrate supervised information during learning. A solution is to enable a Discriminative Autoencoders (DAUT) by adding the standard cost function of the reconstruction error, the error of the classification can be defined as:
m i n W 1 , W 2 , H X W 1 H X 2 2 + λ Y W 2 H X 2 2 ,
and solve this equation, where W 1 and W 2 are the reconstruction and classification weights, and H an encoding matrix and λ a parameter that weights the contribution of each error. The cost function in Equation (5) can be solved by any standard optimizer. Mapped samples can be obtained by the dot product of the original input samples by the encoding matrix. The optimal value of λ can be computed analytically by the l-curve method [31] .

3.4. Transitional Activity Tag Classification

For tag classification, in the present work, the machine learning methods presented in the following sections are employed to analyse their performance on recognizing the true label of the activity tag from RA patients with a single accelerometer sensor.

3.4.1. Proposed Novel Meta-Learning Method: Dichotomous Mapped Forest (DMF)

In this section, we propose a novel ensemble meta-learning algorithm denoted as Dichotomous Mapped Forest (DMF). The basis of the algorithm provides a solution to the fact that the number of activity tags can be unbalanced as static activities may last for long periods while transitions occur over periods of a few seconds only. Additionally, the heterogeneity between the different activities across subjects and the limits on the available amount of training data challenges the approach of using just a single multiclass classifier. In any case, only eight classifiers are used to build the dichotomy, which is a much smaller number than the necessary for a (one-vs.-one or one-vs.-rest) multiclass schema of 10 classes.
Specifically, a dichotomous classifier consists of a hierarchical learning architecture, where in every node an independent classifier is trained with tagged activities from two activity subgroups, with each activity subgroup being represented by a different class to the classifier. The structure of the dichotomy for the activity tags is depicted in Figure 3. It is not necessary to build an ensemble of different dichotomies as we can assess the relationships between the tags by their nature: e.g., continuous passive activities (sitting, standing, lie-down), sudden actions or transitions (sitting-to-standing, lying-to-sitting...) and continuous harmonic activities (walking). Then, going down in the hierarchy, we can narrow down the granularity of nodes into lying-to-upright (lying-to-standing, lying-to-sitting), upright-to-upright (sit-to-stand, stand-to-sit). We employ a random forest (RF) classifier due to its low number of critical hyper-parameters and other advantages such as less sensitivity to outliers and over-fitting. In addition, RF does require a relatively small number of hyperparameters. During training, randomization via bagging and limitation of the maximum tree depth are employed to restrict the complexity of the learned model, thus preventing over-fitting.
A top-down logic hierarchy forming a top-down tree was performed to generate a structure of classification nodes with narrowing specialization. This logic structure is displayed in Figure 3. The subset of data input for each classifier is first mapped into a new data space with higher separability between the classes. This is the same technique used for mapped forests with the difference that we use more complex forms of mapping as described in Section 3.3 rather than principal component analysis.
The training complexity for this algorithm is slightly higher than a standard random forest but still linear with respect to the number of samples. In big-o notation, this can be O ( v g R ( m n v l o g n v ) ) , where g is the number of nodes, R is the number of random trees and n v the subset of input samples considered in that node. The processing for each node can be easily parallelized in a standard multicore computer by setting a processing thread for each classifier during training and testing. The processing pipeline of the dichotomy is explained in Figure 4 for training and testing. The part indicated as “logic tree” in Figure 4 corresponds to the structure in Figure 3, which has the purpose of integrating the classifier ensemble according to the established hierarchy of the dichotomy. This integration is necessary as each classifier forming part of the dichotomy is indeed a specialized classifier, which can be either binary or multiclass depending on the structure of the dichotomy, but no more than three classes are ever used in a single node.
Each classifier contained in the dichotomy yields an output and confidence relative to the binary classification problem, which is defined by each node of the dichotomy. The confidence output of each classifier in the dichotomy must be fused (Figure 3) in order to obtain a single one for each prediction. Objectively, a suggested way to merge the confidences of the classifiers in the dichotomy is simply to multiply their values as:
p ( y = c | x ) = k = 1 l u = 1 g k I ( c C k u ) p ( c C k u | x , c C k ) ,
where y is the predicted label, c is a specific discrete label, C is a subset of labels on a node, where l is the number of levels in the hierarchy, and g k is the number of nodes on that level. The indicator function I ( . ) is a Kronecker delta function that returns 0 when c C k , and 1 when c C k . Here, p ( c C k ) is bounded between 0 and 1. A limitation of this method is that the confidences of the classifier are multiplying, considering that each node in the dichotomy is indeed independent but not all nodes have the same relevance. That is to say, miss-classifying a sample higher in the hierarchy could lead to more conflicting predictions as many leaves still lie further down. Therefore, higher nodes should play a higher role in the confidence, as the further down we get into the tree dichotomy structure, the closer we are to a leaf. For this, we have developed a new method to weight the confidence that is based on the following equation:
p ( y = c | x ) = k = 1 l u = 1 g k I ( c C k u ) ϕ k p ( c C k u | x , c C k ) ,
ϕ k = 2 ( l k + 1 ) l 2 + l .
As a difference with Equation (6), in Equation (7), a weight denoted as ϕ is introduced. These weights become smaller from the upper level node k = 1 to the lower level node k = l in the hierarchy and are computed by decreasing function (8).

3.4.2. Machine Learning Classifiers Used for Comparison

Several popular algorithms used in activity recognition with accelerometers are employed in this work for evaluation purposes. In this section, we provide a description about their learning mechanisms as well as advantages and disadvantages.
Radial Basis Support Vector Machines (RB-SVM). RB-SVMs are popular for training a classifier and work relatively well with small sets containing few outliers [32]. As its learning mechanism is based on a convex optimization problem, it is resilient to local optima. Its training complexity, O ( n 2 m ) , is relatively affordable for standard computers, and the number of parameters to predefined are the regularization coefficient and the kernel parameters. The regularization parameter in the SVM controls the complexity of the margin, hence making the model resistant to over-fitting. The bandwidth and regularization parameters are tuned from the training data via cross-validation and Bayesian optimization as provided in several toolboxes [33,34]. Although multiclass versions of SVMs have been developed, better performance has been achieved when binary SVM classifiers are stacked in one-vs.-one or one-vs.-all fashions. However, this latter methodology requires to define appropriate methods for the fusion of the inference from these classifiers. Another limitation of RB-SVMs is that the selection of the regularization and kernel parameters can highly impact the performance of the classifier and its kernel-based models can be quite sensitive to over-fitting [35]. SVMs are essentially a binary classifier in theory. Multiclass SVMs have been claimed not as efficient as their “one-vs.-all” or “one-vs.-one” multiclass schemes [36]. For this application, we need a minimum of 45 SVM binary classifiers.
Random Forests (RF) are ensemble algorithms of decision trees that can be applied to any multiclass problem with its standard design. Their learning is based on the concepts of information gain and measuring statistical dispersion rather than kernel-learning as in SVM. They can be applied to many problems given that their most critical parameter for performance is the number of trees, whose value can be easily determined by greedy search or calculating out-of-bag error rates [37]. When the latter is large, it then is necessary to provide the model with more data than with other methods such as SVMs. In terms of their computational complexity, however, it is higher than the one of SVMs but still tractable for standard computers with a few CPU cores, as follows: O ( R ( m n l o g n ) ) , where R is the number of trees. They are able to perform implicit feature selection during learning. Although they are not resilient to over-fitting, adding more trees to the forest actually decreases this issue [38]. Over-fitting is controlled by bagging, restricting max three-depth, and the number of potential splits. The effects of applying bagging makes the algorithm more robust with respect to outliers [39]. When working with unbalanced datasets, while SVMs can be tuned to work with penalty weights to penalize the classification of the majority class, and a solution in RFs is a pass for careful management of the data sample partition for each random tree.
Convolutional Deep Belief Networks (CDBN) are a neural network type of classifier consisting of stacked unsupervised networks based on restricted Boltzmann machines (RBM) with a Softmax classifier as the output layer. RBM are generative machines whose main finality is to build a generative model of its inputs. It is composed of two layers, namely the visible layer where the input is first introduced and the hidden layer. Each RBM performs several forwards and backwards passes between the visible and hidden layers so as to adjust the connection weights between the input, hidden and output layers in order to be able to generate an output that is very close to the original input data. The difference between the reconstruction and original input is measured by means of KL divergence, which is typically minimized by applying the contrastive divergence algorithms. When RBMs are stacked together to form a DBN, the hidden layer of one RBM is connected to the visible layer of the one above. As a difference from convolutional networks, DBNs present the advantage that they can yield a reasonable performance with just a small amount of labeled data and when a spatial linkage between the input feature is not existent (for instance, a single row feature vector). The partition of the learning into separate RBMs makes the learning process robust against the vanishing gradient issue. In the present work, for CDBNs, the energy function of the RBM has been modified to add a convolutional operator, which ensures the weights are shared among groups of hidden units [40]. This permits standard DBNs to scale to large dimensional inputs, including spatial information, as well as becoming invariant to input transformations by employing a pooling operator. The risk of over-fitting is solved by forcing activation of the hidden units to become sparse, therefore eluding an over-complete model [41]. The Convolutional Neural Networks (CNNs) architecture of learning is optimized to learn from 2D representations of the data. Although signal data from an accelerometer is one-dimensional, some works have proposed using a time frequency representation as input [42].  This is actually performed by computing a spectrogram applying a short-time Fourier transform over the three accelerometer forces. This spectrogram represents the power-spectral density component and some studies have reported that inherent information about the movement is encoded in this representation [42]. This transformation of the data aims at converting one-dimensional signal data into a two-dimensional data based on its time-frequency representation. This is actually performed by computing a spectrogram applying a short-time Fourier transform over the three accelerometer forces. This spectrogram represents the power-spectral density component and some studies have reported that inherent information about the movement is encoded in this representation [42]. On the one hand, a limitation of this feature with respect to the previous two is that it does not contain class information able to maximize the separability—it is simply a time-frequency representation of the signal. On the other hand, they enable using the data with deep learning architectures such as CNNs.
The disadvantages are that the learning process requires an iterative processing and samplers that may involve a large series of operations; hence, it is recommended to transfer operations such as multiplication between weights and visible or hidden units to a graphical processing units (GPU) or specialized co-processors. Additionally, another disadvantage of these models is that they require pre-defining a large number of hyper-parameters: number of hidden units, weight decay, weight initialization, activation sparsity, learning rate, length of the filter, length of feature maps for both the visual and hidden layers, and so on. Thus, performance is highly dependent on computationally expensive multi-objective optimizers, which may add a significant amount of processing time and machine resources. As standard, hyper-parameters are set to the values suggested in previous works [40].
Continuous Hidden Markov Models (HMMs) are effective models that have long been used in the field of physical activity recognition [43,44]. They are machine learning algorithms for sequential data, particularly sequences of discrete samples. The idea behind HMMs is to learn the likelihood of a sequence of observations given a sequence of hidden states. To this aim, an emission and a transition probabilities between a set of hidden states structured in a Markov chain model are learned from the labeled data using maximum likelihood. To estimate the probability that a sequence of observations can be produced by a particular HMM, backwards and forwards inference are performed by means of a dynamic programming algorithm [45]. When working with signals such as G-forces from an accelerometer, it is necessary to first discretize them into a code-book of discrete samples. These can be obtained by quantization or defining a set of clusters (for instance Gaussian mixtures) in the data and an observation as the label of the cluster to which a sample belongs. A metric of optimal cluster dispersion can be used to assess the optimal number of clusters. Clusters’ co-variances are adjusted for positive definiteness and the maximum number of possible hidden states is set to a limit, hence avoiding highly complex models and evading over-fitting. The advantage of the cluster method is that it can handle samples with multiple dimensions. HMMs are not very expensive to train in modern computers and their parametrization is not excessive as their only critical parameter is the number of hidden states, which can be estimated by any simple greedy or heuristic search. However, as HMMs are inherently designed to be trained with discrete inputs, the performance is highly affected by how well the discrete samples can represent the nature of the original continuous signal.

3.5. A State Machine as a Logic Filter

Most clinical studies present a set of activity tags of interest rather than a universal set. These activity tags are related since a set of activities can occur sequentially in time in a determined order of the logic filter; for instance, one could expect a sit-to-stand prediction after a sitting position but not a stand-to-lying. The number of previous predictions checked to find one of these allowed activities is regulated by the order of the logic filter. In addition, in the case when this is incorrectly predicted, our hypothesis is that a well calibrated classifier should output a low confidence for this prediction. Then, we can check whether this sequence is possible in our activity state machine every time our classifier outputs a prediction with a confidence level lower than a determined threshold. The state machine of possible activity sequences is presented in Figure 5. Most of the transitions between the states of static activities account for a transitional activity in between them, although some exception is made for walking, for which direct transitioning to a static activity state might be permitted, for instance passing from walking to simply standing. Most classifiers yield a confidence value, sometimes in the form of firing strength, probability or distance to a separating plane. In the case of the DMF, this can be done by merging the confidence of all classifiers that form part of the dichotomy, using Equation (7). Figuring out a lower bound, we denote a δ , on the confidence level can be performed by leaving aside a small set of labeled data from the training set, obtaining their predicted values and confidences and finding the optimal operating point among the points in a receiver operating characteristic (ROC) curve. An automatic criterion is to pick out the point corresponding to the minimum Euclidean distance between the point (0,1) and the points in the ROC,
δ = m i n ( d i s t ( i ) ) ; d i s t ( i ) = R O C x 2 + ( 1 R O C y ) 2 ,
where R O C x and R O C y are the x and y values of the ROC curve.
The threshold on the confidence sets a trigger to activate the logical filter of the predictions by looking at the activity automata (Figure 5). This can be effectively achieved by checking the constraints between impossible transitions. If the predicted confidence bound is not surpassed, then the following condition is evaluated:
β : = S t 1 S t p ( y t 1 = c | x ) < δ i o r d e r S t ( i + 1 ) · Θ · S t 1 S t 1 · Θ · S t ,
β = t r u e y t 1 : = y t 1 , β = f a l s e y t 1 : = y t 2 ,
where S t is the state of the current time. The state is a single-entry vector where all the elements are zero apart from one which corresponds to the class index of y t . The symbol Θ is an adjacency matrix of logical true or false values, indicating whether the transition conditions of the finite state-machine are possible. The factor order allows for a level of flexibility in searching for the valid transition between the previous transitions. If a valid sequence is found between the previous activities until t o r d e r + 1 , then the same label is kept despite not surpassing the lower bound on the confidence.
Algorithms 1 and 2 show the sequence of processing steps for training and testing the DMF. Symbols X and Y are the input and tag label data, respectively, and S e l n o d e f e a t are the selected features for each node in the dichotomy after applying SFFS. C n o d e is the list of confidence of the prediction for each node and C n o d e n o r m represents the ones normalized through the dichotomy.
Algorithm 1 Dichotomy mapped forest—Training
Input:  X,Y
Output:  trained model, map & S e l n o d e f e a t
1:
for Each node in logic three according to the dichotomy in Figure 3 do
2:
    Split X into X A n o d e and X B n o d e acording to the node label grouping in each node in Figure 3
3:
    [ X A n o d e s e l , X B n o d e s e l , S e l n o d e f e a t ] ← call SFFS (input=[ X A n o d e , X B n o d e ], labels = [ A n o d e , B n o d e ])
4:
    if mapping == DL then
5:
        [ m a p ( n o d e ) , m a p p e d A , m a p p e d B ] ← call DAUT (input = [ X A n o d e s e l , X B n o d e s e l ], labels = [ A n o d e , B n o d e ])
6:
    else if mapping == Metric then
7:
        [ m a p ( n o d e ) , m a p p e d A , m a p p e d B ] ← call MCM (input = [ X A n o d e s e l , X B n o d e s e l ], labels = [ A n o d e , B n o d e ])
8:
    end if
9:
     m o d e l ( n o d e ) ← call train RF (input = [ m a p p e d A , m a p p e d B ], labels = [ A n o d e , B n o d e ])
10:
end for
Algorithm 2 Dichotomy mapped forest—Test
Input:  Z (test set), m o d e l , S e l n o d e s f e a t , m a p
Output:   Y p r e
1:
for Each node in logic three according to the dichotomy in Figure 3 do
2:
     Z n o d e s e l ← select S e l n o d e f e a t features in Z n o d e
3:
     m a p p e d Z ← map input Z n o d e s e l with m a p ( n o d e )
4:
    [ Y n o d e p r e , C n o d e s ] ← test m a p p e d Z in m o d e l ( n o d e )
5:
end for
6:
[ Y p r e , C n o d e s n o r m ] ← apply Dichotomy (predictions = Y n o d e s p r e , confidence = C n o d e s ) as Figure 3 & Equation (7)
7:
b o u n d n o d e s ← get confidence bound using C n o d e s n o r m as Equation (9)
8:
if C n o d e n o r m < b o u n d n o d e then
9:
     Y p r e ← check logic filter as in (11)
10:
end if

3.6. Training Settings: Mixing Patients and Healthy Volunteers

Clinical studies frequently account for a number of patients and control subjects or volunteers. Patient data concerns a very restricted population and, consequently, a common question when working with patient data is to know whether the dataset can be extended with data from healthy volunteers. Figure 2b highlights the difference in signal patterns between patients and healthy volunteers. In this regard, we aim at investigating the hypothesis that the classification performance of the algorithm can be affected by implementing the training in different ways. Figure 6 shows a graphical representation of the three different training settings.

4. Results and Discussion

This section examines the performance of the methods described in Section 3. The machine learning algorithms proposed in Section 3.3 are trained using a dataset formed of healthy subjects and patients performing the same activities. The results shown in Section 4.1 and Section 4.2 apply the training setting A as depicted in Figure 6. Inertial data is pre-processed to match the desired input method used for the particular classification method. For instance, RF-based methods can be trained with dimensionality reduced inputs from either deep learning (DL) or metric learning (Metric), and the same is applicable to the SVM-based classifiers. CDBN methods are trained using a 2D spectrogram representation of each signal axis as suggested in [46]. For C-HMM, the input features are first quantized to form a transcript (codebook) proposed methods described in [47].

4.1. Activity Classifiers Comparison

This section provides the results of classifying activity patient data using leave-one-subject-out cross-validation to test for the subject independence of the learned model. Figure 7 shows the confusion matrices for each method, which are ordered by decreasing performance from left to right columns and from up to down rows. A major number of segments classified with true activity tags are found in the main diagonal for the suggested best method DMF. Accuracy, recall and specificity are above 90% for the DMF-based methods (Table 1), using inputs spanned by a subspace mapping generated by either a DAUT or metric leaning. However, the F-scores are rather moderate, namely approximately 74% and 73% for DMF-DL and DMF-Metric, respectively. Furthermore, activities that involve transferring from a lying to a stand position are amongst the hardest for the RA patients because they entail a range of sub-movements including pelvis and lumbar torsion as well as joint exertion, which are triggers for pain. The level of persistent muscular pain and stiffness caused by the disease has an important impact on the way patients perform these activities. Additionally, the tagging of the latter is complex as the patient may perform the movements task as a sequence of smooth movement to reduce the feelings of muscular pain, so that rather than just being a single lying-to-stand, it can be seen as a composite of subtle supportive actions until the patient finally achieves a comfortable standing position. Given that the accelerometer cannot provide a measure of angular displacement, if the patient performs repeated balancing and little jumps when transitioning towards the upright position, the signal might exhibit some repeated burst of acceleration and predictions could be likely inaccurate. Likewise, if the patient is performing long gaps of static motion while transitioning, this could be predicted as a static activity.
Table 1 shows the numerical results from the performance test, the best two approaches being DMF-DL and DMF-Metric. Upon statistical test using non-parametric Kruskall–Wallis and Dunn’s test multiple comparison, the performance of DMF-DL is significantly different from SVM-DL and any other classifier down-performing it with (z = 49.80, p < 0.01).Likewise, DMF-Metric is significantly different from SVM-Metric and any other classifier down-performing it with (z = 81.18, p < 0.01). Alternative statistical tests using analysis of variance and Tukey-kramer test show that DMF-DL and DMF-Metric are statistically different from CDBN and any other classifier down-performing it with (q = 0.19, p < 0.01) and (q = 0.19, p < 0.01), respectively. The performance of DMF-Metric and DMF-DL are not significantly different.

4.2. Applying the Logic Filter

A requirement for an actigraphy is that the sequence of activities must be logical for the activity tags defined. The method described in Section 3.5 is applied for the best two methods from Section 4.1 (DMF-DL and DMF-metric). DMF-DL and DMF-metric are the only methods that exhibit both high accuracy and a moderately good F-score as shown in Figure 7 and Table 1. This means that, although most of the ground truth tags are correctly predicted as true positives and true negatives, several other predictions are false positives. As discussed in Section 4.1, this could be produced by the imprecision of the tags; however, a way of ensuring that our tags follow a valid sequence is to activate the logic filter for the prediction as presented in Section 3.5.
The logic filter considerably improves the values of the F-score for the recognition performance, as it can be seen in the lower part of Figure 8. This is important to obtain a well-balanced classifier. Indeed, the logic filter is able to delete those false positives that do not exhibit a logical sequence. The order providing the best results is two, and this means that checking two previous predicted activity tags is enough to find an allowed sequence. While the following orders also have a good impact on the overall classification, the performance either remains the same or decreases (Table 2). Looking for a coherent sequence in many past tags might indeed result in inconsistent logical assumptions.
In Figure 9, a comparison between the performance obtained without applying the logic filter and the one obtained after applying it is shown. As it can be seen, the activity tags that benefit the most are the lying-to-upright transitions as the filter exploits the high sensitivity (above 85%) of recognizing the preceding static activities such as stand, sitting, or lying-down. Overall recognition rates for other transitions such as sit-to-stand and stand-to-sit also improve after applying this correction. Recognition rates are high (over 80%) for all considered activity tags. The differences between applying the mapping powered by the DAUT (DL) or MCM (Metric) are not significant.

4.3. Results with Different Training Settings

In this section, the best two classifiers, i.e., DMF-DL and DMF-Metric, are trained with different subsets of the data. In particular, we use the settings described in Section 3.6. Previous results from Section 4.1 and Section 4.2 were performed using training settings A as described in Figure 6, which consists in using all the available data to train the classifiers. Patient data is difficult to collect as clinical trials require a larger amount of work in terms of preparation as well as the need for an assisting medical team during the study. Hence, one might be interested in looking at the capabilities of the classifier with different training approaches such as using just patient data for training or just for the feature mapping.
Table 3 shows how the training is performed applying settings B and C from Figure 6. It can be seen that, using the patient data only, the performance of the DMF-DL method is diminished by 17% in accuracy and 40% in F-score while the performance of DMF-Metric decreases by 13% in accuracy and 31% in F-score. When applying the patient data just for the mapping, the performance results also lowers the classifier performance, although less than when using patient data only for training. This deterioration in the results reaches 12% in accuracy and 30% in F-score for DMF-DL, compared to 9% and 25% for DMF-Metric. If the logic filter is applied for the predictions, then the negative impact on the results is less dramatic. Specifically, using just patient data for training worsens the results by 8% in accuracy and 26% in F-score for DMF-DL, while the performance decreases by 5% and 16% for DMF-Metric. When the patient data is used for mapping only, then the lowest deterioration is achieved with 5% decrease in accuracy, but 16% in F-score for DMF-DL, while, for DMF-Metric, the deterioration is just 2% in accuracy, but 15% in F-score.
The conclusion from this test is that within the limits investigated, the more (either patient or healthy) subject data is used to jointly train the mapping and the classifier, the more accurate the predictions are. Alternative settings such as B and C are not recommended because they penalize the classification performance for the best approach, particularly the F-score estimates. DMF-Metric exhibits a slightly higher average performance than DMF-DL for this test, and the performance differences between DMF-DL and DMF-Metric are not significant either.

5. Conclusions

In this work, data from a single accelerometer was used to classify static, acyclic and transitional activities. Several methods were adopted as a comparison to analyse the performance behaviour from different machine learning classifiers. Using discriminatory mapping as the ones described in Section 3.3, i.e., which uses the labels from the training data to obtain an optimal subspace to discriminate between activities, proved to be a solution to cope with the signal variations between subjects, particularly the ones arising from the congruity of the inertial data from healthy individuals and the diversity among the RA patients. Building a single multi-classifier from a highly diverse sequence of activities that can also have particular divergences across subjects is a challenging task. We have found positive the utility of an ensemble of classifiers as a form of dichotomy of nodes consisting of a single classification task and described in detail. The division of activities follows a quasi-logical structure, permitting most general conceptual division of activities such as active versus passive activities to be discriminated in advance of specific ones such as sitting or stand-to-sit and the like. It is worth noting that standard ensemble learning depends on the randomization of the input data, which, in this case, may not be as advantageous, as an arbitrary arraignment would ignore the intrinsic categories among activities.
As another relevant finding, when a substantial number of activity tags (ten in the present study) is requested to be classified, with just one sensor and an uncertain validity of the tags, then adding a prediction corrector in the form of a logic filter on top of the less confident predictions helps to achieve more accurate results, as the desired activity tags are always expected to follow a rational and valid sequence.
The best classifications were achieved with the DMF methods, consisting of the dichotomy and mapping plus the classifier. There were no statistical differences between the performance of both types of mapping, the DL mapping being slightly above the one of MCM when all the data in the study is provided to the classifier. It is worth mentioning that the former requires more computational resources than the later so that a powerful GPU computing, dedicated co-processors or a computer cluster are recommended. In the present study, we experienced that the gain of using DL mapping with respect to MCM mapping is not sufficient to justify the higher computational cost for this specific application, as it can be seen from the results in Section 4.2. The best results were approximately 95% accuracy and 81% F-score. These results are also dependent on the evaluation method used. Leave-one-subject-out cross-validation was the chosen method in order to avoid randomly partitioned k-folds, which may end-up including singular activities from the same subject in both the training and test set, hence jeopardizing the subject independence of the learned model.
Although we have intended to simulate a very realistic environment during the experiment, including different types of chairs and walking speeds, to allow natural motion of the subject etc., we might expect that during free-living there could be a set of different scenarios that we cannot replicate during our lab experiments. Nevertheless, the proposed method is designed with robustness in mind to avoid nonsensical predictions by means of the logic filter as well as avoiding signal instabilities by virtue of the proposed pre-processing and subsequent data mappings. Therefore, we do have the certainty that although the accuracy of “free-living” activity predictions might not be identical to the ones reported in this study, they will not degrade significantly. The set of activities provided during training will be recognized to the same extent. The variability of the position of the sensor can be an additional limiting factor during free-living. In previous clinical experiments involving inertial sensors, the lower back position at the L3 to L5 vertebrae has been a popular location due to its closeness to the center of mass [48,49], and acceptable for long-term studies at-home use [50,51]. Nevertheless, proper fitting of the sensor by clinical professionals during the first visit would minimize the impact of possible sensor detachment or displacement.
To conclude, this study showed that it is possible to recognize activities from individuals with disabilities affecting their mobility and postural condition such as RA, in short periods of three seconds, from a single accelerometer attached to the L5 with compelling accuracy. The methods presented in this work can be helpful for the elaboration of clinical studies where wearable data can be used to build actigraphies from patient data.
Last but not least, we would like to highlight for future research the potential of sensors providing novel quantitative physiological bio-markers in order to monitor the health status of the musculoskeletal system of the patients. Along these lines, a novel and promising wearable sensor with contact microphones to pickup acoustical emission from joint sounds has recently been proposed [52], and there are also advances for the in-vivo monitoring of stresses and strains of musculoskeletal tissues by implantable sensor technologies [53,54]. These new sources of local health status in combination with accurate daily-living actigraphies may potentially yield a rich source of information enabling a personalised disease management and optimal care.

Acknowledgments

The authors thank our GSK colleagues from the Clinical Unit at Cambridge Danielle Gerlag, Phil Lawrence and Graham Bentlely, as well as Russell Williamson from GSK’s Immuno Inflamation Therapy Area and other colleagues from the clinical team. The authors also thank all the patients and volunteers who participated in this study. We appreciate the support of the Amazon Web Services (AWS) Cloud Credits for Research program for providing the computational resources to conduct this study.

Author Contributions

L.G.C., J.M., T.K. conceived and designed the experiments; L.G.C., J.M., T.K. performed the experiments; J.A.P., A.V.D. and A.P. developed the methods and analyzed the data; G.Z.Y. supervised and appraised the work and J.A.P. and L.G.C. wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
C-HMMContinuous Hidden Markov Models
CDBNConvolutional Deep Belief Networks
DCNNDeep Convolutional Neural Networks
DLDeep Learning
DMFDichotomous Mapped Forest
DMF-DLDMF with Deep Learning mapping
DMF-MetricDMF with Metric learning mapping
DAUTDiscriminatory Autoencoder
GPUGraphical Processing Unit
HVHealthy Volunteers
KLKullback-Leibler
L5L5 lumbar vertebra
MCMMaximally collapsing metric
MEMSMicroelectromechanical systems
MetricMetric Learning
RBFRadial Basis Function
RFRandom Forest
ROCReceiver Operating Characteristic
RBMRestricted Boltzmann Machine
RARheumatoid Arthristis
RMSRoot Mean Squared
SFFSSequential Fowards Floating Selection
SVMSupport Vector Machine

References

  1. Yazici, Y.; Paget, S.A.; Sibley, J. Elderly-onset rheumatoid arthritis. Rheum. Dis. Clin. N. Am. 2000, 26, 517–526. [Google Scholar] [CrossRef]
  2. Cross, M.; Smith, E.; Hoy, D.; Carmona, L.; Wolfe, F.; Vos, T.; Williams, B.; Gabriel, S.; Lassere, M.; Johns, N. The global burden of rheumatoid arthritis: Estimates from the Global Burden of Disease 2010 study. Ann. Rheum. Dis. 2014, 73, 968–974. [Google Scholar] [CrossRef] [PubMed]
  3. Woolf, A.D.; Pfleger, B. Burden of major musculoskeletal conditions. Bull. World Health Organ. 2003, 81, 646–656. [Google Scholar] [PubMed]
  4. Gabriel, S.E.; Crowson, C.S.; O’Fallon, W.M. Mortality in rheumatoid arthritis: Have we made an impact in 4 decades? J. Rheumatol. 1999, 26, 2529–2533. [Google Scholar] [PubMed]
  5. Jakobsson, U.; Hallberg, I.R. Pain and quality of life among older people with rheumatoid arthritis and/or osteoarthritis: A literature review. J. Clin. Nurs. 2002, 11, 430–443. [Google Scholar] [CrossRef] [PubMed]
  6. Slatkowsky-Christensen, B.; Mowinckel, P.; Loge, J.H.; Kvien, T.K. Health-related quality of life in women with symptomatic hand osteoarthritis: A comparison with rheumatoid arthritis patients, healthy controls, and normative data. Arthritis Rheum. 2007, 57, 1404–1409. [Google Scholar] [CrossRef] [PubMed]
  7. Dickens, C.; McGowan, L.; Clark-Carter, D.; Creed, F. Depression in rheumatoid arthritis: A systematic review of the literature with meta-analysis. Psychosom. Med. 2002, 64, 52–60. [Google Scholar] [CrossRef] [PubMed]
  8. Söderlin, M.K.; Hakala, M.; Nieminen, P. Anxiety and depression in a community-based rheumatoid arthritis population. Scand. J. Rheumatol. 2000, 29, 177–183. [Google Scholar] [PubMed]
  9. Iyengar, S.; Bonda, F.T.; Gravina, R.; Guerrieri, A.; Fortino, G.; Sangiovanni-Vincentelli, A. A framework for creating healthcare monitoring applications using wireless body sensor networks. In Proceedings of the ICST 3rd International Conference on Body Area Networks, Tempe, AZ, USA, 13–17 March 2008; pp. 8–9. [Google Scholar]
  10. Fortino, G.; Giannantonio, R.; Gravina, R.; Kuryloski, P.; Jafari, R. Enabling Effective Programming and Flexible Management of Efficient Body Sensor Network Applications. IEEE Trans. Hum. Mach. Syst. 2013, 43, 115–133. [Google Scholar] [CrossRef]
  11. Galzarano, S.; Giannantonio, R.; Liotta, A.; Fortino, G. A Task-Oriented Framework for Networked Wearable Computing. IEEE Trans. Autom. Sci. Eng. 2016, 13, 621–638. [Google Scholar] [CrossRef]
  12. Gravina, R.; Alinia, P.; Ghasemzadeh, H.; Fortino, G. Multi-sensor fusion in body sensor networks: State-of-the-art and research challenges. Inf. Fusion 2017, 35, 68–80. [Google Scholar] [CrossRef]
  13. Fortino, G.; Galzarano, S.; Gravina, R.; Li, W. A framework for collaborative computing and multi-sensor data fusion in body sensor networks. Inf. Fusion 2015, 22, 50–70. [Google Scholar] [CrossRef]
  14. Bussmann, J.B.J.; Martens, W.L.J.; Tulen, J.H.M.; Schasfoort, F.C.; van den Berg-Emons, H.J.G.; Stam, H.J. Measuring daily behavior using ambulatory accelerometry: The Activity Monitor. Behav. Res. Methods Instrum. Comput. 2001, 33, 349–356. [Google Scholar] [CrossRef] [PubMed]
  15. Doherty, A.; Jackson, D.; Hammerla, N.; Plötz, T.; Olivier, P.; Granat, M.H.; White, T.; van Hees, V.T.; Trenell, M.I.; Owen, C.G. Large Scale Population Assessment of Physical Activity Using Wrist Worn Accelerometers: The UK Biobank Study. PLoS ONE 2017, 12, 1–14. [Google Scholar] [CrossRef] [PubMed]
  16. Stone, A.A.; Broderick, J.E.; Porter, L.S.; Kaell, A.T. The experience of rheumatoid arthritis pain and fatigue: Examining momentary reports and correlates over one week. Arthritis Care Res. 1997, 10, 185–193. [Google Scholar] [CrossRef] [PubMed]
  17. Cutolo, M.; Villaggio, B.; Otsa, K.; Aakre, O.; Sulli, A.; Seriolo, B. Altered circadian rhythms in rheumatoid arthritis patients play a role in the disease’s symptoms. Autoimmun. Rev. 2005, 4, 497–502. [Google Scholar] [CrossRef] [PubMed]
  18. Katz, P.P. The impact of rheumatoid arthritis on life activities. Arthritis Care Res. 1995, 8, 272–278. [Google Scholar] [CrossRef] [PubMed]
  19. Luyster, F.S.; Chasens, E.R.; Wasko, M.C.M.; Dunbar-Jacob, J. Sleep quality and functional disability in patients with rheumatoid arthritis. J. Clin. Sleep Med. 2011, 7, 49–55. [Google Scholar] [PubMed]
  20. Darwish, A.; Hassanien, A.E. Wearable and Implantable Wireless Sensor Network Solutions for Healthcare Monitoring. Sensors 2011, 11, 5561–5595. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Semanik, P.; Song, J.; Chang, R.W.; Manheim, L.; Ainsworth, B.; Dunlop, D. Assessing physical activity in persons with rheumatoid arthritis using accelerometry. Med. Sci. Sports Exerc. 2010, 42, 1493–1501. [Google Scholar] [CrossRef] [PubMed]
  22. Khoja, S.S.; Almeida, G.J.; Chester Wasko, M.; Terhorst, L.; Piva, S.R. Association of Light-Intensity Physical Activity With Lower Cardiovascular Disease Risk Burden in Rheumatoid Arthritis. Arthritis Care Res. 2016, 68, 424–431. [Google Scholar] [CrossRef] [PubMed]
  23. Häkkinen, A.; Kautiainen, H.; Hannonen, P.; Ylinen, J.; Mäkinen, H.; Sokka, T. Muscle strength, pain, and disease activity explain individual subdimensions of the Health Assessment Questionnaire disability index, especially in women with rheumatoid arthritis. Ann. Rheum. Dis. 2006, 65, 30–34. [Google Scholar] [CrossRef] [PubMed]
  24. Munro, B.J.; Steele, J.R.; Bashford, G.M.; Ryan, M.; Britten, N.; Alexander, N.; Schultz, A.; Warwick, D.; Andrews, M.; Arborelius, U.; et al. A kinematic and kinetic analysis of the sit-to-stand transfer using an ejector chair. J. Biomech. 1997, 31, 263–271. [Google Scholar] [CrossRef]
  25. Cieza, A.; Stucki, G. New approaches to understanding the impact of musculoskeletal conditions. Best Pract. Res. Clin. Rheumatol. 2004, 18, 141–154. [Google Scholar] [CrossRef] [PubMed]
  26. Axivity Ltd. Axivity AX3; Axivity Ltd: York, UK. Available online: http://axivity.com/product/ax3 (accessed on 24 August 2017).
  27. Frosio, I.; Pedersini, F.; Borghese, N. Autocalibration of MEMS Accelerometers. IEEE Trans. Instrum. Meas. 2009, 58, 2034–2041. [Google Scholar] [CrossRef]
  28. Pedley, M. Tilt Sensing Using a Three-Axis Accelerometer; Freescale Semiconductor Application Note; Document Number: AN3461; Freescale Semiconductor, Inc.: Austin, TX, USA, 2013; pp. 2012–2013. [Google Scholar]
  29. Yang, G.Z.; Andreu-Perez, J.; Hu, X.; Thiemjarus, S. Multi-sensor Fusion. In Body Sensor Networks; Springer: London, UK, 2014; pp. 301–354. [Google Scholar]
  30. Globerson, A.; Roweis, S. Metric Learning by Collapsing Classes. In Proceedings of the Neural Information Processing Systems, NIPS 2005, Vancouver, BC, Canada, 5–8 December 2005. [Google Scholar]
  31. Engl, H.W.; Grever, W. Using the L–curve for determining optimal regularization parameters. Numer. Math. 1994, 69, 25–31. [Google Scholar] [CrossRef]
  32. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  33. Canu, S.; Grandvalet, Y.; Guigue, V.; Rakotomamonjy, A. SVM and Kernel Methods Matlab Toolbox. Perception Systmes et Information, INSA de Rouen: Rouen, France, 2005. Available online: http://asi.insarouen.fr/enseignants/~arakoto/toolbox/index.html (accessed on 24 August 2017).
  34. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 2951–2959. [Google Scholar]
  35. Cawley, G.C.; Talbot, N.L.C. On Over-fitting in Model Selection and Subsequent Selection Bias in Performance Evaluation. J. Mach. Learn. Res. 2010, 11, 2079–2107. [Google Scholar]
  36. Rifkin, R.; Klautau, A. In Defense of One-Vs-All Classification. J. Mach. Learn. Res. 2004, 5, 101–141. [Google Scholar]
  37. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  38. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  39. Dietterich, T.G. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  40. Lee, H.; Grosse, R.; Ranganath, R.; Ng, A.Y. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; ACM Press: New York, NY, USA, 2009; pp. 1–8. [Google Scholar]
  41. Lee, H.; Ekanadham, C.; Ng, A.Y. Sparse deep belief net model for visual area V2. In Advances in Neural Information Processing Systems 20 (NIPS 2007); Platt, J.C., Koller, D., Singer, Y., Roweis, S.T., Eds.; Curran Associates, Inc.: New York, NY, USA, 2008; pp. 873–880. [Google Scholar]
  42. Ravi, D.; Wong, C.; Lo, B.; Yang, G.Z. Deep learning for human activity recognition: A resource efficient implementation on low-power devices. In Proceedings of the 2016 IEEE 13th International Conference on Wearable and Implantable Body Sensor Networks (BSN), San Francisco, CA, USA, 14–17 June 2016; pp. 71–76. [Google Scholar]
  43. Attal, F.; Mohammed, S.; Dedabrishvili, M.; Chamroukhi, F.; Oukhellou, L.; Amirat, Y. Physical Human Activity Recognition Using Wearable Sensors. Sensors 2015, 15, 31314–31338. [Google Scholar] [CrossRef] [PubMed]
  44. Mannini, A.; Sabatini, A.M. Machine Learning Methods for Classifying Human Physical Activity from On-Body Accelerometers. Sensors 2010, 10, 1154–1175. [Google Scholar] [CrossRef] [PubMed]
  45. Collins, M. The Forward-Backward Algorithm. Available online: http://www.cs.columbia.edu/~mcollins/fb.pdf (accessed on 24 August 2017).
  46. Alsheikh, M.A.; Selim, A.; Niyato, D.; Doyle, L.; Lin, S.; Tan, H.P. Deep Activity Recognition Models with Triaxial Accelerometers. In Proceedings of the Workshops of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  47. Guenterberg, E.; Ghasemzadeh, H.; Loseu, V.; Jafari, R. Distributed Continuous Action Recognition Using a Hidden Markov Model in Body Sensor Networks. In Distributed Computing in Sensor Systems. DCOSS 2009, Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2009; pp. 145–158. [Google Scholar]
  48. Millor, N.; Lecumberri, P.; Gomez, M.; Martinez-Ramirez, A.; Izquierdo, M. Kinematic Parameters to Evaluate Functional Performance of Sit-to-Stand and Stand-to-Sit Transitions Using Motion Sensor Devices: A Systematic Review. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 926–936. [Google Scholar] [CrossRef] [PubMed]
  49. Del Din, S.; Godfrey, A.; Rochester, L. Validation of an Accelerometer to Quantify a Comprehensive Battery of Gait Characteristics in Healthy Older Adults and Parkinson’s Disease: Toward Clinical and at Home Use. IEEE J. Biomed. Heal. Inform. 2016, 20, 838–847. [Google Scholar] [CrossRef] [PubMed]
  50. Marschollek, M.; Rehwald, A.; Wolf, K.H.; Gietzelt, M.; Nemitz, G.; Meyer zu Schwabedissen, H.; Haux, R. Sensor-based Fall Risk Assessment—An Expert ‘to go’. Methods Inf. Med. 2011, 50, 420–426. [Google Scholar] [CrossRef] [PubMed]
  51. Stack, E.; King, R.; Janko, B.; Burnett, M.; Hammersley, N.; Agarwal, V.; Hannuna, S.; Burrows, A.; Ashburn, A. Could In-Home Sensors Surpass Human Observation of People with Parkinson’s at High Risk of Falling? An Ethnographic Study. Biomed Res. Int. 2016, 2016, 1–10. [Google Scholar] [CrossRef] [PubMed]
  52. Inan, O.T.; Whittingslow, D.C.; Teague, C.N.; Hersek, S.; Baran Pouyan, M.; Millard-Stafford, M.; Kogler, G.F.; Sawka, M.N. Wearable Knee Health System Employing Novel Physiological Biomarkers. J. Appl. Physiol. 2017. [Google Scholar] [CrossRef] [PubMed]
  53. D’Lima, D.D.; Fregly, B.J.; Colwell, C.W.J. Implantable sensor technology: Measuring bone and joint biomechanics of daily life in vivo. Arthritis Res. Ther. 2013, 15, 203–211. [Google Scholar] [CrossRef] [PubMed]
  54. Andreu-Perez, J.; Leff, D.R.; Ip, H.M.D.; Yang, G.Z. From Wearable Sensors to Smart Implants–Toward Pervasive and Personalized Healthcare. IEEE Trans. Biomed. Eng. 2015, 62, 2750–2762. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Most common mobility impairments for moderate to severe RA patients.
Figure 1. Most common mobility impairments for moderate to severe RA patients.
Sensors 17 02113 g001
Figure 2. (a) Position of the accelerometer sensor and axes. (b) Signals captured by the accelerometer in the L5 position for seven exemplar activities. The signal plots depict the differences in signal pattern for healthy subjects and RA patients.
Figure 2. (a) Position of the accelerometer sensor and axes. (b) Signals captured by the accelerometer in the L5 position for seven exemplar activities. The signal plots depict the differences in signal pattern for healthy subjects and RA patients.
Sensors 17 02113 g002
Figure 3. Dichotomy of classifiers in a tree-based structure. Each node corresponds to an independent classifier and each leaf is a prediction. This structure is used to generate predictions as will lately work as a logic tree during the inference process. This will subsequently serve the predictions from all nodes in the dichotomy. In the first splitting node, walking is considered a continuously active activity while sitting, standing and lying-down are categorized as passive or temporally active, while transitions can be initially regarded as both. For transitions, these can be performed in a different fashion, they can be either troubled or comforted, depending on the patient conditions. An untroubled transition should be just a temporarily active activity, but if the patient is afflicted enough this could be slower and unorthodox, resembling a continuous activity. Ergo, transitions require a further splitting from cyclical (walking) on one branch and from static (sitting, standing and lying-down) on the other branch.
Figure 3. Dichotomy of classifiers in a tree-based structure. Each node corresponds to an independent classifier and each leaf is a prediction. This structure is used to generate predictions as will lately work as a logic tree during the inference process. This will subsequently serve the predictions from all nodes in the dichotomy. In the first splitting node, walking is considered a continuously active activity while sitting, standing and lying-down are categorized as passive or temporally active, while transitions can be initially regarded as both. For transitions, these can be performed in a different fashion, they can be either troubled or comforted, depending on the patient conditions. An untroubled transition should be just a temporarily active activity, but if the patient is afflicted enough this could be slower and unorthodox, resembling a continuous activity. Ergo, transitions require a further splitting from cyclical (walking) on one branch and from static (sitting, standing and lying-down) on the other branch.
Sensors 17 02113 g003
Figure 4. Training and testing architecture of the dichotomous mapped forest. The logic tree at the end of the testing phase is the same as depicted in Figure 3.
Figure 4. Training and testing architecture of the dichotomous mapped forest. The logic tree at the end of the testing phase is the same as depicted in Figure 3.
Sensors 17 02113 g004
Figure 5. State machine of the logical filter with allowed connections between activity predictions. For instance, once a sample input, representing 3 s, has been predicted as “stand-to-sit”, a “stand”, “lying-to-stand“, “sit-to-stand” or “walking” should have previously been predicted.
Figure 5. State machine of the logical filter with allowed connections between activity predictions. For instance, once a sample input, representing 3 s, has been predicted as “stand-to-sit”, a “stand”, “lying-to-stand“, “sit-to-stand” or “walking” should have previously been predicted.
Sensors 17 02113 g005
Figure 6. Training settings of the classifier. In setting A, data from healthy volunteers is appended to the patient dataset to increase its size. In setting B, only patient data is used to train the classifier model. Finally, in setting C, data from the patient population is used to find the most relevant features and build the mapping; then, the model is trained with healthy data only. In all cases, only patient data is employed to generate the predictions. As one would expect, the data used for tests will come from a different set than the one used for training.
Figure 6. Training settings of the classifier. In setting A, data from healthy volunteers is appended to the patient dataset to increase its size. In setting B, only patient data is used to train the classifier model. Finally, in setting C, data from the patient population is used to find the most relevant features and build the mapping; then, the model is trained with healthy data only. In all cases, only patient data is employed to generate the predictions. As one would expect, the data used for tests will come from a different set than the one used for training.
Sensors 17 02113 g006
Figure 7. Confusion matrices showing predictions as rows and ground truth labels as columns for eight different methods. Numbers inside the confusion matrices represent percentages over the total number of samples predicted as a specific tag.
Figure 7. Confusion matrices showing predictions as rows and ground truth labels as columns for eight different methods. Numbers inside the confusion matrices represent percentages over the total number of samples predicted as a specific tag.
Sensors 17 02113 g007
Figure 8. Performance statistics. Upper part: performance results (accuracy and F-score) for the methods in Table 1; Lower part: performance results for DMF-DL and DMF-Metric applying the logic filter with different orders.
Figure 8. Performance statistics. Upper part: performance results (accuracy and F-score) for the methods in Table 1; Lower part: performance results for DMF-DL and DMF-Metric applying the logic filter with different orders.
Sensors 17 02113 g008
Figure 9. Performance results for applying the logical filter to DMF-Metric (upper) and DMF-DL (lower). The spider diagram shows the percentage of correctly classified input for each independent activity tag.
Figure 9. Performance results for applying the logical filter to DMF-Metric (upper) and DMF-DL (lower). The spider diagram shows the percentage of correctly classified input for each independent activity tag.
Sensors 17 02113 g009
Table 1. Performance values for the methods in Figure 7.
Table 1. Performance values for the methods in Figure 7.
StatisticsDMF-DLDMF-MetricCDBNRF-DLRF-MetricSVM-DLSVM-MetricC-HMM
Accuracy92.44 ± 0.5491.95 ± 0.8284.30 ± 0.9882.25 ± 0.5581.76 ± 0.8672.75 ± 3.7371.84 ± 4.6771.40 ± 3.6
Sensitivity92.20 ± 1.9890.82 ± 1.9878.88 ± 1.1481.42 ± 1.6680.87 ± 0.6274.37 ± 2.0673.78 ± 0.8868.12 ± 2.61
F-score74.22 ± 2.6372.76 ± 2.9654.33 ± 2.9452.02 ± 4.2951.21 ± 3.7539.48 ± 4.4238.70 ± 6.1036.32 ± 4.94
Specificity92.46 ± 0.5592.13 ± 0.9685.04 ± 1.0582.35 ± 0.6781.88 ± 0.9772.49 ± 4.1571.64 ± 5.3171.85 ± 4.92
Table 2. Statistical performance results for DMF-DL and DMF-Metric with activated logic filter.
Table 2. Statistical performance results for DMF-DL and DMF-Metric with activated logic filter.
AlgorithmStatisticsOrder 1Order 2Order 3Order 4Order 5
DMF-DLAccuracy93.78 ± 0.4595.00 ± 0.6594.64 ± 0.5594.43 ± 0.5094.21 ± 0.43
F-score78.05 ± 2.9381.77 ± 2.9080.63 ± 2.8279.97 ± 2.6779.30 ± 2.65
DMF-MetricAccuracy93.75 ± 0.8894.76 ± 1.1594.56 ± 1.1094.31 ± 1.1194.06 ± 1.01
F-score77.92 ± 2.5081.12 ± 3.2080.44 ± 3.1379.67 ± 3.2378.87 ± 2.97
Table 3. Classification results with settings B and C from Figure 6.
Table 3. Classification results with settings B and C from Figure 6.
AlgorithmWithout Logic FilterWith Logic Filter
AccuracyF-scoreAccuracyF-score
DMF-DL patient data only (setting B)78.01 ± 0.9643.78 ± 3.7284.11 ± 1.1253.69 ± 3.22
DMF-DL patient mapping (setting C)83.12 ± 1.4252.31 ± 3.788.84 ± 4.5563.88 ± 4.55
DMF-Metric patient data only (setting B)81.45 ± 1.2850.39 ± 3.8687.63 ± 1.6560.98 ± 4.63
DMF-Metric patient mapping (setting C)85.06 ± 0.9555.75 ± 4.1690.37 ± 1.1561.79 ± 4.46

Share and Cite

MDPI and ACS Style

Andreu-Perez, J.; Garcia-Gancedo, L.; McKinnell, J.; Van der Drift, A.; Powell, A.; Hamy, V.; Keller, T.; Yang, G.-Z. Developing Fine-Grained Actigraphies for Rheumatoid Arthritis Patients from a Single Accelerometer Using Machine Learning. Sensors 2017, 17, 2113. https://doi.org/10.3390/s17092113

AMA Style

Andreu-Perez J, Garcia-Gancedo L, McKinnell J, Van der Drift A, Powell A, Hamy V, Keller T, Yang G-Z. Developing Fine-Grained Actigraphies for Rheumatoid Arthritis Patients from a Single Accelerometer Using Machine Learning. Sensors. 2017; 17(9):2113. https://doi.org/10.3390/s17092113

Chicago/Turabian Style

Andreu-Perez, Javier, Luis Garcia-Gancedo, Jonathan McKinnell, Anniek Van der Drift, Adam Powell, Valentin Hamy, Thomas Keller, and Guang-Zhong Yang. 2017. "Developing Fine-Grained Actigraphies for Rheumatoid Arthritis Patients from a Single Accelerometer Using Machine Learning" Sensors 17, no. 9: 2113. https://doi.org/10.3390/s17092113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop