Next Article in Journal
Twelve-Element MIMO Wideband Antenna Array Operating at 3.3 GHz for 5G Smartphone Applications
Previous Article in Journal
Research on Multi-Layer Defense against DDoS Attacks in Intelligent Distribution Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpretable Support Vector Machine and Its Application to Rehabilitation Assessment

Mobility UX Research Section, Electronics and Telecommunications Research Institute, Daejeon 34129, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(18), 3584; https://doi.org/10.3390/electronics13183584
Submission received: 7 August 2024 / Revised: 23 August 2024 / Accepted: 2 September 2024 / Published: 10 September 2024
(This article belongs to the Section Bioelectronics)

Abstract

:
This paper does present an interpretable support vector machine (SVM) and its application to rehabilitation assessment. We introduce the concept of nearest boundary point to standardize the one-class SVM decision function and determine the shortest path for data from abnormal cases to become those from normal cases. This analytical approach is computationally simple and provides a unique solution. The nearest boundary point of abnormal data can also be used to analyze the cause of abnormal classification and indicate countermeasures for normalization. These properties render the proposed interpretable SVM valuable for medical assessment applications and other problems that require careful consideration of classification results for treatment. Simulation and application results demonstrate the feasibility and effectiveness of the proposed method.

1. Introduction

Recently, machine learning (ML) has been applied to various fields, and its classification algorithms have demonstrated high performance. ML has been actively used in the medical industry, particularly in medical imaging [1,2], and in health metric analyses and predictions [3].
There is a wide range of methods available for assessing muscle function in rehabilitation patients. However, traditional techniques such as manual muscle testing (MMT) provide limited detailed information about muscle function and fail to identify the underlying factors contributing to muscle dysfunction. Additionally, treatment approaches for patients with impaired muscle function are often generalized and not tailored to individual needs. This lack of specificity in both assessment and intervention highlights a significant gap in personalized rehabilitation strategies, underscoring the need for more comprehensive and precise methods to evaluate and address muscle function deficits.
We aimed to develop a classification method that can replace various assessment methods for patients undergoing rehabilitation and performed analysis based on the results obtained from a classification method to establish patient treatment plans. To this end, we formulated the following research questions:
  • How can we handle the imbalance between data from normal and abnormal cases? One challenge in applying ML to healthcare is the imbalance between data from normal and abnormal cases. It is easier to obtain data from normal individuals than from patients presenting specific health conditions, and not all patient data are abnormal. Consequently, accurately labeling abnormality may be challenging.
  • How can we use patient treatment plans considering classification results? In addition to achieving a high classification performance, developing an interpretable classifier is crucial for applying the results to support patient treatment planning. The classifier interpretability may increase trust and usability, thus facilitating the development of treatment plans based on classification results.
The first research question focuses on addressing the issue of data imbalance, where the dataset contains two classes—normal and abnormal—with the abnormal class significantly underrepresented. To tackle this imbalance, several approaches have been considered. One approach is the synthetic minority over-sampling technique (SMOTE) [4], which resamples the minority (rare) class. Another approach is an ensemble method, where the majority (abundant) class is split into smaller subsets, each of which is then trained with the entire rare class to create multiple models, which are subsequently combined [5]. A third method involves clustering the abundant class and training only on representative values (e.g., medians) of these clusters [6].
However, these methods have limitations when applied to our context, where the imbalance is severe. The SMOTE technique is less effective because it requires identifying the distribution pattern of the rare class, which is challenging when the class is extremely underrepresented. Similarly, ensemble and clustering methods are not suitable for our case, as they effectively reduce the amount of data used for modeling, which is problematic given our already limited dataset.
Given these challenges, we chose to address the anomaly detection problem by adopting a one-class classification approach. This method involves training the model exclusively on the abundant class and classifying the remaining data as belonging to the rare class. This approach is particularly well suited to our scenario, as it allows for robust detection of anomalies despite the significant data imbalance.
Several methods have been developed to perform the one-class classification [7,8], from which we used a one-class support vector machine (SVM) [9]. This type of SVM is well-suited for anomaly detection, which identifies datapoints that differ substantially from most of the data. It captures the data distribution and identifies outliers. Another advantage of the one-class SVM is its ability to handle imbalanced datasets, where one class is represented by drastically fewer samples than another one. The one-class SVM can handle situations in which only positive samples are available and negative samples are unknown or difficult to obtain. Hence, it is adequate for medical assessments, especially when collecting data from patients with rare conditions is challenging.
To address the second research question, we devised the nearest boundary point (NBP) to solve for the shortest Mahalanobis distance. Specifically, based on the decision function of the one-class SVM formulated as a mixture of normal distributions, we standardized and analytically derived a series of steps to obtain and reconstruct the NBP. The analytically derived solution is computationally simple and yeilds a unique solution, making it particularly valuable for our rehabilitative assessment application.
To validate the feasibility and applicability of the proposed method, we first performed simulations using artificially generated two-dimensional (2D) nonlinear data. We chose 2D data because they allow for easy visualization and understanding of intermediate steps in addition to verification. After analyzing the simulation results, we tested the proposed interpretable SVM on real data. We trained the one-class classification model on biosensing measurements from normal individuals and classified patient data to solve the NBP problem for identified abnormal data. We analyzed the major factors causing abnormal classifications and investigated feasible treatment methods.
The remainder of this paper is organized as follows. In Section 2, we introduce an interpretable one-class SVM with Gaussian kernels. Section 3 presents the simulation results, and Section 4 reports the analysis of an application to rehabilitation based on the proposed method, including the identification of the main factors and corresponding treatment options for patients classified as abnormal. The simulation and application results validate the effectiveness and interpretability of the proposed SVM. Finally, Section 5 presents our concluding remarks.

2. The Interpretable One-Class SVM

2.1. One-Class Classification of Normally Distributed Data

We address one-class classification as a research question. One-class classification detects anomalies by defining the boundary that encloses normal data. In this subsection, we discuss the case of normal data that follow a normal distribution.
The probability density function of data that following a normal distribution is expressed as
ϕ ( z | μ , Σ ) = 1 ( 2 π ) d / 2 | Σ | 1 / 2 exp 1 2 ( z μ ) T Σ 1 ( z μ ) ,
where μ , Σ , and d denote the mean, covariance, and dimension of z , respectively.
When data follow a normal distribution, an appropriate boundary can be determined by fitting the data to a normal distribution with a certain mean and covariance and then obtaining the contour of the probability density function, which can be determined by a desired acceptance rate or other approaches.
For example, to determine the boundary of 2D data { z } that follow a normal distribution with an acceptance rate of approximately 0.9502, we can determine the distance at which the cumulative distribution function of the normal distribution is 0.9502, as shown in Figure 1, and draw a contour with the same distance to obtain the boundary.
The obtained boundary serves as a criterion for detecting anomalies, and new data can be classified as normal if they fall within the boundary or abnormal if they fall outside the boundary.

2.2. Classification Interpretability Using the NBP

We obtain the boundary for normally distributed data, thereby being able to distinguish between normal and abnormal data based on whether they fall within or outside the boundary. If the data are classified as abnormal because they fall outside the boundary, we analyze the interpretability of classification and the changes that would be necessary for them to be classified as normal.
To ensure interpretability, we aim to solve the NBP problem by finding the boundary point closest to a datapoint z that falls outside the boundary. To determine the concept of nearest point, we must first define a distance, which we take to be the Mahalanobis distance. Given a probability distribution with mean μ and positive-definite covariance Σ , Mahalanobis distance Δ 2 ( z ) is defined as
Δ 2 ( z ) = ( z μ ) T Σ 1 ( z μ ) .
The Mahalanobis distance is an effective multivariate metric of the distance between a point (vector) and a data distribution. It is suitable for multivariate anomaly detection, classification of highly imbalanced datasets, one-class classification, and untapped use cases. In particular, when Mahalanobis distance Δ 2 ( z ) of point z from the distribution in Equation (1) is constant, the point lies on the same contour as the normal probability density function. Therefore, the boundary of normal data can be expressed as
Δ 2 ( z ) = R 2 ,
where R is a real-valued constant.
If the dataset follows normal distribution N ( μ , Σ ) and the boundary is determined by Equation (3), the optimization problem can be solved to find the NBP of z observed outside the boundary. Given z , μ , and Σ , we have
arg min z ¯ ( z z ¯ ) T Σ 1 ( z z ¯ ) s . t . ( z ¯ μ ) T Σ 1 ( z ¯ μ ) = R 2 ,
where z ¯ denotes the NBP. To solve Equation (4), we obtain the following Laplacian:
L ( z ¯ , λ ) = ( z z ¯ ) T Σ 1 ( z z ¯ ) λ ( ( z ¯ μ ) T Σ 1 ( z ¯ μ ) R 2 ) .
The partial derivatives of the Laplacian are given by
L λ = ( z ¯ μ ) T Σ 1 ( z ¯ μ ) R 2 = 0 ,
L z ¯ = 0 ,
From Equation (7), we obtain the NBP as follows:
z ¯ = 1 1 + λ z + λ 1 + λ μ ,
By substituting Equation (8) into Equation (6), we obtain
λ = 1 ± Δ 2 ( z ) R 2 .
If the sign of the square root term in Equation (9) is positive, boundary point z ¯ is the nearest one, and if the sign is negative, z ¯ is the farthest boundary point, as shown in Figure 2.
By obtaining the NBP, z ¯ , for a new observation, z , outside the boundary, we can gain insights into interpretation, such as why z is outside the boundary, how far it is from the boundary, and which element values must be increased or decreased to classify it as a normal point.

2.3. Standardization of Non-Normally Distributed Data

In practice, data may not be normally distributed. To address this issue, we model data as a mixture of multiple normal distributions. When data that do not follow a normal distribution are represented as a mixture of multiple normal distributions, we can apply a normalization method.
Let x be a d-dimensional random variable generated by a mixture of multiple normal distributions as follows:
f ( x ) = i = 1 N α i ϕ ( x | μ i , Σ i ) ,
where α i > 0 , i = 1 N α i = 1 , ϕ ( · | μ i , Σ i ) represents the probability density function of a d-dimensional normal distribution with mean μ i and covariance Σ i , and N is the number of components in the mixture. We introduce membership weight (or posterior probability) w i of class i for x by using Bayes’ theorem as follows:
w i = α i ϕ i ( x ) i = 1 N α i ϕ i ( x ) ,
where ϕ i ( x ) = ϕ ( x | μ i , Σ i ) . By applying the soft assignment proposed in [10], we obtain the following standardization formula:
z = i = 1 N w i Σ i 1 / 2 x i = 1 N w i μ i ,
where z N ( 0 , I d ) is the standardized d-dimensional random vector transformed from x for I d being a d × d identity matrix. Figure 3 illustrates the standardization of a mixture of two normal distributions.

2.4. The NBP of the One-Class SVM

The one-class SVM belongs to a family of ML algorithms known as kernel methods that use kernel functions to represent input features into a new, often higher-dimensional, space. Thus, it facilitates distinguishing between different classes, and the resulting decision boundary is often simpler and more interpretable than that in the original feature space. This approach avoids the need for explicit feature transformations, which can be computationally expensive, and is often referred to as the kernel trick.
A Gaussian kernel is commonly used to perform the kernel trick and is universal, meaning that it can uniformly approximate any arbitrary continuous target function. Gaussian kernels offer a valuable advantage by transferring data from the original plane into a hyperplane for analysis. This capability can resolve a significant limitation of the Mahalanobis distance which is known to perform inadequately for datasets characterized by linearity and heteroscedasticity. By utilizing Gaussian kernels, the shortcomings associated with Mahalanobis distance in these contexts can be effectively mitigated.
After training, the decision function of the SVM for observation x can be obtained as follows:
g ( x ) = i = 1 N β i φ ( x | x i , σ ) b ,
where x i denotes the i-th support vector, b is the bias term, N is the number of support vectors, and φ ( a | b , σ ) is a Gaussian kernel with width σ and given by
φ ( a | b , σ ) = exp ( a b ) T ( a b ) 2 σ 2 .
Because the decision function in Equation (13) is a linear sum of Gaussian functions, we have the following relationship between g ( x ) in Equation (13) and f ( x ) in Equation (10):
g ( x ) = i = 1 N β i exp ( x x i ) T ( x x i ) 2 σ 2 b , K ( g ( x ) + b ) = i = 1 N α i ϕ ( x | μ i , Σ i ) , = f ( x )
where K = { ( 2 π ) d / 2 | Σ | 1 / 2 i = 1 N β i } 1 , α i = β i i = 1 N β i , Σ i = σ 2 I d , and μ i = x i for all i = 1 ,   ,   N .
As shown in Equation (15), K ( g ( x ) + b ) takes the form of multiple mixtures with normal distributions. As K and b are constants, we can transform the distribution of an x space into a standardized normal distribution of another z space using Equations (11) and (12). The posterior probability can be derived as
w i = β i φ ( x | x i , σ ) i = 1 N β i φ ( x | x i , σ ) = β i φ ( x | x i , σ ) g ( x ) + b .
As i = 1 N w i = 1 and Σ i = σ I d , we have
z = 1 σ x i = 1 N w i x i .
As z follows a normal distribution, that is, z N ( 0 , I d ) , we obtain λ from Equation (9) as follows:
λ = z T z R 2 1 .
By numerically computing x b = arg min x g 2 ( x ) from training data, we calculate z b using Equation (17) and derive the Mahalanobis distance of the boundary as R 2 = z b T z b . Substituting λ into Equation (8), we obtain the following NBP in the z space:
z ¯ = z b T z b z T z z ,
and reconstructing z ¯ into the x space, we obtain the NBP of the new observation, x .
x ¯ = σ z ¯ + i = 1 N w ¯ i x i .
As w ¯ i cannot be computed explicitly, we use w i instead of w ¯ i and calculate Equation (17) by letting x = x ¯ and applying Equations (19) and (20) repeatedly until x ¯ converges.
Figure 4 illustrates the proposed process for determining the desired point (that is, NBP) from the SVM decision function.

3. Simulation Results

This section presents simulation results used to validate the proposed method for solving the NBP in the one-class SVM. The results are reported stepwise, and the final outcomes are provided. For the simulation, we used 2D nonlinear data. As shown in Figure 5, we randomly sampled 1571 datapoints (blue circles) from a normal distribution with mean [ 10 cos θ , 10 sin θ ] T and covariance 2 2 I 2 , where θ ( 0 , π ) and I 2 is a 2D identity matrix. We then applied the one-class SVM to extract a region bounded by a decision function (black contour in the figure).
Figure 6 shows the standardized datapoints (red circles) and the boundary in the z space, which correspond to the original datapoints in the x space in Figure 5. After transformation, the data were normally distributed and centered at the origin of the z plane. To quantitatively evaluate the effectiveness of the standardization in Figure 6, we performed a Kolmogorov–Smirnov test [11] on the 2D data. The results confirmed the normality of the standardized datapoints, with a significance level of p < 0.00 .
The decision function for the SVM was obtained through training. Figure 7 shows the test data used to evaluate the decision function. A positive classification was assigned to the test data that fell within the decision boundary, whereas a negative classification was assigned to the test data outside the boundary. As we focused on negative samples, we solved the NBP problem for negative test samples. First, we standardized the negative test samples and computed their NBPs, as shown in Figure 8. The x marks in the figure represent the standardized negative test samples, and the triangular marks represent their NBPs. The NBPs lie on the black contour, which represents the decision boundary.
By mapping the datapoints and the NBPs from the z plane, as shown in Figure 8, onto the x plane, we obtained the final solution (that is, the desired points), as illustrated in Figure 9. The x marks in the figure represent negative test samples, whereas the triangular marks represent their NBPs. The NBPs lie on the black contour, which represents the decision boundary.

4. Applications

We present a practical application of the proposed interpretable SVM to rehabilitation assessment along with experimental results obtained from analysis and testing. Our aims were to evaluate the forearm muscle function of patients undergoing rehabilitation through measurements of surface electromyography (sEMG) and an inertial measurement unit (IMU) as well as to determine appropriate countermeasures for functional electric stimulation (FES) and physical therapy using a robot. Specifically, we employed the exoPill sensor [12], which is a modular device developed by ExoSystems Inc. (Gyeonggi-do, Republic of Korea). It integrates sEMG electrodes and an IMU, as depicted in Figure 10A. We considered wrist extension as the evaluated rehabilitation movement, as illustrated in Figure 10B, by attaching the sEMG electrodes to the extensor carpi radialis muscles and the IMU to the back of the hand, as shown in Figure 10C.
We recruited 40 participants for our study, including 20 healthy individuals and 20 patients who recovered from stroke. Each participant wore a measuring device and alternated resting and wrist extensions 10 times at 5 s intervals. For the stroke patients, the measuring device was worn on the side affected by hemiplegia, and an upper-extremity Fugl–Meyer assessment (FMA) was conducted by a clinical staff member. The maximum score for the FMA was 14.
From the sEMG signal, features such as power (PWR), median frequency (MDF), mean frequency (MNF), and root mean square (RMS) were extracted. From the IMU signal, the velocity (VEL) and range of movement (ROM) were extracted. Using these features from normal individuals as positive samples, we trained the one-class SVM and obtained a decision function by fine-tuning. To improve the training results, we performed data augmentation through a boosting technique.
The one-class SVM decision function trained using the features extracted from normal individuals was applied to the features of stroke patients to classify the muscle function of each patient. For patients classified as abnormal, the NBP was solved to analyze the causes of such classification and possible countermeasures for improving rehabilitation. This process was performed considering the 14 patients whose data were excluded owing to sensor signal omission or mismeasurement.
The evaluation results are listed in Table 1. Out of the 14 patients, 10 were classified as abnormal. For each of these 10 patients, we analyzed the differences between the measurements and NBPs per feature.
A higher difference indicated that more effort was required for patients to be classified as normal. To identify the main factors contributing to patient’s abnormalities, we used the standard deviation of the measurements from normal individuals. The corresponding results are listed in Table 2, where + and − indicate normal and abnormal, respectively.
By applying the interpretable SVM proposed in this paper, it is possible to classify the muscle function of patients as either normal or abnormal based on sEMG and IMU measurements. In cases where the function is classified as abnormal, the method can identify the main factors responsible for the abnormality. Depending on these factors, such as whether the abnormal muscle function is due to issues with muscle recruitment or physical problems like muscle contractures, appropriate treatments can be prescribed. These treatments may include FES for muscle recruitment issues, or robotics approach (e.g., exoskeleton) for physical problems.
As indicated in Table 2, the SVM decision function classified patients with FMA scores of 14 (that is, the highest score) as normal and those with lower scores as abnormal. In the abnormal group, the main abnormality factors were identified by solving the NBP problem, and different factors were identified per patient. Once the main abnormality factors for each patient were determined, PWR, MDF, MNF, and RMS derived from sEMG signals were used to define countermeasures for FES treatment. On the other hand, VEL and ROM provided guidelines for using robotic therapy as physical assistance. For example, for patient 3, PWR and MDF were the main abnormal factors, indicating the need for physical therapy using FES to address muscle recruitment issues. Patients 6, 10, 11, and 18 required FES treatment and additional physical therapy using robotic assistance to address movement dynamic issues.

5. Conclusions

We propose an interpretable SVM that enables the derivation of the causes of data classified as abnormal by one-class classification and the analysis of countermeasures for rehabilitation. The proposed method is based on the Mahalanobis distance and involves obtaining the NBP of abnormal data. The NBP indicates the shortest path from the abnormal data to the normal boundary and allows to identify features that contribute to anomaly along with their contribution extent.
The proposed method for obtaining the NBP of abnormal data involves standardizing the SVM decision function. Given that this function with a Gaussian kernel can be expressed as a mixture of normal distributions in the x space, we can approximate the decision function to a normal distribution in the z space. After standardization, we obtain the NBP in the z space and reconstruct it back to the x space.
We used a 2D nonlinear dataset to simulate and test the proposed method, validating its feasibility and applicability through simulations. Moreover, we applied the proposed method to rehabilitation. Using sEMG and IMU measurements, we classified the status of the patient’s muscle function and analyzed the underlying causes of anomalies and possible therapeutic countermeasures.

Author Contributions

Conceptualization, W.K. and D.Y.; methodology, W.K.; software, W.K.; validation, W.K., H.-S.K. and H.J.; formal analysis, H.-S.K.; investigation, W.K.; resources, W.K.; data curation, W.K.; writing—original draft preparation, W.K.; writing—review and editing, W.K.; visualization, W.K.; supervision, D.Y.; project administration, D.Y.; funding acquisition, W.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Institute of Information and Communications Technology Planning and Evaluation (IITP) under grant funded by the Korea government (MSIT) (2022-0-00501).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Pusan National University Hospital (approval no. 2107-027-105).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to patients’ privacy issue.

Acknowledgments

We thank the staff at Pusan National University Hospital for collecting data of muscle function from rehabilitation patients.

Conflicts of Interest

The authors declare no conflicts of interest and the funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
SVMSupport Vector Machine
MLMachine Learning
NBPNearest Boundary Point 
sEMGSurface Electromyography
IMUInertial Measurement Unit
FESFunctional Electric Stimulation
FMAFugl-Meyer Assessment
PWRPower
MDFMedian Frequency
MNFMean Frequency
RMSRoot Mean Square
VELVelocity
ROMRange of Movement

References

  1. Giger, M.L. Machine learning in medical imaging. J. Am. Coll. Radiol. 2018, 15, 512–520. [Google Scholar] [CrossRef] [PubMed]
  2. Currie, G.; Hawk, K.E.; Rohren, E.; Vial, A.; Klein, R. Machine learning and deep learning in medical imaging: Intelligent imaging. J. Med. Imaging Radiat. Sci. 2019, 50, 477–487. [Google Scholar] [CrossRef] [PubMed]
  3. Massaro, A.; Ricci, G.; Selicato, S.; Raminelli, S.; Galiano, A. Decisional support system with artificial intelligence oriented on health prediction using a wearable device and big data. In Proceedings of the 2020 IEEE International Workshop on Metrology for Industry 4.0 and IoT, Roma, Italy, 3–5 June 2020; pp. 718–723. [Google Scholar]
  4. Chawala, N.; Bowyer, K.; Hall, L.; Kegelmeyer, W. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  5. Liu, L.; Wu, X.; Li, S.; Li, Y.; Tan, S.; Bai, Y. Solving the class imbalance problem using ensemble algorithm: Application of screening for aortic dissection. BMC Med. Inform. Decis. Mak. 2022, 22, 82. [Google Scholar] [CrossRef] [PubMed]
  6. Xuan, L.; Zhigang, C.; Fan, Y. Exploring of clustering algorithm on class-imbalanced data. In Proceedings of the 8th International Conference on Computer Science and Education, Colombo, Sri Lanka, 26–28 April 2013. [Google Scholar]
  7. Liu, Y.-H.; Huang, H.-P. Towards a high-stability EMG recognition system for prosthesis control: A one-class classification based non-target EMG pattern filtering scheme. In Proceedings of the 2009 IEEE International Conference on Systems, Man and Cybernetics, San Antonio, TX, USA, 11–14 October 2009; pp. 4752–4757. [Google Scholar]
  8. Faula, Y.; Eglin, V.; Bres, S. One-class detection and classification of defects on concrete surfaces. In Proceedings of the 2021 IEEE International Conference on Systems, Man, and Cybernetics, Melbourne, Australia, 17–20 October 2021; pp. 826–831. [Google Scholar]
  9. Schölkopf, B.; Williamson, R.C.; Smola, A.; Shawe-Taylor, J.; Platt, J. Support vector method for novelty detection. Adv. Neural Inf. Process. Syst. 1999, 12, 582–588. [Google Scholar]
  10. Li, M.; Schwartzman, A. Standardization of multivariate Gaussian mixture models and background adjustment of PET images in brain oncology. Ann. Appl. Stat. 2018, 12, 2197–2227. [Google Scholar] [CrossRef]
  11. Marsaglia, G.; Tsang, W.; Wang, J. Evaluating Kolmogorov’s Distribution. J. Stat. Softw. 2003, 8, 1–4. [Google Scholar] [CrossRef]
  12. exoPill: AI-Based Wearable Healthcare Solution. Available online: https://www.exosystems.io/exopill?lang=en (accessed on 1 August 2023).
Figure 1. Cumulative distribution function of 2D normal data and boundary at acceptance rate of 0.9502. We used the squared distance Δ 2 as the Mahalanobis distance, the details of which are provided in Section 2.2.
Figure 1. Cumulative distribution function of 2D normal data and boundary at acceptance rate of 0.9502. We used the squared distance Δ 2 as the Mahalanobis distance, the details of which are provided in Section 2.2.
Electronics 13 03584 g001
Figure 2. Solved NBP and farthest boundary point of normally distributed 2D data with boundary shown in Figure 1. The red point is the NBP of the green point, z , which is a new observation, while the light purple point is the farthest boundary point (FBP) in terms of the Mahalanobis distance given by Equation (9).
Figure 2. Solved NBP and farthest boundary point of normally distributed 2D data with boundary shown in Figure 1. The red point is the NBP of the green point, z , which is a new observation, while the light purple point is the farthest boundary point (FBP) in terms of the Mahalanobis distance given by Equation (9).
Electronics 13 03584 g002
Figure 3. Standardization of mixture of multiple normal distributions. (a) Probability density function of mixture of two normal distributions with means of ( 0.5 , 0 ) and ( 0.5 , 2 ) and covariances of [ 1 , 0.7 ; 0.7 , 0.7 ] and [ 0.7 , 0.7 ; 0.7 , 1 ] . (b) Probability density function of standardized mixture.
Figure 3. Standardization of mixture of multiple normal distributions. (a) Probability density function of mixture of two normal distributions with means of ( 0.5 , 0 ) and ( 0.5 , 2 ) and covariances of [ 1 , 0.7 ; 0.7 , 0.7 ] and [ 0.7 , 0.7 ; 0.7 , 1 ] . (b) Probability density function of standardized mixture.
Electronics 13 03584 g003
Figure 4. Block diagram for finding the desired point (or the NBP) x ¯ when a new observation x is classified as negative. Each standardized z , NBP z ¯ , and reconstructed point x ¯ can be computed by Equations (17), (19) and (20), respectively.
Figure 4. Block diagram for finding the desired point (or the NBP) x ¯ when a new observation x is classified as negative. Each standardized z , NBP z ¯ , and reconstructed point x ¯ can be computed by Equations (17), (19) and (20), respectively.
Electronics 13 03584 g004
Figure 5. Randomly sampled (1571) datapoints (blue circles) and their one-class SVM boundary (black contour). The boundary effectively encloses the majority of the datapoints, excluding a few outliers, demonstrating a smooth and robust classification performance that is less sensitive to noise.
Figure 5. Randomly sampled (1571) datapoints (blue circles) and their one-class SVM boundary (black contour). The boundary effectively encloses the majority of the datapoints, excluding a few outliers, demonstrating a smooth and robust classification performance that is less sensitive to noise.
Electronics 13 03584 g005
Figure 6. Standardized datapoints (red circles) and their boundary (black contour) in the z space. The boundary effectively encloses the majority of the datapoints while excluding outliers, similar to the original space. In the 2D standardized space, the boundary appears geometrically circular.
Figure 6. Standardized datapoints (red circles) and their boundary (black contour) in the z space. The boundary effectively encloses the majority of the datapoints while excluding outliers, similar to the original space. In the 2D standardized space, the boundary appears geometrically circular.
Electronics 13 03584 g006
Figure 7. Decision boundary (black contour) that separates positive samples (blue circles) from negative samples (red circles). The boundary effectively distinguishes between the two classes, even for datapoints near the decision boundary.
Figure 7. Decision boundary (black contour) that separates positive samples (blue circles) from negative samples (red circles). The boundary effectively distinguishes between the two classes, even for datapoints near the decision boundary.
Electronics 13 03584 g007
Figure 8. Standardized negative test points (x marks) and their NBPs (triangular marks). The NBPs lie on the decision boundary (black contour) and are accurately identified in the direction toward the center of the circular boundary.
Figure 8. Standardized negative test points (x marks) and their NBPs (triangular marks). The NBPs lie on the decision boundary (black contour) and are accurately identified in the direction toward the center of the circular boundary.
Electronics 13 03584 g008
Figure 9. Negative test samples (x marks) and their NBPs (triangular marks). The NBPs lie on the decision boundary (black contour). The results show that the nearest boundary points identified in the x plane are consistently mapped back to the x plane, effectively locating the nearest points on the decision boundary.
Figure 9. Negative test samples (x marks) and their NBPs (triangular marks). The NBPs lie on the decision boundary (black contour). The results show that the nearest boundary points identified in the x plane are consistently mapped back to the x plane, effectively locating the nearest points on the decision boundary.
Electronics 13 03584 g009
Figure 10. (A) exoPill measurement system developed by ExoSystems. (B) Wrist extension for evaluation. (C) Experimental setup.
Figure 10. (A) exoPill measurement system developed by ExoSystems. (B) Wrist extension for evaluation. (C) Experimental setup.
Electronics 13 03584 g010
Table 1. Subtraction of NBPs and negative test samples.
Table 1. Subtraction of NBPs and negative test samples.
No.PWRMDFMNFRMSVELROM
20.0216 28.6384   2.1470 2.7642   6.20590.3870
3 0.0891   20.4617   4.18151.1141−6.86740.1913
4 0.0746   27.4724   −3.1780 2.2990   0.15510.2810
6 0.0788   47.4210   24.4754   2.6820   23.0072   1.0584  
7 0.0694   35.6850   −7.1576 2.4398   0.80680.2192
8 0.0726   −0.34921.3856 4.0477   5.46160.2313
10 0.1037   56.0339   9.5026 3.7628   18.7112   1.2062  
110.0496 38.2881   35.0693   0.4746 18.8491   0.3364
17−0.0014−0.8882−2.4736−0.43335.43640.0162
18 0.0883   28.7705   2.2433−0.1190 15.5825   0.8308  
σ 0.02166.32664.04560.48054.09950.1378
† When the difference between the NBP and corresponding feature of the negative test sample was larger than 3 σ , we identified the feature as the main factor, where σ denotes the standard deviation of each feature from the value for normal individuals.
Table 2. Results of muscle function assessment.
Table 2. Results of muscle function assessment.
No.FMASVMMain Factors from NBP
212MDF, RMS
39PWR, MDF
412PWR, MDF, RMS
60PWR, MDF, MNF, RMS, VEL, ROM
712PWR, MDF, RMS
812PWR, RMS
101PWR, MDF, RMS, VEL, ROM
111MDF, MNF, VEL
1214+
1514+
1614+
1712None
186PWR, MDF, VEL, ROM
2014+
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, W.; Joe, H.; Kim, H.-S.; Yoon, D. Interpretable Support Vector Machine and Its Application to Rehabilitation Assessment. Electronics 2024, 13, 3584. https://doi.org/10.3390/electronics13183584

AMA Style

Kim W, Joe H, Kim H-S, Yoon D. Interpretable Support Vector Machine and Its Application to Rehabilitation Assessment. Electronics. 2024; 13(18):3584. https://doi.org/10.3390/electronics13183584

Chicago/Turabian Style

Kim, Woojin, Hyunwoo Joe, Hyun-Suk Kim, and Daesub Yoon. 2024. "Interpretable Support Vector Machine and Its Application to Rehabilitation Assessment" Electronics 13, no. 18: 3584. https://doi.org/10.3390/electronics13183584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop