Next Article in Journal
The Dynamic Photometric Stereo Method Using a Multi-Tap CMOS Image Sensor
Next Article in Special Issue
Experimental Robot Model Adjustments Based on Force–Torque Sensor Information
Previous Article in Journal
A New Method for Immobilization of His-Tagged Proteins with the Application of Low-Frequency AC Electric Field
Previous Article in Special Issue
Approach for the Development of a Framework for the Identification of Activities of Daily Living Using Sensors in Mobile Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Virtual Sensors for Advanced Controllers in Rehabilitation Robotics

by
Aitziber Mancisidor
1,†,
Asier Zubizarreta
1,*,†,
Itziar Cabanes
1,†,
Eva Portillo
1,† and
Je Hyung Jung
2,†
1
Department of Automatic Control and System Engineering, Faculty of Engineering in Bilbao, University of the Basque Country (UPV/EHU), Plaza Ingeniero Torres Quevedo 1, 48013 Bilbao, Spain
2
Neurorehabilitation Area, Health Division, TECNALIA Research and Innovation, Mikeletegi Pasealekua 1-3, Donostia-San Sebastian 20009, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2018, 18(3), 785; https://doi.org/10.3390/s18030785
Submission received: 26 January 2018 / Revised: 2 March 2018 / Accepted: 2 March 2018 / Published: 5 March 2018
(This article belongs to the Special Issue Smart Sensors for Mechatronic and Robotic Systems)

Abstract

:
In order to properly control rehabilitation robotic devices, the measurement of interaction force and motion between patient and robot is an essential part. Usually, however, this is a complex task that requires the use of accurate sensors which increase the cost and the complexity of the robotic device. In this work, we address the development of virtual sensors that can be used as an alternative of actual force and motion sensors for the Universal Haptic Pantograph (UHP) rehabilitation robot for upper limbs training. These virtual sensors estimate the force and motion at the contact point where the patient interacts with the robot using the mathematical model of the robotic device and measurement through low cost position sensors. To demonstrate the performance of the proposed virtual sensors, they have been implemented in an advanced position/force controller of the UHP rehabilitation robot and experimentally evaluated. The experimental results reveal that the controller based on the virtual sensors has similar performance to the one using direct measurement (less than 0.005 m and 1.5 N difference in mean error). Hence, the developed virtual sensors to estimate interaction force and motion can be adopted to replace actual precise but normally high-priced sensors which are fundamental components for advanced control of rehabilitation robotic devices.

1. Introduction

Global aging and a sedentary lifestyle in industrialized countries have increased the number of patients with reduced mobility due to stroke [1]. Nowadays, more than 33 million people around the world have to live with stroke sequels [2]. In order to recover the lost mobility and improve the quality of life of these patients, it is necessary to perform appropriate rehabilitation treatments which usually require several years to provide the desired results [3]. However, in most rehabilitation programs trainings hours are limited due to staff and financial constraints [4].
Due to the need of increasing the effectiveness and the resources of rehabilitation for stroke patients, over last couple of decades robotic devices have been proposed to execute upper [5,6] and lower limb [7,8,9] rehabilitation programs. Rehabilitation robots emulate and replicate the movements and exercises performed by a physiotherapist, obtaining treatments of higher intensity, frequency and repetitiveness that can allow better recovery of patients with reduced mobility [3]. In addition, they can work as a measurement tool of the interaction forces and motions, providing data for the analysis of the state of recovery of the patient [4]. Finally, integrating an adequate graphic interface, they can motivate the patients to continue with the rehabilitation program by using games [10].
The long training process is composed by several phases, in which the patient mobility is gradually increased [11]. Hence, robotic devices need to adapt their performance and task execution to the particular needs of each patient. This way, in the early stages of rehabilitation, patients typically cannot move autonomously, and the movement must be executed by the robot. After a few months of training, patients gradually recover part of their mobility, and they must be encouraged to generate movement on their own. This is, passive training mode is required for initial stages, while the so-called active training mode must be used in more advanced phases [12].
In order to implement these training modes, it is necessary to control both the force and motion of the robot/user interaction. In passive mode, a suitable position controller is necessary, so that the robotic device executes the desired trajectory. On the other hand, in active mode, a force controller is required to assist the user to perform the rehabilitation task [13]. These controllers calculate the torque/force to be applied by the actuators of the robotic device based on the reference generated by the selected training mode (passive or active) and the robot/user interaction feedback (force or motion).
Usually, this interaction feedback is measured directly using position and force sensors. However, although various research groups work in this line [14], there are still no specific sensors for this type of applications. Hence, most rehabilitation robots that can be found in the literature use industrial sensors, which are not adapted for training therapies [15,16,17] and present several drawbacks: (1) rehabilitation robots usually present complex structures, and the placement of sensors in the user/robot interaction point is a challenging task; (2) the introduction of a number of sensors in wearable robots may increase their weight, which is not desirable; (3) noise and temperature dependency of the force sensors require adequate processing of sensors signals; and finally; (4) the use of accurate sensors increases the cost of robotic rehabilitation devices, reducing its market growth [18].
The lack of proper sensors to measure the robot-human interaction has motivated some researchers to develop sensorless controllers. Usually, these advanced algorithms estimate the interaction forces and motions based on the dynamic/kinematic model of the robot and the measurements provided by the integrated sensors of the actuator systems [19,20]. However, most of sensorless approaches are focused on industrial [21] or surgical [22] applications, while very few are applied to rehabilitation robots [23], where the main focus is estimating the robot-human interaction.
In this context, this work presents the use of virtual sensors as an alternative to direct measurement using accurate sensors in rehabilitation robotics. Virtual sensors allow to estimate the robot/user interaction force and motion, reducing the overall cost and the complexity of the robots. For that purpose, a kinematical/dynamical mathematical model of the robotic device and low cost position sensors, such as optical encoders or linear potentiometers are used. This approach is applied to the Universal Haptic Pantograph (UHP) upper limb rehabilitation robot, used for the training of shoulder, elbow and wrist in people who suffer motor deficit after a stroke [24].
The rest of this paper is organized as follows: Section 2 introduces the Universal Haptic Pantographth rehabilitation robot. Section 3 describes the concept and theoretical development of the presented force and motion virtual sensors. The approach is validated Section 4, comparing the performance of the robot controller with real sensors and virtual ones. Finally, the most important ideas are summarized in Section 5.

2. UHP Rehabilitation Robot

The UHP robotic device is a multifunctional upper limb rehabilitation robot designed for the treatment of people with reduced mobility due to stroke [24]. Thanks to the use of three lockable/unlockable joints ( P B , P D and P E ), the UHP has the ability to configure its mechanical structure (Figure 1). This reconfigurable structure allows to define eight operating modes. However, for the case of study analyzed in this work, two of the most used will be detailed: ARM and WRIST modes [25].
The ARM operating mode allows to perform planar motions that result in arm flexion/extension exercises. This allows to execute rehabilitation exercises associated with the three movements of the shoulder (rotation, flexion/extension and abduction/adduction) and the movement of flexion/extension of the elbow [26]. On the other hand, WRIST operating mode allows to perform rehabilitation exercises associated with the rotating movements of the wrist (abduction/adduction and flexion/extension) and the supination/pronation movement of the forearm [27]. So, all joints of the upper limb can be trained by combining these two operating modes (Figure 1).
The UHP rehabilitation robot is divided into two subsystems: a pantograph-based robotic device and a drive system based on elastic components (Figure 2). The pantograph structure interacts with the patient at the contact point ( P Cn ), while the drive system provides the force/motion to the robotic device by means of two ortogonally placed Serial Elastic Actuators (SEA) [28]. Each SEA actuator is composed of a rotary motor and two linear springs connected by rigid cables, as shown in Figure 2. Hence, the motion of the drive system based on elastic components ( P Ac ) is generated by the rotative motions of the motors ( q m ) and the variable length of the springs ( n S ). The two subsystems are connected in a single transmission point ( P Tr ).
In order to measure the actuator system motion ( P Ac ), the UHP rehabilitation robot has two optical rotative encoders (HEDS-5500 of Avago technologies, 1024 p u l s e s / r e v o l u t i o n resolution) and two linear potentiometers (TEX 0150 of NovoTechnik, 0 , 01 m resolution). The variable length of the springs ( n S ) is measured with linear potentiometers, while the encoders are used to measure the rotation angle of the motors ( q m ) (Figure 2).

3. Motion and Force Virtual Sensors

The aim of the virtual sensors is to provide an accurate feedback of the robot/patient interaction, this is, to provide estimations of the force ( F Cn ) and motion ( P Cn ) of the contact point between the robot and the patient. For that purpose, the proposed approach is based on the mathematical model of the UHP rehabilitation robot and the motion measurements provided by the SEA drive system of the robot ( P Ac ).

3.1. Motion Virtual Sensor of the UHP Rehabilitation Robot

The development of the Motion Virtual Sensor for the UHP rehabilitation robot is divided into two steps. First, the kinematical model of the drive system based on elastic components is analysed to determine the motion of the transmission point P Tr ; second, the user/robot contact point motion P Cn is estimated based on the kinematical model of the pantograph and the motion of the transmission point P Tr .

3.1.1. Kinematical Analysis of the Drive System Based on Elastic Components

As shown in Figure 2 and Figure 3, the UHP presents a single transmission point ( P Tr ) where 3 degrees-of-freedom motions can be transmitted bilaterally among the two subsystems.
P Tr = x T r y T r z T r
The motion of the transmission point ( P Tr ) depends directly on the motion generated by two SEA drive systems ( Ac 1 and Ac 2 ),
P Tr = f ( P A c 1 , P A c 2 )
where P A c 1 is the distance from point P m 1 to the transmission point ( P Tr ) and P A c 2 is the distance between P m 2 and P Tr (Figure 3).
From the kinematic model of the SEA drive system [26], it is obtained by,
P A c 1 = l 1 + n S A q m 1 r p P A c 2 = l 1 + n S B q m 2 r p
where l 1 is the distance from the equilibrium point of the transmission point ( P 0 = P Tr when x T r = y T r = z T r = 0 m) to the motors m 1 and m 2 , n S A and n S B are the variable lengths of springs S A and S B , q m 1 and q m 2 are the rotation angles of the motors m 1 and m 2 , and r p is the radius of pulleys.
On the other hand, by applying trigonometry (Figure 3), the following relationships are calculated,
P A c 1 = ( x T r + l 1 ) 2 + y T r 2 + z T r 2 P A c 2 = x T r 2 + ( y T r l 1 ) 2 + z T r 2 x T r 2 + y T r 2 + ( l 3 z T r ) 2 = l 2 2
where l 2 is the distance from the transmission point ( P Tr ) to joint P A of the pantograph structure.
Combining Equations (3) and (4), the expression that defines the motion of the transmission point ( P Tr ) based on the the variables measured by the actuator sensors ( q m 1 , q m 2 , n S A and n S B ) is obtained by,
( x T r + l 1 ) 2 + y T r 2 + z T r 2 = n S A q m 1 r p + l 1 x T r 2 + ( y T r l 1 ) 2 + z T r 2 = n S B q m 2 r p + l 1 x T r 2 + y T r 2 + ( l 2 z T r ) 2 = l 2 2

3.1.2. Kinematical Analysis of the Pantograph

In order to estimate the motion of the contact point ( P Cn = [ x C n y C n z C n ] T ) based on the motion of the transmission point ( P Tr = [ x T r y T r z T r ] T ), the closure loop equation associated to the pantograph structure that relates both parameters has to be defined as,
Γ ( P Tr , P Cn ) = 0
Equation (6) depends on the mechanical configuration of the pantograph structure, this is, on the particular operating mode (ARM or WRIST) selected (Figure 4).
ARM operating mode
In ARM operating mode, joint P B is locked. Hence, the closure loop equation is (Figure 4),
P Tr + l 2 + l 3 + d 1 + l 4 P Cn = 0
solving for P Cn = x C n y C n z C n T ,
x C n y C n z C n = x T r y T r z T r + l 2 + l 3 + d 1 + l 4 l 2 x T r y T r l 2 z T r = l 3 d 1 l 4 l 2 x T r y T r z T r + 0 0 l 2 + l 3 + d 1 + l 4
WRIST operating mode
On the other hand, in WRIST operating mode, joint P B is unlocked and P D and P E are locked. Hence, the closure loop equation can be calculated as (Figure 4),
l 3 + d 1 + l 5 + l 6 l 7 l 8 = 0
where l 3 + d 1 depends on the motion of the transmission point ( P Tr ) and l 6 depends on the motion of the contact point ( P Cn ),
l 3 + d 1 l 2 x T r y T r l 3 z T r + l 5 l 5 l 4 x C n l 6 l 8 y C n l 2 + l 7 z C n 0 l 8 l 6 l 7 = 0 0 0
Solving for P Cn = x C n y C n z C n T ,
x C n y C n z C n = l 3 + d 1 ( l 5 l 4 ) l 2 l 5 x T r y T r z T r + 0 0 l 2 + l 7 + ( l 5 l 4 ) ( l 3 + d 1 l 7 ) l 5
Therefore, the Motion Virtual Sensor of the UHP rehabilitation robot is defined by Equations (5) and (8) for ARM operating mode and Equation (11) in WRIST operating mode. These allow to estimate the motion of the robot/patient contact point P Cn in terms of the sensorized variables of the drive system ( q m 1 , q m 2 , n S A and n S B ): the rotation of the actuators, using encoders, and the length of the elastic springs, using linear potentiometers.

3.2. Force Virtual Sensor of the UHP Rehabilitation Robot

Similarly to the derivation of the Motion Virtual Sensor, the Force Virtual Sensor of the UHP is calculated in two steps: first, the dynamical model of the elastic drive system is used to estimate forces/torques generated by the actuators in the transmission point ( F Tr ); second, the relationship between the forces of the transmission point ( F Tr ) and the forces exerted at the robot/user contact point ( F Cn ) are estimated based on the dynamical model of the robotic device and the motion of transmission point ( P Tr ).

3.2.1. Dynamical Analysis of the Drive System Based on Elastic Components

As shown in Figure 3, the transmission force ( F Tr ) of the UHP rehabilitation robot is the sum of the forces exerted by the four springs (A,B,C,D) that are connected to the transmission point ( P Tr ),
F Tr = F AC 1 + F AC 2 = F S A + F S B + F S C + F S D
where each spring force ( F S i ) is calculated using Hooke’s law based on its variable length ( n S i ), its stifness constant ( K S i ) and its unitary directional vector ( v i ) [26],
F S i = k S i n S i v i
where v i depends on the current position of the transmission point ( P Tr ), which has been defined in Equation (5).

3.2.2. Dynamical Analysis of the Pantograph

Once the transmission force is estimated ( F Tr ), the dynamical model of the pantograph is used to estimate the robot/user contact force ( F Cn ). Although different formulations can be applied, in this work the Lagrangian formulation is selected [29], as it allows to calculate the dynamic model of the rehabilitation robot without considering the internal forces of the mechanism, resulting in a compact dynamic model appropriate to be implemented in advanced controllers.
The development of the dynamical model using the Lagrangian formulation requires of the calculation of both kinetic ( K ) and potential ( U ) energies of the pantograph in ARM and WRIST operating modes, whose substraction defines the Lagrangian ( L ).
In ARM operating mode (Figure 4), the five elements of the pantograph ( E 1 , E 2 , E 3 , E 4 and E 5 ) move [26]. Therefore, the Lagrangian function ( L ARM ) can be calculated as,
L ARM = K ARM U ARM = K E 1 + K E 2 + K E 3 + K E 4 + K E 5 ( U E 1 + U E 2 + U E 3 + U E 4 + U E 5 )
In WRIST operating mode (Figure 4) element E 5 is locked, so its energy is be zero. In addition, element E 4 can only rotate along its axis, so its potential energy will be zero [27].
L WRIST = K WRIST U WRIST = K E 1 + K E 2 + K E 3 + K E 4 ( U E 1 + U E 2 + U E 3 )
Note that in both cases K E i represents the kinetic energy and U E i the potential energy of element E i ,
K E i = 1 2 ω CM E i T I E i ω CM E i + 1 2 m E i v CM E i T v CM E i
U E i = m E i g h C M E i
where I E i is the inertia of element E i , m E i is its mass, h C M E i the z coordinate of the position of its center of mass, and v CM E i and ω CM E i the linear and the angular velocity of the center of mass.
Therefore, applying the Lagrangian formulation, the relationship between the contact force ( F Tr ) and the transmission force ( F Cn ) in terms of the motion of the transmission points ( P Tr ) and the motion of the contact point ( P Cn ) is obtained.
d d t L P ˙ Tr L P Tr = λ i Γ ( P Tr , P Cn ) P Tr + F Tr
d d t L P ˙ Cn L P Cn = λ i Γ ( P Tr , P Cn ) P Cn + F Cn
where, λ i is the set of Lagrange multipliers and Γ ( P Tr , P Cn ) is the closure loop equation that relates P Tr and P Cn , as defined in Section 3.1.2.
Combining this relationship with Equation (12) the Force Virtual Sensor of the UHP rehabilitation robot is obtained, which estimates robot/patient contact force F Cn in terms of the sensorized variables of the drive system ( q m 1 , q m 2 , n S A and n S B ).

4. Validation and Results

In order to validate the developed Virtual Sensors, they were integrated into an advanced position/force controller designed for the UHP rehabilitation robot and their performance was evaluated by performing active and passive rehabilitation exercises. The integrated controller is composed by a set of controllers that operate at two levels. In the low level, two controllers operate in parallel: a kinematical model-based position control and a dynamical model-based force control. The high level controller, which defines the rehabilitation task (active or passive), is used to select the force or position controller and their reference. The overall structure of the controller is summarized in Figure 5.
As shown in Figure 5, the low level position and force controllers require contact point motion ( P Cn ) and contact force ( F Cn ) measurements respectively to operate properly. Hence, the Virtual Sensors were integrated in the control loop to estimate these variables instead of measuring them using high accuracy sensors.
To evaluate the applicability and reliability of the Virtual Sensor approach, the control performance of the robot based on the proposed Virtual Sensors has been compared with that using actual high accuracy sensors in different experimental tests. For that purpose, a three step procedure has been followed for the validation, as shown in Figure 6: (1) Each experimental test has been carried out with the rehabilitation robot, where the required low level controllers have been fed with the data provided by high accuracy sensors that measure P Cn and F Cn ; (2) the same test has been performed using controllers whose data feedback data is provided by the proposed Virtual Sensors; (3) the results of both approaches have been compared . In the first step, A MINI40 force/torque sensor (ATI, 6 D O F , 1 / 100 N resolution) has been used to measure F Cn , while the precise YNGS1 inclinometer (Sensor-Technik Wiedemann GmbH, 3 D O F , 0.25°/s resolution) has been adopted for the measurement of P Cn .
Three experimental tests have been designed to evaluate both Virtual Sensors in ARM and WRIST operating modes following the aforementioned procedure: in the first test, in order to validate the Motion Virtual Sensor, the position controller has been used, while in the last two tests, the force controller has been used to validate the Force Virtual Sensor. Results for each test are detailed next.

4.1. Motion Virtual Sensor Validation

The goal of the first test is to evaluate the performance of the position controller using the Motion Virtual Sensor. For this purpose, a sinusoidal reference has been designed for both x and y directional movement. The frequency varied from 0.05 Hz to 0.2 Hz as a function of time. The magnitude of the reference has been set to 0.1 m for ARM mode and 0.05 m for WRIST mode.
Figure 7 and Figure 8 show the results of the contact point P Cn displacement for ARM and WRIST operating modes, respectively. In the upper plot, the x coordinate of P Cn is shown, while the lower one represents the y coordinate. In addition, the blue line illustrates the reference motion of the contact point ( P Cn Des ), while the red line is associated to the case using additional sensors, and the green one to the case in which the proposed Motion Virtual Sensor is used. Table 1 summarizes mean and maximum distance error | P Cn Des P Cn | for both cases.

4.2. Force Virtual Sensor Validation

In the following two tests, the performance of the force controller with the Virtual Sensor compared to the performance with a high accuracy sensor are analysed. The first test simulates the active mode of the robot, where the desired contact force has been set to zero ( F Cn Des = 0 N ). In the second one, a sinusoidal force reference with 20 N magnitude and variable frequencies of 0.05 Hz, 0.1 Hz and 0.2 Hz has been selected.
For all the figures in this section, the upper plot depicts the evolution of the x coordinate of the contact force ( F Cn ) while the lower one shows F Cn in y coordinate. In addition, the blue line defines the desired contact force ( F Cn Des ), the red line is related to the contact force provided by the controller with real sensors ( F Cn Sensor ), while the green one corresponds to the case in which the Virtual Sensor ( F Cn VirtualSensor ) is used.
Test 1: Zero force reference
Figure 9 and Figure 10 show the control results of zero force reference mode in ARM and WRIST operating modes. Table 2 summarizes the mean and maximum error of the force controller when the contact force is measured with the force sensor or estimated with the Force Virtual Sensor ( | F Cn Des F Cn | ).
Test 2: Variable force reference
Figure 11 and Figure 12 show the control results of variable force in ARM and WRIST operating modes, respectively. Table 3 summarizes the mean and maximum force errors carried out when considering sensor data or the proposed Virtual Sensor feedback ( | F Cn Des F Cn | ).

4.3. Suitability of the Developed Virtual Sensors for Advanced Controllers in Rehabilitation Robotics

In this subsection results from the previously detailed tests will be discussed, and conclusions drawn. The results of the first validation test show that the performance of the position controller using Virtual Sensors is similar to the one with real sensors. According to the literature [30], healthy people are able to perform movements of the upper extremities with an average resolution of 0.005 m. Therefore, it can be stated that the obtained error difference (0.005 m in ARM and 0.0001 m in WRIST operating mode) between virtual and real sensor is acceptable for rehabilitation therapies.
On the other hand, results of the second validation test show that in ARM operating mode, the performance of the force controller in active rehabilitation tasks (zero force reference) is very similar when considering actual sensor data, or the proposed Virtual Force Sensor. The existing estimations errors are acceptable for rehabilitation purposes ( Δ of force error < 0.6 N) due to the sensitivity limit of humans. The performance of the controller when using Virtual Sensors is even better than the use of real sensor data in WRIST mode due to the mechanical configuration of the robot and its influence on the force sensor ( Δ of the mean error = 0.07 N and Δ of the maximum error = 1.15 N). It is also important to note that the tests have been carried by a healthy person, which has moved arbitrarily and randomly the robot in this case. This way, the misaligment of peaks in Figure 9 and Figure 10 are due to being different executions of the same test.
In the proposed third validation test, in which variable force references are tested, the controller has been able to follow the desired force reference with small and acceptable errors for both cases. The difference between the two implementation (sensor and Virtual Sensor) is almost negligible for the user ( Δ of the average error is 1.5 N in ARM and 0.85 N in WRIST operating mode). In addition, it should be noted that these tests have been carried out by a healthy person, who can exert much more force than a stroke patient. Hence, less error is expected for these cases.
It is important to emphasize that in a rehabilitation application, safety and robustness of the device are more critical than the precision of the movements [31,32]. For these applications, fast controllers capable of responding quickly to the movement generated by the patient are required. Therefore, the use of compact Virtual Sensors like ones proposed are preferred to complex and precise Virtual Sensors that need more computational cost and slow down the speed of the controller.
Therefore, after performing different validation tests with the UHP rehabilitation robot in ARM and WRIST operating modes, it is concluded that the developed Force and Motion Virtual Sensors are suitable for their use in advanced controllers for rehabilitation robots.

5. Conclusions

Robotic devices can improve the rehabilitation process of patients with reduced mobility after a neurological disorder, such as a stroke. One of the key components for successful robot-mediated therapy is to design advanced controllers that can adjust interactive force and motion between the patient and the robot as needed. Force and motion values are usually measured through accurate sensors, which increases the cost and complexity of the device.
In this work, a novel virtual sensor approach is proposed for the estimation of the contact force and motion in rehabilitation robots. The developed Virtual Sensors are based on the mathematical model of the robotic device and the measurement provided by low cost position sensors, such as optical encoders or linear potentiometers.
The proposed Virtual Sensors have been implemented and evaluated in an advanced force/position controller of the UHP rehabilitation robot for upper limb training. Control performance based on Virtual Sensors has been compared to that using real sensors in performing passive and active rehabilitation exercises with a healthy subject through three different experimental conditions.
Results shows that the use of the Virtual Sensors produces performance similar to that obtained with real sensors in terms of motion and force errors (less than 0.005 m of mean motion error and 1.5 N of mean force error), which is acceptable for rehabilitation use. While the results reveal that Virtual Sensors can be used as an alternative of real sensors, further experiments with actual patients who have different level of impairments will be future interesting topic in order to ensure clinical adaptation and acceptance of the developed Virtual Sensors.

Acknowledgments

This work was supported in part by the Basque Country Governments (GV/EJ) under grant PRE-2014-1-152, UPV/EHU’s PPG17/56 project, Basque Country Governments IT914-16 project, Spanish Ministry of Economy and Competitiveness’ MINECO & FEDER inside DPI2017-82694-R project, Euskampus, FIK and Spanish Ministry of Science and Innovation PDI-020100-2009-21 project.

Author Contributions

A. Mancisidor has been in charge of the development of the Virtual Sensor approach and performing the tests. A. Zubizarreta and I. Cabanes, have supervised and coordinated the research work, participating in the design of the approach. E. Portillo and J.H. Jung have provided technical advice related with the UHP development and testing. All authors have participated in the organization of the paper and the approval of the final manuscript.

Abbreviations

The following abbreviations are used in this manuscript:
SEASerial Elastic Actuator
UHPUniversal Haptic Pantograph

References

  1. Van Eeden, M.; Van Heugten, C.M.; Evers, S.M. The economic impact of stroke in The Netherlands. BMC Public Health 2012, 12, 122. [Google Scholar] [CrossRef] [PubMed]
  2. Feigin, V.L.; Forouzanfar, M.H.; Krishnamurthi, R.; Mensah, G.A.; Connor, M.; Bennett, D.A.; Moran, A.E.; Sacco, R.L.; Anderson, L.; Truelsen, T.; et al. Global and regional burden of stroke during 1990–2010: Findings from the Global Burden of Disease Study 2010. Lancet 2014, 383, 245–255. [Google Scholar] [CrossRef]
  3. Pons, J.L.; Raya, R.; González, J. Emerging Therapies in Neurorehabilitation II. Biosyst. Biorobot. 2016, 10, 29–64. [Google Scholar]
  4. Xie, S. Advanced Robotics for Medical Rehabilitation. Springer Tracts Adv. Robot. 2016, 108, 1–357. [Google Scholar]
  5. Bai, J.; Song, A.; Xu, B.; Nie, J.; Li, H. A Novel Human-Robot Cooperative Method for Upper Extremity Rehabilitation. Int. J. Soc. Robot. 2017, 9, 265–275. [Google Scholar] [CrossRef]
  6. Sarac, M.; Solazzi, M.; Leonardis, D.; Sotgiu, E.; Bergamasco, M.; Frisoli, A. On the Use of Cable-Driven Robots in Early Inpatient Stroke Rehabilitation. Adv. Italian Mech. Sci. 2017, 47, 551–558. [Google Scholar]
  7. Yang, X.; She, H.; Lu, H.; Fukuda, T.; Shen, Y. State of the Art: Bipedal Robots for Lower Limb Rehabilitation. Appl. Sci. 2017, 7, 1182. [Google Scholar] [CrossRef]
  8. Zi, B.; Yin, G.; Zhang, D. Design and Optimization of a Hybrid-Driven Waist Rehabilitation Robot. Sensors 2016, 16, 2121. [Google Scholar] [CrossRef] [PubMed]
  9. Meng, W.; Liu, Q.; Zhou, Z.; Ai, Q.; Sheng, B.; Xie, S.S. Recent development of mechanisms and control strategies for robot-assisted lower limb rehabilitation. Mechatronics 2015, 31, 132–145. [Google Scholar] [CrossRef]
  10. Zhou, S.H.; Fong, J.; Crocher, V.; Tan, Y.; Oetomo, D.; Mareels, I. Learning control in robot-assisted rehabilitation of motor skills—A review. J. Control Decis. 2016, 7706, 1–25. [Google Scholar] [CrossRef]
  11. Lo, C.L.; Tseng, H.T. Predicting rehabilitation treatment helpfulness to stroke patients: A supervised learning approach. Artif. Intell. Res. 2017, 6, 1. [Google Scholar] [CrossRef]
  12. Proietti, T.; Crocher, V.; Roby-Brami, A.; Jarrasse, N. Upper-limb robotic exoskeletons for neurorehabilitation: A review on control strategies. Biomed. Eng. 2016, 9, 4–14. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, S.; Guo, S.; Gao, B.; Hirata, H.; Ishihara, H. Design of a Novel Telerehabilitation System with a Force-Sensing Mechanism. Sensors 2015, 15, 11511–11527. [Google Scholar] [CrossRef] [PubMed]
  14. Grosu, V.; Grosu, S.; Vanderborght, B.; Lefeber, D.; Rodriguez-Guerrero, C. Multi-axis force sensor for human-robot interaction sensing in a rehabilitation robotic device. Sensors 2017, 17, 1294. [Google Scholar] [CrossRef] [PubMed]
  15. Ochoa Luna, C.; Rahman, M.H.; Saad, M.; Archambault, P.; Zhu, W.H. Virtual decomposition control of an exoskeleton robot arm. Robotica 2016, 34, 1587–1609. [Google Scholar] [CrossRef]
  16. Li, H.Y.; Chien, L.Y.; Hong, H.Y.; Pan, S.H.; Chiao, C.L.; Chen, H.W.; Fu, L.C.; Lai, J.S. Active control with force sensor and shoulder circumduction implemented on exoskeleton robot NTUH-II. In Proceedings of the International Conference on Intelligent Robots and Systems, Daejeon, Korea, 9–14 October 2016; pp. 2406–2411. [Google Scholar]
  17. Song, A.; Pan, L.; Xu, G.; Li, H. Adaptive motion control of arm rehabilitation robot based on impedance identification. Robotica 2014, 33, 1–18. [Google Scholar] [CrossRef]
  18. Etedali, S.; Talebi, H.A.; Mohammadi, A.D. A robust force observer for robot manipulators subjected to external disturbance. In Proceedings of the International Conference on Robotics and Mechatronics, Tehran, Iran, 7–9 October 2015; pp. 539–544. [Google Scholar]
  19. Yao, B.; Zhou, Z.; Wang, L.; Xu, W.; Liu, Q.; Liu, A. Sensorless and adaptive admittance control of industrial robot in physical human-robot interaction. Robot. Comput.-Integr. Manuf. 2018, 51, 158–168. [Google Scholar] [CrossRef]
  20. Zhao, B.; Nelson, C.A. Sensorless Force Sensing for Minimally Invasive Surgery. J. Med. Devices 2015, 9, 012–041. [Google Scholar] [CrossRef] [PubMed]
  21. De Luca, A.; Mattone, R. Sensorless Robot Collision Detection and Hybrid Force/Motion Control. IEEE Int. Conf. Robot. Autom. 2005, 2005, 999–1004. [Google Scholar]
  22. Aviles, A.I.; Alsaleh, S.; Sobrevilla, P.; Casals, A. Sensorless force estimation using a neuro-vision-based approach for robotic-assisted surgery. In Proceedings of the International IEEE/EMBS Conference on Neural Engineering (NER), Montpellier, France, 22–24 April 2015; pp. 86–89. [Google Scholar]
  23. Choi, J.H.; Kwak, J.H.; An, J.; Oh, S. Force Sensorless Multi-functional Impedance Control for Rehabilitation Robot. IFAC-PapersOnLine 2017, 50, 12077–12082. [Google Scholar] [CrossRef]
  24. Mancisidor, A.; Zubizarreta, A.; Cabanes, I.; Bengoa, P.; Jung, J.H. Multifunctional robotic device for the rehabilitation of the upper limbs. Rev. Iberoam. Autom. Inform. Ind. 2018, 15. Available online: https://polipapers.upv.es/index.php/RIAI/article/view/8820 (accessed on 5 March 2018).
  25. Perry, J.C.; Oblak, J.; Jung, J.H.; Cikajlo, I.; Veneman, J.F.; Goljar, N.; Bizoviar, N.; Matjai, Z.; Keller, T. Variable structure pantograph mechanism with spring suspension system for comprehensive upper-limb haptic movement training. J. Rehabil. Res. Dev. 2011, 48, 317–334. [Google Scholar] [CrossRef] [PubMed]
  26. Mancisidor, A.; Zubizarreta, A.; Cabanes, I.; Bengoa, P.; Jung, J.H. Kinematical and dynamical modeling of a multipurpose upper limbs rehabilitation robot. Robot. Comput.-Integr. Manuf. 2018, 49, 374–387. [Google Scholar] [CrossRef]
  27. Mancisidor, A.; Zubizarreta, A.; Cabanes, I.; Bengoa, P.; Jung, J.H. Kinematic and Dynamic Modeling of a Multifunctional Rehabilitation Robot UHP. In New Trends in Medical and Service Robots. Design, Analysis and Control; Husty, M., Hofbaur, M., Eds.; Springer: Berlin, Germany, 2018; Volume 48, pp. 117–130. [Google Scholar]
  28. Pratt, G.; Williamson, M. Series elastic actuators. In Proceedings of the International Conference on Intelligent Robots and Systems. Human Robot Interaction and Cooperative Robots, Pittsburgh, PA, USA, 5–9 August 1995; pp. 399–406. [Google Scholar]
  29. Chen, H.; Liu, Y. Robotic assembly automation using robust compliant control. Robot. Comput.-Integr. Manuf. 2013, 29, 293–300. [Google Scholar] [CrossRef]
  30. Germanotta, M.; Vasco, G.; Petrarca, M.; Rossi, S.; Carniel, S.; Bertini, E.; Cappa, P.; Castelli, E. Robotic and clinical evaluation of upper limb motor performance in patients with Friedreich Ataxia: an observational study. J. NeuroEng. Rehabilit. 2015, 12, 1–13. [Google Scholar] [CrossRef] [PubMed]
  31. Virk, G.S.; Cameron, S. ISO—IEC standardization efforts in robotics. In Proceedings of the Standardized Knowledge Representation and Ontologies for Robotics and Automation, Chicago, IL, USA, 18 September 2014. [Google Scholar]
  32. Zhang, F.; Hua, L.; Fu, Y.; Chen, H.; Wang, S. Design and development of a hand exoskeleton for rehabilitation of hand injuries. Mech. Mach. Theory 2014, 73, 103–116. [Google Scholar] [CrossRef]
Figure 1. ARM and WRIST operating modes of the UHP rehabilitation robot (Red color denotes locked joint, Green color denotes unlocked joint).
Figure 1. ARM and WRIST operating modes of the UHP rehabilitation robot (Red color denotes locked joint, Green color denotes unlocked joint).
Sensors 18 00785 g001
Figure 2. Two main subsystems of the UHP rehabilitation robot.
Figure 2. Two main subsystems of the UHP rehabilitation robot.
Sensors 18 00785 g002
Figure 3. Drive system of the UHP rehabilitation robot.
Figure 3. Drive system of the UHP rehabilitation robot.
Sensors 18 00785 g003
Figure 4. Pantograph structure of the UHP rehabilitation robot.
Figure 4. Pantograph structure of the UHP rehabilitation robot.
Sensors 18 00785 g004
Figure 5. Advanced Controller for the UHP rehabilitation robot with Motion and Force Virtual Sensors.
Figure 5. Advanced Controller for the UHP rehabilitation robot with Motion and Force Virtual Sensors.
Sensors 18 00785 g005
Figure 6. Procedure to evaluate the applicability and reliability of the developed Virtual Sensors.
Figure 6. Procedure to evaluate the applicability and reliability of the developed Virtual Sensors.
Sensors 18 00785 g006
Figure 7. Validation of the Motion Virtual Sensor in ARM operating mode.
Figure 7. Validation of the Motion Virtual Sensor in ARM operating mode.
Sensors 18 00785 g007
Figure 8. Validation of the Motion Virtual Sensor in WRIST operating mode.
Figure 8. Validation of the Motion Virtual Sensor in WRIST operating mode.
Sensors 18 00785 g008
Figure 9. Validation of the Force Virtual Sensor in the ARM operating mode with zero force reference.
Figure 9. Validation of the Force Virtual Sensor in the ARM operating mode with zero force reference.
Sensors 18 00785 g009
Figure 10. Validation of the Force Virtual Sensor in the WRIST operating mode with zero force reference.
Figure 10. Validation of the Force Virtual Sensor in the WRIST operating mode with zero force reference.
Sensors 18 00785 g010
Figure 11. Validation of the Force Virtual Sensor in the ARM operating mode with variable force reference.
Figure 11. Validation of the Force Virtual Sensor in the ARM operating mode with variable force reference.
Sensors 18 00785 g011
Figure 12. Validation of the Force Virtual Sensor in the WRIST operating mode with variable force reference.
Figure 12. Validation of the Force Virtual Sensor in the WRIST operating mode with variable force reference.
Sensors 18 00785 g012
Table 1. Mean and maximum motion errors of the position controller.
Table 1. Mean and maximum motion errors of the position controller.
OperatingWith the SensorWith the Virtual Sensor Δ of the Error
ModeMeanMaximumMeanMaximumMeanMaximum
ARM 0.015 m 0.098 m 0.02 m 0.141 m 0.005 m 0.043 m
WRIST 0.0022 m 0.021 m 0.0021 m 0.015 m 0.0001 m 0.007 m
Table 2. Mean and maximum errors of the force controller with zero force reference.
Table 2. Mean and maximum errors of the force controller with zero force reference.
OperatingWith the SensorWith the Virtual Sensor Δ of the Error
ModeMeanMaximumMeanMaximumMeanMaximum
ARM 0.88 N 4.38 N 1.42 N 4.87 N 0.54 N 0.49 N
WRIST 1.45 N 7.72 N 1.38 N 6.57 N 0.07 N 1.15 N
Table 3. Mean and maximum errors of the force controller with variable force reference.
Table 3. Mean and maximum errors of the force controller with variable force reference.
OperatingWith the SensorWith the Virtual Sensor Δ of the Error
ModeMeanMaximumMeanMaximumMeanMaximum
ARM 2.89 N 14.72 N 4.34 N 14.61 N 1.45 N 0.11 N
WRIST 2.92 N 14.10 N 3.77 N 15.20 N 0.85 N 1.10 N

Share and Cite

MDPI and ACS Style

Mancisidor, A.; Zubizarreta, A.; Cabanes, I.; Portillo, E.; Jung, J.H. Virtual Sensors for Advanced Controllers in Rehabilitation Robotics. Sensors 2018, 18, 785. https://doi.org/10.3390/s18030785

AMA Style

Mancisidor A, Zubizarreta A, Cabanes I, Portillo E, Jung JH. Virtual Sensors for Advanced Controllers in Rehabilitation Robotics. Sensors. 2018; 18(3):785. https://doi.org/10.3390/s18030785

Chicago/Turabian Style

Mancisidor, Aitziber, Asier Zubizarreta, Itziar Cabanes, Eva Portillo, and Je Hyung Jung. 2018. "Virtual Sensors for Advanced Controllers in Rehabilitation Robotics" Sensors 18, no. 3: 785. https://doi.org/10.3390/s18030785

APA Style

Mancisidor, A., Zubizarreta, A., Cabanes, I., Portillo, E., & Jung, J. H. (2018). Virtual Sensors for Advanced Controllers in Rehabilitation Robotics. Sensors, 18(3), 785. https://doi.org/10.3390/s18030785

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop