Next Article in Journal
Research on the Deviation Correction Control of a Tracked Drilling and Anchoring Robot in a Tunnel Environment
Previous Article in Journal
Calibration Optimization of Kinematics and Dynamics for Delta Robot Driven by Integrated Joints in Machining Task
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impedance Learning-Based Hybrid Adaptive Control of Upper Limb Rehabilitation Robots

1
Institute of Artificial Intelligence, Donghua University, Shanghai 201620, China
2
Shanghai Engineering Research Center of Industrial Big Data and Intelligent System, Shanghai 201620, China
3
School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
4
The Second Research Institute of China Electronics Technology Group Corporation, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Actuators 2024, 13(6), 220; https://doi.org/10.3390/act13060220
Submission received: 15 May 2024 / Revised: 6 June 2024 / Accepted: 10 June 2024 / Published: 12 June 2024
(This article belongs to the Section Actuators for Medical Instruments)

Abstract

:
This paper presents a hybrid adaptive control strategy for upper limb rehabilitation robots using impedance learning. The hybrid adaptation consists of a differential updating mechanism for the estimation of robotic modeling uncertainties and periodic adaptations for the online learning of time-varying impedance. The proposed hybrid adaptive controller guarantees asymptotical control stability and achieves variable impedance regulation for robots without interaction force measurements. According to Lyapunov’s theory, we proved that the proposed impedance learning controller guarantees the convergence of tracking errors and ensures the boundedness of the estimation errors of robotic uncertainties and impedance profiles. Simulations and experiments conducted on a parallel robot validated the effectiveness and the superiority of the proposed impedance learning controller in robot-assisted rehabilitation. The proposed hybrid adaptive control has potential applications in rehabilitation, exoskeletons, and some other repetitive interactive tasks.

1. Introduction

Comfort and safety are of significant value in robot-assisted rehabilitation and can be improved through robot impedance regulation. As a well-known compliance control approach, impedance control has shown its powerful abilities in robot impedance regulation and has been extensively applied in service robots and industrial robots [1,2,3,4]. However, robot modeling uncertainties and human impedance variation bring in difficulties for impedance control implementation in rehabilitation training.
Robot modeling uncertainties affect the control stability and the control robustness of impedance control. Up to now, considerable research attention has been paid to robust impedance control [5], adaptive impedance control [6,7], neural impedance control [8,9,10,11], and iterative learning impedance control [12]. These impedance control methods improve the stability and robustness of closed-loop control systems; however, these methods mainly focus on constant impedance regulation.
In robot-assisted rehabilitation, human impedance varies along with the variation in interaction forces and tasks. Time-varying impedance control [13] has better robustness and compliance in rehabilitation training through robotic variable impedance regulation, in comparison with constant impedance control [14,15,16]. Although categories of variable impedance control methods were designed for robot-assisted rehabilitation [17], many of these results neglected the stability of the desired variable impedance dynamics (DVID) which is of significant value for human–robot cooperative motion. Actually, the DVID are described by linear time-varying systems and may be violated by the variation in impedance profiles. In [18], the stability region of impedance parameters was evaluated experimentally for the DVID with zero stiffness. In [19], an adaptive variable impedance controller was designed for force tracking in interaction with an unknown environment with constant impedance. In [20,21,22], novel constraints on time-varying impedance profiles were proposed to guarantee the stability of the DVID, but how to design impedance profiles in robot-assisted rehabilitation was not given in these results.
Designing appropriate impedance profiles to balance between the desired control stability and the desirable interaction performance is of significant value for robot-assisted rehabilitation. For impedance profile construction, the algorithms of Learning from Demonstration and sEMG-based Dynamic Movement Primitives (DMPs) were applied for time-varying impedance control in physical HRI. In [23,24,25], two-loop-based optimal admittance controllers were proposed for robot-assisted rehabilitation. An inner-control loop makes robots behave in a prescribed impedance model, and an outer-loop optimal control aims at finding the optimal parameters of the prescribed impedance model by minimizing a performance function of interaction forces and tracking errors. Almost all impedance/admittance controllers mentioned above were designed using interactive forces; however, the force measurements may become impossible without related force sensors or ineffective due to measurement delay or noises.
From the above analysis, the variable impedance regulation of robots can improve human–robot interactive performances, but impedance variation brings in control stability problems. Recently, model-based impedance learning controllers were proposed in [26,27,28] for human–robot interaction with a control stability guarantee and without the requirement of interaction forces. Specifically, human impedance uncertainties were estimated by using the periodicity characteristics of task reference trajectory and human impedance profiles in [26,27], and an adaptive impedance learning controller was proposed in [28], where robot modeling uncertainties and human impedance uncertainties were estimated by using traditional differential updating laws. However, the results in [26,27,28] have the following deficiencies: (i) these adaptive impedance learning controllers only guarantee the closed-loop control systems be uniformly ultimately bounded (UUB) but not be asymptotically stable; (ii) precise knowledge about robotic dynamic parameters is required in [26,27]; and (iii) the result in [28] requires the unknown impedance profiles be slowly varying due to the weakness of differential adaptation in estimating time-varying signals.
In this paper, we propose an impedance learning-based hybrid adaptive control strategy with potential applications in repetitive rehabilitation training. In comparison with the related literature, the innovations and contributions of this paper are stated as follows:
(1) A differential updating mechanism is designed for estimating robotic parameter uncertainties, while periodic adaptive learning laws are designed for the learning of human time-varying impedance profiles.
(2) The projection function used in the adaptive law and the saturation functions in the periodic adaptive laws avoid the possible drift of estimators.
(3) The proposed hybrid adaptive robot controller guarantees asymptotical control stability and achieves variable impedance regulation for rehabilitation robots without interaction force measurements.
(4) The asymptotical convergence of tracking errors and the finiteness of the estimation errors of the robot parameters and the human impedance profiles are validated through a Lyapunov-like analysis.
Compared with the constant impedance control, the proposed impedance learning-based controller can provide time-varying impedance regulation for rehabilitation robots. In comparison with the impedance learning controllers in [13,18,19,20,21,22], the proposed model-based impedance learning controller can guarantee the control stability without using interaction force signals. Compared with the model-based impedance learning controllers in [26,27,28], the proposed controller guarantees the asymptotical convergence of tracking errors. In addition, robotic parameter uncertainties are considered and estimated in this paper in comparison with [26,27], and the impedance profiles are not required to be slowly varying in comparison with [28]. A summary of currently used adaptive controllers with impedance regulation and the gap where this paper is located are listed in Table 1.

2. Problem Formulation

Consider the robot dynamics with the following form
M ( q ) q ¨ + C ( q , q ˙ )   q ˙ + G ( q ) + F q ˙ = τ + τ h
where q     R n   denotes the vector of joint angles, M ( q ) the inertial matrix, C ( q ,   q ˙ ) the Coriolis and centrifugal matrix, G ( q ) the gravity torque, and F q ˙ the viscous force with   F being a constant matrix. In (1), τ h   and τ   are the interaction force vector in joints and the system control input, respectively.
Denote x as the end-effector Cartesian position. The robot has the following kinematic transformations
x = Δ q ,
x ˙ = J ( q ) q . ˙
In (3),   J ( q ) = d ( q ) / d q is the Jacobian matrix from the joint space to the Cartesian coordinates. Based on (2) and (3), one can convert the joint space dynamics in (1) into the following Cartesian dynamics
M x x x ¨ + C x ( x + x ˙ ) x ˙ + G x x + F x x = u + f h
where M x = J T M J 1 , C x = J T ( C q M J 1 J ˙ ) J 1 , G x = J T G , F x = J T F , f h = J T τ h , and u = J T τ .
Property 1.
M x x    is a symmetric positive definite matrix and satisfies
σ 1 I     M x ( x )   σ 2 I
where  σ 1 ,   σ 2 > 0 , and   σ 1  and  σ 2  can be regarded as the minimum and maximum eigenvalues of  M x ( x ) , respectively.
Property 2.
  M ˙ x x 2 C x ( x , x ˙ )  is skew symmetric, i.e.,
ξ T ( M ˙ x x 2 C x x , x ˙ ) ξ = 0 ,   ξ R n .
Property 3.
The left part of (4) can be parameterized in the following form
M x x ϕ 1 + C x x , x ˙ ϕ 2 + G x x = Y ( ϕ 1 , ϕ 2 , x , x ˙ ) θ
where  ϕ 1 , ϕ 2 R n ,   Y ( ϕ 1 , ϕ 2 , x , x ˙ )  is a regression matrix, and θ is a constant vector that contains unknown parameters and satisfies  | | θ | |   θ c .
Property 4.
If  x , x ˙ L , then    M x x , C x x , x ˙ , G x x ,    and their partial derivatives with respect to their respective arguments are bounded. If    ϕ 1 , ϕ 2 , x , x ˙ L ,  Y ( ϕ 1 , ϕ 2 , x , x ˙ ) , and its partial derivatives with respect to its arguments are bounded.
Assumption 1.
Suppose the desired trajectory  x r  and its first and second time derivatives are bounded for the considered robot in the HRI.
Define the trajectory tracking error e as
e = x r x
As proven in [26], the interaction force can be expanded as
f h = f ¯ f t K h t e B h ( t ) e ˙
where   f ¯ f t   denotes the feedforward force of the human, and   K h t   and   B h ( t ) denote the stiffness term and the damping term of the interaction force, respectively. The interaction force f h   can be further expressed as
f h = f f t + K h t x + B h ( t ) x ˙
where   f f t : = f ¯ f t K s * t x r K d * t x ˙ r   denotes a modified feedforward force. In repetitive rehabilitation training, the impedance profiles   f f t , K h t ,   and   B h ( t ) can be assumed to be periodic with   T , i.e.,
f f t + T = f f t , K h t + T = K h t ,
B h t + T = B h t .
It is assumed that f f t ,   K h t and B h ( t ) are continuous functions and satisfy
f f t f * , K h t K * , B h t B * .
where f * , K * , and B * are constant matrices; ‘ ’ in the last equation means each element of the left part of ‘ ’ is no more than the related element of the right part of ‘ ’.
Remark 1.
In [25], the feedforward force, the stiffness term, and the damping term are assumed to be slowly varying and are estimated by using traditional adaptive laws. In this paper,   f f t ,   K h t , and  B h t  are only required to be continuous and bounded functions that will be estimated in Section 3 using periodic adaptation.

3. Hybrid Adaptive Control via Impedance Learning

Define the auxiliary error ϵ as
ϵ = e ˙ + k 1 e
where k 1 > 0 .
From (4), ϵ has the following dynamics
M x ( x ) ϵ ˙ = C x ( x , x ˙ ) ϵ + Y ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) θ   f h u .
It is convenient to design the following control law
u = k 2 ϵ + Y ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) θ ^   f ^ h
where θ ^   and   f ^ h   are estimators of θ and f h , respectively; f ^ h is designed as
f ^ h = f ^ f + K ^ h x + B ^ h x ˙
with f ^ h , K ^ h , and B ^ h   being estimators of f f , K h , and B h , respectively.
Substituting (11) and (12) into (10) yields
M x ( x ) ϵ ˙ = C x ( x , x ˙ ) ϵ k 2 ϵ + Y ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) θ ~   f ~ f K ~ h x B ~ h x ˙ .
where   f ~ f = f f f ^ f , K ~ h = K h K ^ h , and B ~ h = B h B ^ h denote estimation errors.
In (15), θ is a constant vector that can be estimated by the sing traditional differential updating mechanism, while the impedance profiles f f   ( t ) , K h   ( t ) , and B h   t are time-varying signals that may not be slowly varying and cannot be effectively estimated using traditional adaptive laws. By exploiting their periodicity characteristics, the impedance profiles can be estimated by using periodic adaptations. Therefore, we propose a hybrid adaptive law for estimating θ and the impedance profiles.
For   θ ^ , we design the following adaptive update law
θ ^ ˙ = p r o j θ { α 0 Y T ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) ϵ }
where α 0 is the learning rate, and p r o j θ ( ) is a projection function defined as
p r o j θ = i f   θ ^ c θ       o r   | | θ ^ | | = c θ & θ ^ T · 0   θ ^ θ ^ T · | | θ ^ | | 2 i f   θ ^ = c θ & θ ^ T · 0  
with θ c being defined in Property 3.
To estimate the time-varying terms f f ( t ) , K h ( t ) , and B h ( t ) , we design the following periodic learning laws
f ^ f ( t ) = s a t f * ( f ^ f ( t T ) ) α 1 ϵ , t [ T , ) α 10 ( t ) ϵ , t [ 0 , T ]
K ^ h ( t ) = s a t K * ( K ^ h ( t T ) ) α 2 ϵ x T , t [ T , ) α 20 ( t ) ϵ x T , t [ 0 , T ]
B ^ h ( t ) = s a t B * ( B ^ h ( t T ) ) α 3 ϵ x ˙ T , t [ T , ) α 30 ( t ) ϵ x ˙ T , t [ 0 , T ]
where f ^ f t = 0 , K ^ h t = 0 , and B ^ h t = 0 for t < 0 ; α 1 , α 2 , and α 3 are positive gains; α 10 ( t ) , α 20 ( t ) , and α 30 ( t ) are a continuous and strictly increasing function that satisfy α 10 0 = α 20 0 = α 30 0 = 0 , α 10 T = α 1 , α 20 T = α 2 , and α 30 T = α 3 ; s a t f * ( f ^ f ( t T ) ) is defined by
s a t f * ( f ^ f i ( t T ) ) = f ^ f i ( t T ) , i f   | f ^ f i ( t T ) | f i * f i * s g n ( f ^ f i ( t T ) ) , o t h e r w i s e
and   s a t K * ( K ^ h ( t T ) ) , s a t B * ( K ^ h ( t T ) ) are defined similarly to the definition of s a t f * ( f ^ f ( t T ) ) .
Based on (12), (13) and (21)–(23), the estimation errors f ~ f ( t ) : = f f ( t ) f ^ f ( t ) , K ~ h ( t ) : = K h ( t ) K ^ h ( t ) , and   B ~ h ( t ) : = B h ( t ) B ^ h ( t ) , satisfy
f ~ f ( t ) = s a t f * ( f f ( t T ) ) s a t f * ( f ^ f ( t T ) ) + α 1 ϵ , K ~ h ( t ) = s a t K * ( K h ( t T ) ) s a t K * ( K ^ h ( t T ) ) + α 2 ϵ x T , B ~ h ( t ) = s a t B * ( B h ( t T ) ) s a t B * ( B ^ h ( t T ) ) + α 3 ϵ x ˙ T ,
for t T .
Remark 2.
If  α 10 T = α 1 ,  α 20 T = α 2 , and  α 30 T = α 3 , then the impedance profile estimators  f ^ f t ,  K ^ h ( t ) , and    B ^ h ( t )  will be discontinuous at  t = i T  with  i  = 0, 1, 2, · · ·, just as the impedance profile estimators in [26,27]. In this paper, the designed  α 10 t ,  α 20 t , and  α 30 t  in (21)–(23) guarantee the continuity of the impedance profiles estimators at  t = i T  with  i  = 0, 1, 2, ⋯.
Theorem 1.
For the robot dynamics in (1), design the periodic adaptive control strategy in (16), where  θ ^    is updated by the differential type updating law in (19), and  f ^ f t ,  K ^ h ( t ) , and  B ^ h ( t )    are updated by the periodic adaptive laws in (21)–(23). Then, the estimation errors  θ ~   ,  f ~ f ,  K ~ h , and  B ~ h  are uniformly bounded, and the tracking error  e  converges to zero.
Proof. 
As ϵ and θ ^   are with differential dynamics and f ^ f t , K ^ h ( t ) , and B ^ h ( t ) are with periodic difference update learning laws, it is convenient for us to consider the following non-negative function
V t = V 1 t + V f t + V K t + V B t
where V 1 t , V f t , V K t , and V B t are chosen as
V 1 ( t ) = 1 2 ϵ T M ( q ) ϵ + 1 2 α 0 θ ~ T θ ~ , V f ( t ) = 1 2 α 1 t T t   Z 1 T ( τ ) Z 1 ( τ ) d τ , V K ( t ) = 1 2 α 2 t T t   tr { Z 2 T ( τ ) Z 2 ( τ ) } d τ , V B ( t ) = 1 2 α 3 t T t   tr { Z 3 T ( τ ) Z 3 ( τ ) } d τ
with Z 1 t , Z 2 t , and   Z 3 ( t ) .
Z 1 ( t ) = s a t f * ( f f ( t ) ) s a t f * ( f ^ f ( t ) )
Z 2 ( t ) = s a t K * ( K h ( t ) ) s a t K * ( K ^ h ( t ) )
Z 3 ( t ) = s a t B * ( B h ( t ) ) s a t B * ( B ^ h ( t ) )
with Z 1 ( t ) = 0 , Z 2 ( t ) = 0 , Z 3 ( t ) = 0 for t < 0 . □
i. Finiteness of  V t on  [ 0 , T ]
Since θ ^ , f ^ f t , K ^ h ( t ) , and B ^ h ( t ) are continuous functions, the righthand side of Equation (15) is continuous. According to the existence theorem of differential equation, there exists a finite solution to (15) in an interval [ 0 , T 1 ] [ 0 , T ] with   0 < T 1 T . It is obvious that V   ( t ) is finite on the interval [ 0 , T 1 ] . In the following, we will show the finiteness of V   ( t ) on [ T 1 , T ] , where   0 < α 10 T 1 α 10 T < α 1 , 0 < α 20 T 1 α 20 T < α 2 , and 0 < α 30 T 1 α 30 T < α 3 . Taking the time derivative of V   ( t ) for   t T 1 , T   and substituting (18), one can obtain
V ˙ = k 2 ϵ T ϵ ϵ T f ~ f ϵ T K ~ h x ϵ T B ~ h x ˙   1 α 0 θ ~ T ( θ ^ ˙ α 0 Y T ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) ϵ )   + 1 2 α 1 Z 1 T ( t ) Z 1 ( t ) + 1 2 α 2 t r { Z 2 T ( t ) Z 2 ( t ) }   + 1 2 α 3 t r { Z 3 T ( t ) Z 3 ( t ) } .
The used projection modification in (20) guarantees the following: (1) θ ^ t c θ   for t 0   i f θ ^ 0 c θ ; (2) θ ~ T ( θ ^ ˙ α 0 Y T ( x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ ) ϵ ) 0 . From the definitions of   s a t f * ( ) , s a t K * ( ) , and   s a t B * ( ) , we have the following inequalities [29,30]: Z 1 T ( t ) Z 1 ( t ) f ~ f f ~ f , t r { Z 2 T ( t ) Z 2 ( t ) } t r { K ~ h T ( t ) K ~ h } , and t r { Z 3 T ( t ) Z 3 ( t ) } t r { B ~ h T ( t ) B ~ h } . Substituting these results into (31), one obtains
V ˙ k 2 ϵ T ϵ ϵ T f ~ f t r { K ~ h T ϵ x T } t r { B ~ h T ϵ x ˙ T } + 1 2 α 1 f ~ f f ~ f   + 1 2 α 2 t r { K ~ h T ( t ) K ~ h ( t ) } + 1 2 α 2 t r { B ~ h T ( t ) B ~ h ( t ) }
for t [ T 1 , T ] .
From (21)–(23), ϵ = ( f f f ~ f ) / α 10 ( t ) , ϵ q T = ( K h K ~ h ) / α 20 ( t ) , and ϵ q ˙ T = ( B h B ~ h ) / α 30 t for t [ T 1 , T ] .
Applying these equations into (32) yields
V ˙ k 2 ϵ T ϵ α ¯ 1 f ~ f T f ~ f α ¯ 2 t r { K ~ h T K ~ h } α ¯ 3 t r { B ~ h T B ~ h }   + 1 α 10 ( t ) f ~ f f f + 1 α 20 ( t ) t r { K ~ h T ( t ) K h }   + 1 α 30 ( t ) t r { B ~ h T ( t ) B h }
for t T 1 , T , where α ¯ 1 = 1 / α 10 ( t ) 1 / ( 2 α 1 ) > 0 , α ¯ 2 = 1 / α 20 ( t ) 1 / ( 2 α 2 ) > 0 , α ¯ 3 = 1 / α 30 ( t ) 1 / ( 2 α 3 ) > 0
Using Young’s inequality,
f ~ f T f f ξ 1 2 f ~ f T f ~ f + 1 2 ξ 1 f f T f f
t r { K ~ h T K h } ξ 2 2 t r { K ~ h T K ~ h } + 1 2 ξ 2 t r { K h T K h }
t r { B ~ h T B h } ξ 3 2 t r { B ~ h T B ~ h } + 1 2 ξ 3 t r { B h T B h }
where ξ 1 , ξ 2 , ξ 3 are chosen positive constants, such that
β 1 : = α ¯ 1 ξ 1 / ( 2 α 10 ( t ) ) > 0 , β 2 : = α ¯ 2 ξ 2 / ( 2 α 20 ( t ) ) > 0 , β 3 : = α ¯ 3 ξ 3 / ( 2 α 30 ( t ) ) > 0 .
From inequalities in (33)–(36), t [ T 1 , T ]
V ˙ k 2 ϵ T ϵ β 1 f ~ f T f ~ f β 2 t r { K ~ h T K ~ h } β 3 { B ~ h T B ~ h } + c d
where c d is defined as
c d = f d T f d / ( 2 ξ 1 α 10 ( T 1 ) ) + t r { K d T K d } / ( 2 ξ 2 α 20 ( T 1 ) )   + t r { B d T B d } / ( 2 ξ 3 α 30 ( T 1 ) ) .
From (38), V ˙ is negative-definite outside the following region
Ω = { [ ϵ , f ~ f , K ~ h , B ~ h ] : k 2 ϵ T ϵ + β 1 f ~ f T f ~ f + β 2 t r { K ~ h T K ~ h }   + β 3 { B ~ h T B ~ h } c d }
Furthermore, θ ~ is bounded from the projection modification (20). Therefore, V is bounded on [ T 1 , T ] .
ii. Convergence of e
For t T , the time derivatives of V 1   and V f   satisfy
V ˙ 1 ( t ) k 2 ϵ T ϵ ϵ T f ~ f ϵ T K ~ h x ϵ T B ~ h x ˙ ,
V ˙ f t = 1 2 α 1 ( Z 1 T ( t ) Z 1 ( t ) Z 1 T ( t T ) Z 1 ( t T ) )
Substituting (29) and   Z 1 T ( t ) Z 1 ( t ) f ~ f f ~ f into (42), t T
V ˙ f ( t ) = 1 2 α 1 ( f ~ f ( t ) α 1 ϵ ( t ) ) T ( f ~ f ( t ) α 1 ϵ ( t ) )   + 1 2 α 1 Z 1 T ( t ) Z 1 ( t ) α 1 2 ϵ T ( t ) ϵ ( t ) + ϵ T ( t ) f ~ f ( t ) .
Similarly, one obtains   t T
V ˙ K ( t ) α 2 2 ϵ T ( t ) x ( t ) x T ( t ) ϵ ( t ) + t r { K ~ h T ϵ x T }
V ˙ B t α 3 2 ϵ T t x ˙ t x ˙ T t ϵ t + tr B ~ h T ϵ x ˙ T .
Since ϵ T K ~ h x = tr K ~ h T ϵ x T and ϵ T B ~ h x ˙ = tr B ~ h T ϵ x ˙ T , from V ˙ t = V ˙ 1 t + V ˙ f t + V ˙ K t + V ˙ B t and (41)–(45),
V ˙ ( t ) k 2 ϵ T ϵ , t T
from which one can obtain V ( t ) L for t T and
k 2 0   ϵ T ( τ ) ϵ ( τ ) d τ V ( T ) < .
From the above analysis, V ( t ) L for t 0 which implies ϵ , θ ~ L and e , e ˙ L Combining e , e ˙ , ϵ L and x r , x ˙ r L ,   one can obtain x , x ˙ L and C x x , x ˙ , Y x ¨ r + k 1 e ˙ , x ˙ r + k 1 e , x , x ˙ L based on Property 4. Based on x , x ˙ , ϵ L   and (21)–(23),   f ~ f , K ~ h , B ~ h L . Therefore, the righthand side of (18) is bounded and then ϵ ˙ L which shows that ϵ is uniformly continuous in T , . Babalat’s Lemma can be conducted to show that l i m t ϵ = 0 which implies l i m t e = l i m t e ˙ = 0 .
Remark 3.
The proposed impedance learning control method requires that the human impedance profiles are temporally periodic, and the robotic dynamics can be linearly parameterized. If any of these requirements are not satisfied, the proposed control method cannot be implemented.
Remark 4.
The hybrid adaptive impedance learning control can provide variable impedance regulation for repetitive robot-assisted rehabilitation without the requirement of the interactive force sensing.
Remark 5.
In applications, control parameters can be selected step by step. Firstly, PD control gains k1 and k2 are adjusted to make the closed-loop control system stable. Secondly, the adaptive learning rate α0 is designed to make the robot follow the reference trajectory with satisfactory tracking accuracy in the case of no human–robot interaction. Next, periodic learning parameters for time-varying impedance are set. Finally, the control parameters are further fine-tuned to achieve satisfactory interactive performances. Under the proposed hybrid adaptive control strategy, the human–robot system follows the reference trajectory cooperatively and repetitively.

4. Simulations

The planar five-bar parallel mechanism depicted by Figure 1 with the structure in Figure 2 is considered as the mechanism for some upper limb rehabilitation robots and is adopted here for simulation validation. To show the effectiveness of the proposed hybrid adaptive control (HAC) and its advantages in comparison with the adaptive control (AC) in [28] and the biomimetic adaptive control (BAC) in [26,27], simulations are conducted on a planar parallel robot depicted by the render graph in with a five-bar parallel.
In Figure 2, A i , i = 0 ,   1 ,   2 ,   3 ,   4   are the five joint points, where A 0 is the HRI point, and A 2 A 4 , A 4 A 0 ,     A 0 A 3 ,   A 3 A 1 denote four robotic links whose masses, lengths, and inertial moments are given in Table 2. The distances between the joints and respective centers of mass are all 0.25 m, i.e., l c i = 0.25   m for i = 1 ,   2 ,   3 ,   4 , and the length of A 1 A 2 is   0.15   m . In addition to its advantages over a series mechanism, a parallel mechanism results in a more complex model for the robot. In robot-assisted rehabilitation, the reference trajectory is x r = [ 0.075 + 0.15 s i n ( π t / 2 ) ,   0.55 + 0.15 c o s ( π t / 2 ) ] T .
The feedforward force, the stiffness matrix, and the damping matrix are
f ¯ f = [ 10 | cos   ( π / 3 t ) | , 10 | sin   ( π / 3 t ) | ] T K h = diag   { 5 | cos   ( π / 3 t ) | , 5 | sin   ( π / 3 t ) | } B h = diag   { 3 | sin   ( π / 3 t ) | , 3 | cos   ( π / 3 t ) | }
For the proposed HAC, we choose the following control parameters: k 1 = 10 , k 2 = 70 , α 0 = 125 , α 1 = α 2 = α 3 = 600 , α 10 ( t ) = t 2 + ( α 1 T 2 ) t / T , α 20 ( t ) = t 2 + ( α 2 T 2 ) t / T , α 30 ( t ) = t 2 + ( α 3 T 2 ) t / T . In the PD control, the proportional control parameters are determined according to the requirements for the response speed and the static speed error based on the transfer function of the robot model, and the derivative control parameters are designed based on the regulation time and overshooting. Then, the hybrid adaptive learning parameters are determined by trial and evaluation to obtain satisfactory interactive performance.
Figure 3a shows the tracking performance by the proposed HAC, the AC in [28], and the BAC in [26,27] with locally enlarged parts shown in the right part of the figure. From Figure 3a, we can see that the actual trajectory (AT) and the desired trajectory (DT) are almost coincident after several periodic motions using the proposed method, while the AT consistently deviates from the DT using the AC in [28] and the BAC in [26,27]. To better display the effectiveness and advantages of the proposed control method, comparisons of tracking errors among the three methods are illustrated in Figure 3b. It can be seen that the HAC makes the tracking error converge to zero, while the AC in [28] and the BAC in [26,27] only make the tracking error be bounded. The not satisfactory tracking performance by the AC is mainly due to the limitations of traditional differential adaptation in estimating time-varying signals. The better tracking performance validates the superiority of the HAC in comparison to the BAC in [28].
It should be noted that the estimated force cannot be very close to the actual interactive force (see Figure 4a) due to the strong coupling between robot modeling uncertainties and human impedance uncertainties. Such derivation is allowable and does not affect control performance shown in Figure 3. In addition, the stiffness and damping variation is finite, which is illustrated by Figure 4b.

5. Experiments

To further verify the effectiveness and validity of the proposed control method, experiments are implemented on the self-designed planar five-bar parallel mechanism (see Figure 5) which was driven by rotary actuators (erob80, ZeroErrInc., Beijing, China).
Twincat 3.0 eXtended automation engineering running on a PC acts as the fieldbus master in the EtherCAT network. The control algorithm is implemented by the TwinCAT Component Object Model (TcCOM) programed in structured Text (ST) languages. Both the control frequency and the sampling period for the controller are 1 ms. The rotary actuator is operated in torque mode. The planar parallel robot used in our study is an end-effector rehabilitation robot. The trajectory of the robot is usually predefined by therapists during rehabilitation training. During the rehabilitation process, the patient’s upper limbs are trained with the help of a rehabilitation robot, as in Figure 5. The emerging interactive forces between subjects and the robot may give rise to trajectory deviation for the robot. The proposed impedance learning control can make the human robot system cooperatively follow the desired trajectory. In the experiment, the robot and a subject repetitively track an ellipse trajectory. A typical set of experimental results of the proposed method is illustrated in Figure 6 which presents the tracking performance from the 1st task completion to the 40th completion. It can be inferred from Figure 6 that the tracking error is satisfactory after 30 task completions, i.e., after 120 s. The tracking error in the experiment will not converge to zero due to measurement noises, device inherent deficiencies, and so on. The proposed HAC provides finite stiffness variation for robot–environment interaction (see Figure 7). Although there exists some chattering in the experiment results by the proposed HAC due to measurement noises and the difference form of the periodic adaptation in (21)–(23), the chattering is very limited and is allowable. The experimental results by the AC [28] are shown in Figure 8. It is obvious that the tracking performance and the tracking error are unsatisfactory in the whole interactive process. The reason comes from the inherent deficiencies of the traditional differential adaptation in estimating time-varying signals. Taking Figure 6 and Figure 8, we can conclude the superiority of the proposed HAC in comparison with the AC [28].

6. Conclusions

This paper presents an impedance learning-based HAC strategy for HRI with robotic modeling parameter uncertainties and human impedance uncertainties. The robotic parameter vector is constant and is estimated by a differential adaptive law with projection modification, while the time-varying impedance profile including the feedforward force, the stiffness, and the damping is estimated by periodic adaptations in (21)–(23). In Theorem 1, we prove the convergence of tracking errors and the boundedness of estimation errors using Babalat’s Lemma. Though the illustrative example on a parallel robot, we validate the following: (i) the HAC achieves time-varying impedance regulation for HRI with a stability guarantee; (ii) the asymptotic convergence of the tracking error is achieved by the HAC as opposed to the uniformly ultimately bounded results of the tracking errors in [25]; and (iii) the periodic adaptations have better performances in the estimation of time-varying impedance profiles than the differential adaptation in [25] by using the periodicity.
The subjects in this study are healthy individuals. Next, we will apply the device to patients with spasticity in upper limbs for assisted rehabilitation to study its robustness, as well as its capability in assistive movement with low tracking error in the presence of spasticity.

Author Contributions

Conceptualization, Z.J., Q.L. and J.Y.; Investigation, Z.W.; Methodology, J.Y.; Writing—original draft, Z.W.; Writing—review and editing, Z.J. and J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 62103280.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

Author Qipeng Lv was employed by the company The Second Research Institute of China Electronics Technology Group Corporation. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhang, F.; Hou, Z.G.; Cheng, L.; Wang, W.; Chen, Y.; Hu, J.; Peng, L.; Wang, H.B. iLeg—A lower limb rehabilitation robot: A proof of concept. IEEE Trans. Hum.-Mach. Syst. 2016, 46, 761–768. [Google Scholar] [CrossRef]
  2. Hogan, N. Impedance control: An approach to manipulation: Part II—Implementation. J. Dyn. Syst. Meas. Control 1985, 107, 8–16. [Google Scholar] [CrossRef]
  3. Spong, M.W.; Hutchinson, S.; Vidyasagar, M. Robot Modeling and Control, 2nd ed.; John Wiley & Sons: New York, NY, USA, 2006; pp. 1–18. [Google Scholar]
  4. He, W.; Tang, X.; Wang, T.; Liu, Z. Trajectory tracking control for a three-dimensional flexible wing. IEEE Trans. Contr. Syst. Technol. 2022, 30, 2243–2250. [Google Scholar] [CrossRef]
  5. Chan, S.P.; Yao, B.; Gao, W.B.; Cheng, M. Robust impedance control of robot manipulators. Int. J. Robot. Autom. 1991, 6, 220–227. [Google Scholar]
  6. Li, Z.; Huang, Z.; He, W.; Su, C.Y. Adaptive impedance control for an upper-limb robotic exoskeleton using biological signals. IEEE Trans. Ind. Electron. 2017, 64, 1664–1674. [Google Scholar] [CrossRef]
  7. Sun, T.; Peng, L.; Cheng, L.; Hou, Z.G.; Pan, Y. Composite learning enhanced robot impedance control. IEEE Trans. Neur. Net. Lear. 2010, 31, 1052–1059. [Google Scholar] [CrossRef]
  8. He, W.; Dong, Y.; Sun, C. Adaptive neural impedance control of a robotic manipulator with input saturation. IEEE Trans. Syst. Man Cybern. 2016, 46, 334–344. [Google Scholar] [CrossRef]
  9. Yang, C.; Chen, C.; He, W.; Cui, R.; Li, Z. Robot learning system based on adaptive neural control and dynamic movement primitives. IEEE Trans. Neur. Net. Lear. 2019, 30, 777–787. [Google Scholar] [CrossRef]
  10. Yang, C.; Peng, G.; Cheng, L.; Na, J.; Li, Z. Force sensorless admittance control for teleoperation of uncertain robot manipulator using neural networks. IEEE Trans. Syst. Man Cybern. A 2021, 51, 3282–3292. [Google Scholar] [CrossRef]
  11. He, W.; Xue, C.; Yu, X.; Li, Z.; Yang, C. Admittance-based controller design for physical human-robot interaction in the constrained task space. IEEE Trans. Autom. Sci. Eng. 2022, 19, 1937–1949. [Google Scholar] [CrossRef]
  12. Li, X.; Liu, Y.H.; Yu, H. Iterative learning impedance control for rehabilitation robots driven by series elastic actuators. Automatica 2018, 90, 1–7. [Google Scholar] [CrossRef]
  13. Buchli, J.; Stulp, F.; Theodorou, E.; Schaal, S. Learning variable impedance control. Int. J. Robot. Res. 2011, 30, 820–833. [Google Scholar] [CrossRef]
  14. Ajoudani, A.; Tsagarakis, N.; Bicchi, A. Tele-impedance: Teleoperation with impedance regulation using a body-machine interface. Int. J. Robot. Res. 2012, 31, 1642–1656. [Google Scholar] [CrossRef]
  15. Howard, M.; Braun, D.J.; Vijayakumar, S. Transferring humanimpedance behavior to heterogenous variable impedance actuators. IEEE Trans. Robot. 2013, 29, 847–862. [Google Scholar] [CrossRef]
  16. Kronander, K.; Billard, A. Learning compliant manipulation through kinesthetic and tactile human-robot interaction. IEEE Trans. Haptics 2013, 7, 367–380. [Google Scholar] [CrossRef]
  17. Abu-Dakka, F.J.; Saveriano, M. Variable impedance control and learning-a review. Front. Robot. AI 2020, 7, 590681. [Google Scholar] [CrossRef]
  18. Ficuciello, F.; Villani, L.; Siciliano, B. Variable impedance control of redundant manipulators for intuitive human-robot physical interaction. IEEE Trans. Robot. 2015, 31, 850–863. [Google Scholar] [CrossRef]
  19. Duan, J.; Gan, Y.; Chen, M.; Dai, X. Adaptive variable impedance control for dynamic contact force tracking in uncertain environment. Robot. Auton. Syst. 2018, 102, 54–65. [Google Scholar] [CrossRef]
  20. Kronander, K.; Billard, A. Stability considerations for variable impedance control. IEEE Trans. Robot. 2016, 32, 1298–1305. [Google Scholar] [CrossRef]
  21. Dong, Y.; Ren, B. UDE-based variable impedance control of uncertain robot systems. IEEE Trans. Syst. Man Cybern. A 2019, 49, 2487–2498. [Google Scholar] [CrossRef]
  22. Sun, T.; Peng, L.; Cheng, L.; Hou, Z.G.; Pan, Y. Stability-guaranteed variable impedance control of robots based on approximate dynamic inversion. IEEE Trans. Syst. Man Cybern. A 2021, 51, 4193–4200. [Google Scholar] [CrossRef]
  23. Modares, H.; Ranatunga, I.; Lewis, F.L.; Popa, D.O. Optimized assistive human-robot interaction using reinforcement learning. IEEE Trans. Cybern. 2016, 46, 655–667. [Google Scholar] [CrossRef]
  24. Peng, G.; Chen, C.P.; Yang, C. Robust admittance control of optimized robot-environment interaction using reference adaptation. IEEE Trans. Neur. Net. Learn. 2023, 34, 5804–5815. [Google Scholar] [CrossRef]
  25. Sadrfaridpour, B.; Wang, Y. Collaborative assembly in hybrid manufacturing cells: An integrated framework for human robot interaction. IEEE Trans. Autom. Sci. Eng. 2018, 15, 1178–1192. [Google Scholar] [CrossRef]
  26. Yang, C.; Ganesh, G.; Haddadin, S.; Parusel, S.; Albu-Schaeffer, A.; Burdet, E. Human-like adaptation of force and impedance in stable and unstable interactions. IEEE Trans. Robot. 2011, 27, 918–930. [Google Scholar] [CrossRef]
  27. Li, Y.; Ganesh, G.; Jarrassé, N.; Haddadin, S.; Albu-Schaeffer, A.; Burdet, E. Force, impedance, and trajectory learning for contact tooling and haptic identifification. IEEE Trans. Robot. 2018, 34, 1170–1182. [Google Scholar] [CrossRef]
  28. Sharifi, M.; Azimi, V.; Mushahwar, V.K.; Tavakoli, M. Impedance learning-based adaptive control for human-robot interaction. IEEE Trans. Contr. Syst. Technol. 2022, 30, 1345–1358. [Google Scholar] [CrossRef]
  29. Dixon, W.E.; Zergeroglu, E.; Dawson, D.M.; Costic, B.T. Repetitive learning control: A lyapunov-based approach. IEEE Trans. Syst. Man Cybern. B 2002, 32, 538–545. [Google Scholar] [CrossRef]
  30. Costic, B.T.; de Queiroz, M.S.; Dawson, D.N. A new learning control approach to the active magnetic bearing benchmark system. In Proceedings of the 2000 American Control Conference, Chicago, IL, USA, 28–30 June 2000; pp. 2639–2643. [Google Scholar]
Figure 1. The render graph of a planar parallel robot.
Figure 1. The render graph of a planar parallel robot.
Actuators 13 00220 g001
Figure 2. The robotic structure diagram.
Figure 2. The robotic structure diagram.
Actuators 13 00220 g002
Figure 3. Simulation performances by the proposed HAC, the AC in [28], and the BAC in [26,27]. (a) The tracking performance, (b) The tracking error, (c) The control input.
Figure 3. Simulation performances by the proposed HAC, the AC in [28], and the BAC in [26,27]. (a) The tracking performance, (b) The tracking error, (c) The control input.
Actuators 13 00220 g003aActuators 13 00220 g003b
Figure 4. The estimation of the interactive force and the estimated stiffness and damping by the proposed HAC. (a) Interactive force estimation, (b) Stiffness and damping.
Figure 4. The estimation of the interactive force and the estimated stiffness and damping by the proposed HAC. (a) Interactive force estimation, (b) Stiffness and damping.
Actuators 13 00220 g004
Figure 5. Physical HRI.
Figure 5. Physical HRI.
Actuators 13 00220 g005
Figure 6. The experimental results by the proposed HAC. (a) The tracking performance, (b) The tracking error, (c) The control input.
Figure 6. The experimental results by the proposed HAC. (a) The tracking performance, (b) The tracking error, (c) The control input.
Actuators 13 00220 g006aActuators 13 00220 g006b
Figure 7. The estimated stiffness by the proposed HAC.
Figure 7. The estimated stiffness by the proposed HAC.
Actuators 13 00220 g007
Figure 8. The experimental results by the AC in [29]. (a) The tracking performance, (b) The tracking error, (c) The control input.
Figure 8. The experimental results by the AC in [29]. (a) The tracking performance, (b) The tracking error, (c) The control input.
Actuators 13 00220 g008aActuators 13 00220 g008b
Table 1. A summary of the currently used adaptive controllers with impedance regulation in rehabilitation.
Table 1. A summary of the currently used adaptive controllers with impedance regulation in rehabilitation.
IndexCurrently Used MethodsInnovations or Contributions of This Paper
1The constant impedance control [14,15,16]Time-varying impedance regulation for rehabilitation robots are provided
2The impedance learning controllers [13,18,19,20,21,22]The control stability without using interaction force signals is guaranteed
3The model-based impedance learning controllers [26,27,28](a) The asymptotical convergence of tracking errors is guaranteed
(b) Robotic parameter uncertainties are considered and estimated
(c) The impedance profiles are not required to be slowly varying
Table 2. The parameters of the parallel robot.
Table 2. The parameters of the parallel robot.
LinkMass (kg)Length (m)Inertial
Moment   ( kg · m 2 )
A 2 A 4 m 2 = 4.859 l 2 = 0.5 I 2 = 0.128
A 4 A 10 m 4 = 4.053 l 4 = 0.5 I 4 = 0.093
A 0 A 3 m 3 = 1.423 l 3 = 0.5 I 3 = 0.0327
A 3 A 1 m 1 = 4.859 l 1 = 0.5 I 1 = 0.128
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, Z.; Wang, Z.; Lv, Q.; Yang, J. Impedance Learning-Based Hybrid Adaptive Control of Upper Limb Rehabilitation Robots. Actuators 2024, 13, 220. https://doi.org/10.3390/act13060220

AMA Style

Jiang Z, Wang Z, Lv Q, Yang J. Impedance Learning-Based Hybrid Adaptive Control of Upper Limb Rehabilitation Robots. Actuators. 2024; 13(6):220. https://doi.org/10.3390/act13060220

Chicago/Turabian Style

Jiang, Zhenhua, Zekai Wang, Qipeng Lv, and Jiantao Yang. 2024. "Impedance Learning-Based Hybrid Adaptive Control of Upper Limb Rehabilitation Robots" Actuators 13, no. 6: 220. https://doi.org/10.3390/act13060220

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop