Next Article in Journal
On the Benefits of Providing Timely Information in Ticket Queues with Balking and Calling Times
Next Article in Special Issue
Toward Embedded System Resources Relaxation Based on the Event-Triggered Feedback Control Approach
Previous Article in Journal
Vehicle Routing Problem with Deadline and Stochastic Service Times: Case of the Ice Cream Industry in Santiago City of Chile
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Cognitive Control for Unknown Dynamic Systems Tracking

Department of Automation and Applied Informatics, Politehnica University of Timisoara, 300223 Timisoara, Romania
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(21), 2752; https://doi.org/10.3390/math9212752
Submission received: 28 September 2021 / Revised: 22 October 2021 / Accepted: 26 October 2021 / Published: 29 October 2021
(This article belongs to the Special Issue Mathematical Methods for Nonlinear Control)

Abstract

:
A general control system tracking learning framework is proposed, by which an optimal learned tracking behavior called ‘primitive’ is extrapolated to new unseen trajectories without requiring relearning. This is considered intelligent behavior and strongly related to the neuro-motor cognitive control of biological (human-like) systems that deliver suboptimal executions for tasks outside of their current knowledge base, by using previously memorized experience. However, biological systems do not solve explicit mathematical equations for solving learning and prediction tasks. This stimulates the proposed hierarchical cognitive-like learning framework, based on state-of-the-art model-free control: (1) at the low-level L1, an approximated iterative Value Iteration for linearizing the closed-loop system (CLS) behavior by a linear reference model output tracking is first employed; (2) an experiment-driven Iterative Learning Control (EDILC) applied to the CLS from the reference input to the controlled output learns simple tracking tasks called ‘primitives’ in the secondary L2 level, and (3) the tertiary level L3 extrapolates the primitives’ optimal tracking behavior to new tracking tasks without trial-based relearning. The learning framework relies only on input-output system data to build a virtual state space representation of the underlying controlled system that is assumed to be observable. It has been shown to be effective by experimental validation on a representative, coupled, nonlinear, multivariable real-world system. Able to cope with new unseen scenarios in an optimal fashion, the hierarchical learning framework is an advance toward cognitive control systems.

Graphical Abstract

1. Introduction

Cognitive control systems are characterized by perception, learning, planning and memorization, defining a clear path toward general intelligent systems that are able to optimally handle new, unseen before situations and display adaptability. The best examples of collections of such complex control systems are offered by certain biological systems (living organisms, humans in particular) who intelligently solve complex tasks by combining knowledge gathered through experience via learning and planning mechanisms that are encoded in the brain. As such, the brain acts as a processing unit that is able, at a higher level, to project future strategies in order to achieve goals through reasoning and planning, decompose complex strategies into well-known possibly simpler strategies (planning), memorize new experiences to augment the current knowledge base, and process feedback signals by fusing visual, tactile, auditory, and olfactory information for task execution improvement through learning. Finally, it guides the lower-level neuromuscular control system to act in the environment to achieve higher-level goals.
This intelligence of living organisms has developed without solving explicit mathematical equations; therefore, this kind of complex intelligence serves as inspiration for developing model-free control, learning, and planning techniques. One suitable paradigm for explaining the above techniques is the process of learning neuromotor skills, which rely on well-acknowledged control formulations such as those found in feed-forward- and feedback-based control and predictive-based control. A common trait of all these techniques is that they are almost always posed and solved in optimization-based settings [1]. In such settings (e.g., taking error-based learning), the error gradient is exploited to improve the control solutions. Alternatively, use-dependent learning addresses the permanent compromise between repeatable tasks whose improved execution is improvable for high-performance and non-repeatable tasks that require a significant relearning effort.
One illustrative natural behavior for some of the above concepts is the imitating behavior of biological systems: the first execution of an imitation task is nearly optimal. The brain can analyze, memorize, and decompose the imitation task into subtasks for each limb and plan for corresponding immediate future tracking tasks. The control of each limb is already encoded in the neuromuscular system through an inverse dynamical model that is not explicitly represented. Another illustrative situation occurs when new neuromotor tasks are solved by first decomposing them into tasks with known solutions and then recomposing the new solution from existing ones. The previously described behavior unites several concepts that can be formally called a primitive-based learning framework [1].
Primitive-based control has been researched for at least two decades, in various forms, by transforming the time scale [2,3], concatenation-based mechanisms [4,5,6], and decomposition/re-composition in time [7,8]. More recently, primitive-based control has been studied in [9,10,11,12,13], mostly as part of hierarchical learning frameworks [14,15,16,17,18]. However, the application of the Iterative Learning Control (ILC, a full list of the acronyms used in the paper is presented in abbreviations part) technique [19,20,21,22,23,24,25] over linearized feedback closed-loop control systems (CLSs) as a primer mechanism for primitive-based learning was proposed in [8]. An experiment-driven ILC (EDILC) variant was crafted in a norm-optimal tracking setting to learn reference-input controlled output pairs called primitives. Such a primitive tuple contains a reference input to the CLS (called the input primitive) coupled with its paired control system output (called the output primitive). Subsequently, the output primitive is usually shaped as a Gaussian function, but it can be any shape suitable for approximation purposes. Delayed copies of the output primitives are used to approximate a new trajectory that must be tracked. The coefficients that combine the output primitives linearly to approximate the new trajectory are used to combine the delayed and copied input primitives to obtain the optimal reference input. For the assumed linear CLS, this reconstructed reference input is optimal with respect to the criterion by which the original primitives are learned; therefore, when set as input to the CLS, it leads to theoretically perfect new trajectory tracking. It does this without repetitive-based relearning using the EDILC.
The primitive-based approach is sensible to the CLS’s linearity assumption; therefore, a control that makes the CLS as linear as possible has to be learned. One framework to ensure this linearization is the approximate (adaptive) dynamic programming (ADP) [26,27,28,29,30], also known as reinforcement learning (RL) [31,32,33,34,35], designed for the output reference model (ORM) paradigm. Specifically, a model-free iterative Value Iteration (IVI) algorithm as a representative ADP/RL approach can serve this goal by using general function approximators such as neural networks (NNs). Hence, it is called the approximated IVI (AIVI).
With a linear ORM’s output well tracked by the CLS output, an approximate linearized CLS from the reference input to the controlled output is ensured in a wide range, together with some inherent disturbance attenuation ability. Two major problems with ADP/RL are exploration and state knowledge [36,37,38,39,40]. The former is still an open issue, although much research is underway. Enhanced and accelerated exploration is achievable with pre-stabilizing controllers, which can also be designed based on the unknown dynamics assumption principles. The state availability was solved with virtual state-space models constructed from input-output (I/O) samples of general unknown but assumed observable nonlinear systems [41,42]. Using historical I/O samples from the system as an alternate system, state representation has been used in more complex but unformalized environments, such as video games [43].
This work aims to integrate the learning control techniques and machine learning techniques previously presented in a hierarchical cognitive-based control framework tailored to extrapolate optimally learned tracking control to novel tracking tasks. Tracking should be achievable without relearning through trials by exploiting the already learned optimal tracking skills. It is obtained through a hierarchical learning architecture where (a) the closed-loop system at a lower level (called L1 level herein) is first linearized using the ORM principle. This is achieved through nonlinear state-feedback control, where a virtual state is built from finite sequences of I/O system data samples. Learning such a controller is performed using AIVI; (b) an EDILC is used at the secondary upper L2 level to learn the optimal tracking skills called primitives, by repetitions; (c) the tertiary and last L3 level is dedicated to extrapolating the optimally learned tracking skills, without repetitions, to new tracking tasks.
Progress and contributions with respect to work [8] are:
  • At level L1, CLS linearization is strongly ensured using a virtual state-feedback neuro-controller learned by AIVI. The new state-space representation relies on a virtual transformation empowered by observability theory.
  • At level L2, the model-free EDILC dedicated to primitive learning is improved in several aspects: (1) the gradient information is extracted using a single gradient experiment in the vicinity of the current iteration nominal trajectory, irrespective of the CLS’s number of controlled outputs; (2) EDILC monotonic convergence is derived by optimally selecting the learning rate parameter in order to trade-off learning speed and robustness. This is achieved using two approximate CLS models for a double-safe mechanism; (3) proper initialization of the gradient-based search specific to EDILC results in fewer iterations for convergence.
  • The primitive-based mechanism at level L3, for optimally predicting the reference input ensuring theoretical perfect tracking of a previously unseen desired trajectory, is designed to: (1) handle desired trajectories longer than the learned primitives, and (2) indirectly handle constraints on the CLS’s output.
The following bottom-up presentation of the proposed hierarchical learning framework results in an effective validation of a hybrid software-electrical system that is coupled, multivariable, and has a delay. The case study is representative of many mechatronic systems.

2. The Linear ORM Tracking for Observable Systems with Unknown Dynamics

A common and general representation for a deterministic controlled system model in a discrete-time state space is
{ s k + 1 e x t = E ( s k e x t , a k ) , y k = v k , 1 ,   y k O R M = L s k O R M ,
where E() is a partially known nonlinear state map, k is the sample index, a k = [ a k , 1 , , a k , m u ] T Ω A m u lumps the mu control action inputs of the system within a known domain Ω A . A number of p measurable system outputs are formally collected in the partial output vector y k = [ y k , 1 , , y k , p ] T Ω Y p with known domain Ω Y . Another p measurable system outputs formally complete the remaining output as y k O R M = [ y k , 1 O R M , , y k , p O R M ] T Ω Y O R M p with known domain Ω Y O R M , to be contextualized later on. The state s k e x t = [ v k T , ( s k O R M ) T , ρ k T ] T lumps three types of state variables stemming from three independent systems. First, a virtual state-space system is defined as follows:
{ v k + 1 = F ( v k , a k ) , y k = v k , 1 ,
which is a transformation of the true state–space system
{ s k + 1 = f ( s k , a k ) , y k = g ( s k ) ,
where s k = [ s k , 1 s k , n ] T Ω S n gathering the n (which is unknown) true system states that cannot be measured, ak are the control action inputs of the true system, and the measurable true system outputs are yk. The system functions as nonlinear maps f ( s k , a k ) : Ω S × Ω A Ω S , g ( s k ) : Ω S Ω Y are unknown and are assumed to be continuously differentiable. Some assumptions regarding (1) were introduced. The pair ( f , g ) is entirely state observable, meaning that the true state s k from (3) is recoverable from the present and past I/O measurements of a k , y k . The observability property is used in the sense defined for linear systems. Further, (3) is assumed to be controllable, and the system’s relative degree (delay included) is known and constant.
With unknown system functions ( f , g ), observability and controllability can only be inferred from the working experience with the system, from technical manuals, or from the literature. The system’s relative degree and its non-minimum phase (NMP) type are recognizable from historical I/O response data.
Relying on the system’s (3) complete observability, using Lemma 1 from [41] shows how to build transformation (2) whose virtual state vector is defined in terms of I/O samples of (3): v k = [ ( Y k , k τ ¯ ) T , ( A k 1 , k τ ¯ ) T ] T = [ ( v k , 1 ) T , ( v k , 2 ) T , , ( v k , 2 τ + 1 ) T ] T Ω Y × × Ω Y τ + 1   times × Ω A × × Ω A τ   times = Ω v p ( τ + 1 ) + m u τ . For a more detailed element-wise correspondence, the indexing Y k , k τ ¯ = [ ( y k ) T ( y k τ ) T ] T = Δ [ ( v k , 1 ) T , ( v k , τ + 1 ) T ] T ,   A k 1 , k τ ¯ = [ ( a k 1 ) T ( a k τ ) T ] T = Δ [ ( v k , τ + 2 ) T ( v k , 2 τ + 1 ) T ] T is utilized. Remarkably, knowing Ω Y ,   Ω A , makes Ω v known. The τ (ℕ is the set of positive integers) indexes the historical I/O measurements, and its value is related to the order of (3); therefore, it is also correlated with the unknown observability index of (3).
Definition 1.
The unknown observability index of (3) is the minimal value τmin of τ for which state sk is fully recoverable from the I/O measurements  Y k , k τ ¯ ,   A k 1 , k τ ¯ . The role of this index is similar to that of linear systems.
Under the observability assumption, there exists a minimal value τ = τ m i n which makes s k fully recovered from I/O data and uniquely associated with v k which is a sort of alias for s k , but in a different dimensional space [41]. For τ > τ m i n , the size increase of v k does not add valuable information about s k . In practice, τ should be as large as possible without negatively affecting the computational complexity of the subsequent learning of state-feedback control based on high-dimensional v k .
Based on transformation (2), system (2) is also I/O controllable, having the same I/O as (1). The main feature of (2) is its complete state availability, making it fully state observable but without entirely known dynamics because the map F ( . ) is partially unknown. Input delay (dead-time) systems also hold transforms such as (2) to render them fully state observable.
The second component of s k e x t is the state s k O R M of the ORM model which will be matched by the CLS by proper control design; let this known ORM state-space model be
{ s k + 1 O R M = G s k O R M + H ρ k , y k O R M = L s k O R M ,
where s k O R M = [ s k , 1 O R M , , s k , n m O R M ] T Ω S O R M n m gathers the nm ORM states in known domain Ω S O R M , ρ k = [ ρ k , 1 , , ρ k , p ] T Ω ρ p lumps the ORM’s inputs in their known domain Ω ρ ; ρ k will also serve as CLS input. The ORM outputs are collected in y k O R M introduced in the first place. Note that ρ k completes the definition of s k e x t in (1).
The tuple ( G , H , L ) characterizes the linear ORM for which an I/O pulse transfer matrix (t.m.) is y k O R M = T O R M ( q ) ρ k (q is the unit time step advance operator). Information about (3), such as relative degree, non-minimum phase character, and bandwidth, must be synthesized within the ORM (4), as dictated by classical control rules. Commonly, this selection is better reflected in a T O R M ( q ) , as it is more straightforward and easier to interpret (bandwidth, overshoot).
To fulfill the Markovian property of (1), the exogenous reference input signal ρ k is described by the known generation model dynamics ρ k + 1 = Γ ( ρ k ) , where ρ k acts as a measurable state variable. Note that the extended model (1) has two outputs stemming from (3) and (4). In summary, the dynamics E ( . ) will be partially unknown owing to the partially unknown dynamics of its component F ( . ) . The main feature of (1) is its full state observability, that is, the state is fully measurable, and the domain of s k e x t is known as Ω S e x t = Ω v × Ω S O R M × Ω ρ .
The ORM tracking problem finally poses as the optimal control search over an infinite-horizon cost V O R M :
a k * = arg   m i n a k   V O R M ( a k ) ,   V O R M ( a k ) = k = 0 y k ( a k ) y k O R M 2 2 , s . t .   d y n a m i c s   ( 1 ) .
The norm . 2 defined as the vector-wise Euclidean distance, penalizes at each sampling instant the CLS’s output deviation from the ORM’s output. Problem (5) is a type of imitation (or apprentice) learning with the ORM (3) acting as a teacher (or expert or supervisor) and with system (1) trying to mimic (or follow) the ORM dynamics. Assuming that (5) is solvable in the following, ADP/RL techniques are well suited for finding closed-form optimal control laws of the type a k = C ( s k e x t ) with the control action input as a nonlinear mapping of the extended state. An AIVI is in fact employed to solve (5) as a competitive alternative to the recently proposed Virtual State-Feedback Reference Tuning (VSFRT) [42]. This is considered the L1 level control and is presented next.

3. The L1 Level—ORM Tracking AIVI-Based Solution

AIVI is a type of model-free offline batch reinforcement Q-learning that is used to find the optimal control solution to (5). AIVI most often uses NNs as function approximators because of their easily customizable architecture, training software availability, and generalization capacity. As such, AIVI uses one NN to parameterize the nonlinear control action as a k = C ( s k e x t , ψ ) and another NN to parameterize the well-known Q-function as Q ( s k e x t , a k , π ) . Here, ψ and π formally gather all the NN tunable weights corresponding to some generic NN architectures. Starting from some initial solutions ψ 0 and π 0 , two AIVI-specific steps are called iteratively, and in this particular order, the Q-function update and controller improvement. The steps are formally expressed as
π j = arg   m i n π 1 B i = 1 B [ Q ( s k e x t [ i ] , u k [ i ] , π ) r ( s k e x t [ i ] , a k [ i ] ) Q ( s k + 1 e x t [ i ] , C ( s k + 1 e x t [ i ] , ψ j 1 ) , π j 1 ) ] 2 ,
ψ j = arg   m i n ψ 1 B i = 1 B Q ( s k e x t [ i ] , C ( s k e x t [ i ] , ψ ) , π j ) ,
respectively, with penalty r ( s k e x t [ i ] , a k [ i ] ) = y k ( a k ) y k O R M 2 2 , over a batch consisting of B transition samples (also known tuples or experiences) of the form { ( s k e x t [ i ] , a k [ i ] , s k + 1 e x t [ i ] ) } , i = 1 , B ¯ . Here, j indexes the AIVI iteration. It is easily shown that (6) and (7) are equivalent to the NN training of the Q-function and controller, respectively. The transition samples set is collectable from the controlled system (5) under any sufficiently exploratory (i.e., it visits many state-action combinations which uniformly span their space) controller. Even a random control strategy may be suitable for this purpose; however, a priori stabilizing controllers accelerate the exploration process: faster visiting more state-action combinations while simultaneously dealing with operational (e.g., safety) constraints. This off-policy offline is more data efficient than online on-policy reinforcement learning, where the exploration-exploitation dilemma throughout learning, combined with the lower data efficiency, is a major challenge. When all (or a random subset) of the transition samples are used in each AIVI step, the approach ensures the so-called experience replay.
The AIVI steps converge to the optimal virtual state feedback controller, ensuring that the CLS tracks the ORM in terms of reference input-controlled output at the L1 level [31,32,40,41]. It is assumed that the CLS is indirectly linearized in the ORM tracking sense. Thus, it is prepared to receive the L2 level learning phase, next presented.

4. The L2 Level—EDILC for Learning Primitives

The L2 level aims to learn pairs of reference inputs and CLS outputs called primitive pairs or simply primitives, where the CLS outputs have a basis function shape (e.g., Gaussian), to be used later for function approximation. To this end, we search for ρ k which makes y k track a desired trajectory y k d . The CLS is assumed to be reset with each ILC iteration to zero initial conditions. The CLS response to non-zero (but constant) initial conditions and repeated disturbances is easily absorbed into y k d . At the given ILC iteration j, the CLS reference input-controlled output relationship is written in lifted (or supervector) notation over an N-length experiment, as in
Y j = T P j ,
where the reference input at iteration j is P j = [ ( P 1 j ) T ( P 2 j ) T ( P p j ) T ] T p N × 1 , P l j = [ ρ 0 , l j ,   ρ 1 , l j , ,   ρ N 1 , l j ] T   N × 1 ,   l = 1 , p ¯ , the CLS output at iteration j is Y j = [ ( Y 1 j ) T ( Y 2 j ) T ( Y p j ) T ] T p N × 1 , Y l j = [ y 0 , l j , y 1 , l j , , y N 1 , l j ] T N × 1 ,   l = 1 , p ¯ . The impulse response terms of the pulse transfer matrix (t.m.) operator T ( q ) characterizing the CLS are used to build the matrix T. To see how T emerges, the p × p causal CLS at the current iteration j is (in time notation)
[ y k , 1 j y k , 2 j y k , p j ] = [ T 1 , 1 ( q ) T 2 , 1 ( q ) T p , 1 ( q ) T 1 , 2 ( q ) T 2 , 2 ( q ) T p , 2 ( q ) T 1 , p ( q ) T 2 , p ( q ) T p , p ( q ) ] T ( q ) [ ρ k , 1 j ρ k , 2 j ρ k , p j ] ,
where j = 0 , 1 , 2 , is the iteration index, k = 0 , 1 , is the discrete-time index, y k , l j is the lth output at time k in iteration j, ρ k , l j is the lth input at time k in iteration j, and T i , l ( q ) are the pulse transfer functions (t.f.) from input i to output l. The t.m. T ( q ) encompasses the feedback CLS with some kind of a multiple input multiple output (MIMO) (possibly nonlinear) feedback controller over a MIMO (possibly nonlinear) controlled system. The feedback CLS is square; that is, each reference input drives one controlled output. In general, T ( q ) was coupled.
With each rational t.f. T i , l ( q ) admitting an infinite T i , l ( q ) = + t 2 i , l q 2 + t 1 i , l q + t 0 i , l + t 1 i , l q 1 + t 2 i , l q 2 + power series expansion, a truncated lifted notation of length N for the ILC system is
[ y 0 , 1 j y 1 , 1 j y N 1 , 1 j y 0 , p j y 1 , p j y N 1 , p j ] = [ T 1 , 1 T 2 , 1 T p , 1 T 1 , p T 2 , p T p , p ] T [ ρ 0 , 1 j ρ 1 , 1 j ρ N 1 , 1 j ρ 0 , p j ρ 1 , p j ρ N 1 , p j ] Y j = T P j ,
where each T i , l = [ t 0 i , l t 1 i , l t 1 N i , l t 1 i , l t 0 i , l t 2 N i , l t N 1 i , l t N 2 i , l t 0 i , l ] N × N is built from the impulse response coefficients of T i , l ( q ) and all the T i , l build the T p N × p N (note the difference between T and T ( q ) ).
Let the desired output trajectory be Y d = [ ( Y 1 d ) T   ( Y 2 d ) T     ( Y p d ) T ] T p N × 1 , Y l d   =   [ y 0 , l d ,   y 1 , l d ,   ,   y N 1 , l d ] T   N × 1 ,   l = 1 , p ¯ . The learning goal is formally expressed as an optimization problem without constraints as:
P * = arg   m i n P J ( P ) ,   J ( P ) = 1 N E ( P ) 2 2 = 1 N T P Y d 2 2 + λ N P 2 2 ,
where the tracking error in the current iteration is E j = Y d T P j . The non-zero regularization coefficient λ may be useful in cases where, for example, Y d does not acknowledge a possible NMP T ( q ) . For simplicity, no regularization is used next ( λ = 0 ). Directly working with T in practice raises several numerical difficulties: ill-conditioning, large size, and its identification is costly. A convenient numerical solution to (11) is P * = ( T T T ) 1 T T Y d . However, without using T explicitly, (11) is solved based on an ILC-specific iterative update law as a gradient descent search starting from an initial P 0 :
[ P 1 j + 1 P 2 j + 1 P p j + 1 ] = [ P 1 j P 2 j P p j ] ξ [ J ( P ) P 1 J ( P ) P p ] | P = P j = [ P 1 j P 2 j P p j ] ξ J ( P j ) P ,
with a positive-definite matrix ξ = d i a g ( ξ i ) ,   i = 1 , p ¯ is a user-selectable learning rate. The gradient in (12) computed according to the cost in (11) is 2 N · T T E j and is based on an unknown T. The next EDILC Algorithm 1 extracts this gradient information experimentally, and its steps are:
Algorithm 1
  Step 0. Initialize P 0 = Y d , then for each j = 0 , ϵ ¯ call the subsequent steps.
  Step 1. In a nominal experiment, with the current iteration reference input Pj, measure Ej. Flip Ej to obtain udf(Ej): the upside-down flipped version of Ej.
  Step 2. In a so-called gradient experiment, use μ·udf(Ej) (with scaling factor μ) as an additive disturbance on the current iteration Pj. With the disturbed reference input, the CLS output Y G j . It is valid that Y G j = T ( P j + μ · u d f ( E j ) ) = Y j + μ T u d f ( E j ) . Irrespective of the number of control channels p, a single gradient experiment was performed.
  Step 3. Calculate J ( P j ) P = 2 N · u d f ( Y G j Y j μ ) , since T T E j = u d f ( T u d f ( E j ) ) provably holds.
  Step 4. Call (8) to update P j .
  Step 5. Go to Step 1 if j < ϵ, otherwise exit with P ϵ .
The output of the model-free EDILC algorithm is the primitive pair { P ϵ = Δ P ˜ * , Y } where the input primitive is  P ˜ * , the output primitive is Y. Several such pairs can be learned, and they can be indexed { P ˜ * [ δ ] , Y [ δ ] } . Note that Y is used instead of Y d in the primitive because of the following reasons: after the algorithm, Y and Y d are close; it is the pair { P ˜ * , Y } that intrinsically encodes the behavior T ( q ) . In addition, P ˜ * is not exactly the P * solving (11) because EDILC stops after a finite number of iterations. Algorithm 1 runs in a mixed-mode: each iteration requires two real-time experiments which are proportional to the number N of samples of the desired trajectory. These computations are repeated for a number of iterations ϵ . We conclude that this algorithm complexity is O ( N 2 ) . Note that the gradient computation has been greatly optimized: it requires a single gradient experiment than the previous variants of the EDILC algorithm in [8] which required a number of gradient experiments equal to the number of CLS inputs, in which case the complexity would have been O ( N 3 ) .
Monotonic convergence is desirable for EDILC and is ensured by the proper selection of step size ξ . The learning level L1 makes the CLS match T O R M ( q ) to some extent; however, a mismatch is expected. The I/O samples from the EDILC trajectories used to learn { P ˜ * , Y } offer the opportunity to identify a linear lower-order model for the CLS, which commonly has low-pass behavior. We denote the model identified from the I/O data as T ˜ ( q ) . Then, T O R M ( q ) and T ˜ ( q ) are both rough CLS estimates, which can be used to derive suitable step sizes for monotonic EDILC convergence. The most conservative selection of ξ is selected by solving the subsequent optimization:
ξ 1 * = arg   m a x ξ i = 1 p ξ 1 , i ,   s . t . ξ 1 , i > 0 ,   I p × p 2 N T O R M ( q ) ξ 1 T O R M T ( 1 q ) < 1 ; ξ 2 * = arg   m a x ξ i = 1 p ξ 2 , i ,   s . t . ξ 2 , i > 0 ,   I p × p 2 N T ˜ ( q ) ξ 2 T ˜ T ( 1 q ) < 1 ; ξ * = d i a g [ ξ i ] ,   ξ i = min ( ξ 1 , i , ξ 2 , i ) ,   i = 1 , p ¯ ,
where . is the norm calculated as the greatest singular value of a t.m. over the frequency domain, while the identity transfer matrix is I p × p . Derivation of the convergence condition for the t.m. norm inequality constraints in (13) is presented in Appendix A.
T O R M ( q ) is usually selected diagonally to ensure decoupling between the control channels through controller design, at least in the steady state, for example, by including integral action in the controllers. For example, for a first-order square reference model of size 2-by-2, the continuous-time unitary-gain t.m. is selected as T O R M ( s ) = d i a g ( e s T m 1 1 + s T 1 , e s T m 2 1 + s T 2 ) (s is the continuous-time Laplace complex argument), where T1 and T2 are the time constants that impose the transient behavior, T m 1 , T m 2 are delays. The real achieved behavior in continuous time can be modeled as T ˜ ( s ) = [ T ˜ 11 T ˜ 12 ; T ˜ 21 T ˜ 22 ] with T ˜ 11 ( s ) = k 11 ( T 11 s + 1 ) e s T m 11 s 2 + b 11 s + c 11 , T ˜ 21 ( s ) = k 21 s ( T 21 s + 1 ) e s T m 21 s 2 + b 21 s + c 21 , T ˜ 12 ( s ) = k 12 s ( T 12 s + 1 ) e s T m 12 s 2 + b 12 s + c 12 , T ˜ 22 ( s ) = k 22 ( T 22 s + 1 ) e s T m 22 s 2 + b 22 s + c 22 where the second-order t.f.s on the anti-diagonal capture any coupling transient that possibly vanishes in time and oscillates at most, while the second-order t.f.s from the main diagonal captures any transient responses that are oscillatory at most and may or may not achieve the desired unit gains. Time delay and NMP zeros must be included in the ORM definition without difficulty, if present. The equivalent discretized causal behavior model T ˜ ( q ) of the CLS T ˜ ( s ) , identifiable from I/O data, is (where m i j are the delay steps)
T ˜ ( q ) = [ b 11 1 q 1 + b 11 2 q 2 1 + a 11 1 q 1 + a 11 2 q 2 q m 11 b 12 0 + b 12 1 q 1 + b 12 2 q 2 1 + a 12 1 q 1 + a 12 2 q 2 q m 12 b 21 0 + b 21 1 q 1 + b 21 2 q 2 1 + a 21 1 q 1 + a 21 2 q 2 q m 21 b 22 1 q 1 + b 22 2 q 2 1 + a 22 1 q 1 + a 22 2 q 2 q m 22 ] .
During ILC operation, T ˜ ( q ) is identified using I/O data either from the nominal or gradient experiment, or from both. The model can be re-estimated or can be continuously refined with every ILC iteration, with no computational burden, owing to the offline experimental nature of ILC. Model precision is not crucial because uncertainty only affects the monotonic convergence condition, and more cautious learning can be attempted. In any case, using low-order models such as (14) is recommended because the ORM control ensures that the CLS approximately matches T O R M ( q ) .
A better initial value than P 0 = Y d in the EDILC law (12) could theoretically be found based on the desired ORM T O R M ( q ) . This could further improve the convergence speed of the subsequent EDILC. The initialization is expressed as:
P 0 = arg   m i n P 1 N ( T O R M P Y d ) T ( T O R M P Y d ) + λ N P T P ,
where the term T O R M P is evaluated by simulation, based on T O R M ( q ) , on a finite-time scenario with length N. However, experimental investigations reveal that this initialization renders oscillatory P 0 near its ends because of noncausal filtering involved in the model-based solution of (15).
The EDILC learns primitives under constraints imposed on the operational variables, such as on the control action a k (and also on its derivative) and on the system (and also the CLS) output y k (and also on its derivative). The reference input constraints are not of interest in this type of tracking problem. These constraints are mostly of the inequality type owing to the underlying dynamical system. They can be directly handled either as hard constraints appended to (7) or as soft constraints by additional terms in the cost function of the optimization (7). It was shown [44] that inequality-type hard constraint handling is still possible with EDILC in an experimental-driven fashion with unknown system dynamics. For primitive-based learning with EDILC, inequality-type constraints on a k are possible, but generally not an issue, as long as the CLS operates in the linearized range resulting from the level L1 learning phase. Any constraints on a k (and possibly on its derivative) may negatively impact the level of L2 learning trajectory tracking performance. Therefore, the constraints on y k are more interesting, mainly for safety reasons such as saturation, which when violated in mechanical systems could lead to permanent damage. The constraints on y k will not be handled explicitly at this level L2, but rather indirectly at the next level L3 and deferred to a later section.
Concluding the level L2 learning aspects discussion, the final tertiary learning level L3 is presented next.

5. The L3 Level—New Desired Trajectories Optimally Tracked by Extrapolating Primitives’ Behavior

Suppose a new desired trajectory Y d is to be tracked; this time, without relearning by EDILC. This amounts to finding the optimal reference input P * to obtain a CLS output that perfectly tracks Y d . A solution to use and extrapolate the behavior of learned primitive pairs was proposed in [8]. Because the solution uses delayed copies of the output primitives from the learned pairs serving as basis functions for approximating Y d , the reason for using EDILC to learn output primitives that have good approximation capability is transparent. Among others, Gaussian shapes are well-known in function approximation theory; therefore, primitive pairs { P ˜ * [ δ ] , Y [ δ ] } where the output primitives Y [ δ ] have a Gaussian-like shape, are commonly learned with EDILC.
In the following, the desired trajectory Y d is assumed to start from zero and end in non-zero values, component-wise. To accommodate for fine approximations whenever Y d ends in a non-zero value and starts from zero with a non-zero slope, both Y d and the copied primitive pairs { P ˜ * [ δ ] , Y [ δ ] } will undergo a time-extension process. The length of Y d is N (same as the primitives’ length), and with no generality loss, N is assumed even. A longer desired trajectory was deferred to a later discussion. The optimal reference input calculation Algorithm 2 is summarized below.
Algorithm 2
  Step 0. Initialize P 0 = Y d , then for each j = 0 , ϵ ¯ call the subsequent steps.
  Step 1. Let y k , 1 d , y k , 2 d , , y k , p d be the components of Y d . In lifted notation, these trajectories are Y 1 d , Y 2 d , , Y p d .
  Step 2. Extend each Y 1 d , , Y p d to size 2N, padding with the first and last sample of Y 1 d , , Y p d , respectively with N/2 values to the right and to the left, respectively. We denote the extended trajectories Y 1 d [ e ] , Y 2 d [ e ] , , Y p d [ e ] .
  Step 3. Arrange Y d [ e ] = [ Y 1 d [ e ] T , Y 2 d [ e ] T , , Y p d [ e ] T ] T 2 p N .
  Step 4. The N-length input primitives ( P ˜ 1 * [ δ ] ) , ( P ˜ 2 * [ δ ] ) , , ( P ˜ p * [ δ ] ) and N-length output primitives ( Y 1 [ δ ] ) , ( Y 2 [ δ ] ) , , ( Y p [ δ ] ) , from the δth  learned pair, are extended to size 2N by padding to the left and right N/2 zeros, respectively. Padding with zeros is employed because both P ˜ * [ δ ] , Y [ δ ] start and end at zero, owing to the Gaussian-like shape of Y [ δ ] . We denote { P ˜ * [ δ e ] , Y [ δ e ] } as the extended primitives.
  Step 5. Make M copies of the randomly occurring δ extended pairs { P ˜ * [ δ e ] , Y [ δ e ] } , let them be called as { P ˜ * [ π e ] , Y [ π e ] } ,   π = 1 , M ¯ .
  Step 6. Delay each component P ˜ 1 * [ π e ] , , P ˜ p * [ π e ] and Y 1 [ π e ] , , Y p [ π e ] of the π pairs by an integer uniform number θ [ N , N ] . Because the delay is a non-circular shift, padding is again required to either end, based on the sign of θ. The value used for padding for a number of |θ| samples is the first or last sample value of the unshifted signals. Let P ˜ 1 * [ θ π e ] , , P ˜ p * [ θ π e ] , Y 1 [ θ π e ] , , Y p [ θ π e ] be the notation for the delayed input and output primitive of each of the π pairs.
  Step 7. Obtain P ˜ * [ θ π e ] = [ P ˜ 1 * [ θ π e ] T , , P ˜ p * [ θ π e ] T ] T 2 p N and Y [ θ π e ] = [ Y 1 [ θ π e ] T , , Y p [ θ π e ] T ] T 2 p N . The basis function matrix is Y [ b ] = [ Y [ θ 1 e ] , , Y [ θ M e ] ] 2 p N × M . A visual summary of the extension with padding, delay with padding, and final concatenation is shown for the output primitives in Figure 1.
  Step 8. Compute w = [ w 1 , , w M ] T M as the minimizer of Y [ b ] w Y d [ e ] 2 with linear least-squares. Each component of w multiplies a specific Y [ θ π e ] .
  Step 9. Based on Theorem 1 from [8], the optimal reference to ensure the tracking of Y d is:
P ˜ * = π = 1 M w π P ˜ * [ θ π e ] ,
  Step 10. In (16), P ˜ * = [ P ˜ 1 * T P ˜ 2 * T P ˜ p * T ] T 2 p N . To obtain N-length channel-wise optimal reference inputs, the p components are clipped in the middle of their intervals.
  Step 11. The clipped optimal references are set as CLS inputs, and the CLS output tracks Y d . The algorithm ends.
Algorithm 2 is entirely performed offline, not impacting the real-time control task. Of course, the planner is allowed sufficient computation time. If the trajectories to be tracked on each subinterval are very long, this may impact performance, but we have tried to solve this particular issue by dividing the longer trajectories in shorter, manageable trajectories. Essentially, Algorithm 2 has two complexity components: the least squares matrices preparation part which has complexity O ( N 3 ) and the least squares regression solution which combines matrix multiplication and the SVD factorization. For this numerical solution, the complexity is O ( M 2 X ) where M is the size of the regression vector w and X = 2 pN is the number of examples ( N is the experiment length and p is the number of CLS inputs and outputs). Hence, the complexity of Algorithm 2 is O ( N 4 ) but since p is usually a small constant, it can be approximated to O ( N 3 ) .
The optimality of P ˜ * in (16) with respect to the tracking performance and its relation to the true optimal solution P * of (11) is analyzed next.
Theorem 1.
When  Y d [ e ]  is approximated with a bounded reconstruction error ζ, then  P *  can be computed with arbitrary precision.
Proof. 
The desired trajectory is decomposed as in Y d [ e ] = π = 1 M w π Y [ θ π e ] + ζ , where ζ is the bounded reconstruction error. The optimal reference input ensuring perfect tracking is P * = ( T T T ) 1 T T Y d [ e ] and is expressed as P * = ( T T T ) 1 T T ( π = 1 M w π Y [ θ π e ] + ζ ) = π = 1 M w π P ˜ * [ θ π e ] + ( T T T ) 1 T T ζ . As ζ 2 0 , then ( T T T ) 1 T T ζ 2 ( T T T ) 1 T T 2   ζ 2 0 follows, where 2 applied to a matrix is the induced 2-norm by a vector, which represents the matrix’s largest singular value, 2 applied to a vector is the Euclidean norm. □
Theorem 1 shows that a good approximation of the desired trajectory in terms of a small error leads to a reference input close to the optimal one. This ensures good tracking performance. To achieve a small approximation error, the number of basis functions can be adjusted along with the desired shape.

Managing Longer Desired Trajectories

Suppose that each signal component y k , 1 d , y k , 2 d , , y k , p d of Yd has length Ω as a multiple of N (padding is again usable when Ω is not a multiple of N). One solution to obtain the desired optimal reference input that achieves CLS output tracking of Yd is to extend the primitive pairs { P ˜ * [ δ e ] , Y [ δ e ] } to length Ω with appropriate padding. In addition, proportionally more M copies of these extended pairs were used. This will increase the size of w and the ill-condition the least squares. An intuitive mitigation is to perform tracking experiments on subintervals of length N as divisions of Yd.
For example, a longer trajectory of length 2N, comprising two shorter trajectories of length N, is displayed in Figure 2 for a single input single output (SISO) system. The end of trajectory 00 ¯ is the start of trajectory 0 B ¯ . A base reference frame in discrete-time is < k 0 y d > with a discrete-time index k.
One solution is to translate the origin of the base reference frame < k 0 y d > to the starting point of each piece of the desired trajectory of length N. In Figure 2, < k 0 y d > is translated to < k 0 y d > . This is equivalent to having the desired trajectory 0 B ¯ starting from the origin. Thus, the planner always computes the optimal reference inputs for a desired trajectory starting in zero initial conditions. Considering that the tracked trajectory 0 A ¯ does not perfectly track the desired one 00 ¯ and there exists a vertical distance ε between the endpoints of 0 A ¯ and 00 ¯ , then in the translated reference frame < k 0 y d > the closed-loop CS will have a non-zero initial condition in the ε -vicinity near 0 . In practice, ε is expected to be small. The CS response due to the non-zero initial condition vanishes over time.
After executing the tracking task on the second subinterval with desired trajectory 0 B ¯ , the execution is translated back in the base frame < k 0 y d > for visualization; afterwards, the reference frame < k 0 y d > is moved in the start of the desired trajectory of the third subinterval, and the entire process repeats itself for all subintervals. After all subinterval tracking tasks are executed, the actual obtained trajectories are translated back to the base reference frame and then concatenated to be presented as a longer tracking task.
Inequality-type constraints on Y can be handled indirectly at the tertiary learning level. The aim of this learning level is to ensure the theoretical perfect tracking of Y d , meaning that Y = Y d . This implies that any constraint on Y d is imposed on Y also. We note that inequality constraints on the rate of Y are again not of particular interest when trajectory tracking is aimed; therefore, the magnitude constraints are more feasible in the lifted form as Y l o w d Y d Y u p d . As a result, the constraints on Y are softly handled.
With the discussed aspects of the L3 learning level, the entire hierarchical three-level framework shown in Figure 3 was validated on a real-world system.

6. Validation Case Study

6.1. The Hybrid Software-Electrical Controlled System

Figure 4 shows the proposed multivariable nonlinear controlled system. The USB-based communication between an Arduino-based board and a computer (PC) manages the pulse-width modulated interface (PWM) used for voltage input action setting and for the analog-to-digital interface to read controlled voltages. Here, the PWM duty cycles are the system’s input actions, a = [ a 1 , a 2 ] T . Their domain [0;1] is shifted to [−0.5,0.5] for symmetric operation. Let Vo1 and Vo2 be the capacitors C1 and C2 voltages, respectively. The system controlled outputs are shifted to [ 2.5 ; 2.5 ]   V for symmetric operation; therefore, they are calculated on the PC as y 1 = V o 1 2.5 , y 2 = V o 2 2.5 . All communication, processing and control tasks are performed using MATLAB software package.
Part of this Hammerstein-like system is electrical (circuit parameters are C 1 = C 2 = 47   µ F , C 3 = 16.5   µ F ,   R 1 = R 2 =   10 k ,), while the input nonlinearities (characterized by a hyperbolic tangent function) and a linear t.f. on the first input action (which has a resonant mode at approximately 7.3 rad/s) run on a PC. The system is zero-order-hold sampled at T s = 0.1   s , rendering the I/O as a k = [ a k , 1 , a k , 2 ] T and yk = [yk,1, yk,2]T. Prior experiments reveal a unit step delay T s to be included when defining the ORM for level L1 learning and also in defining y k d for level L2 learning [45].

6.2. ORM Tracking Learned for Feedback Linearization (Level L1)

First, a PI-type controller of the form a k , 1 = [ 0.15 + 0.1 T s / ( q 1 ) ] ( ρ k , 1 y k , 1 ) , a k , 2 = [ 0.35 + 0.25 T s / ( q 1 )   ] ( ρ k , 2 y k , 2 ) is employed to collect transition sample experiences from the CLS under noisy additive disturbances on the control action inputs [45]. These disturbances were used to enhance exploration. For a prescribed time of 1500 s, sequences of piecewise constant reference inputs with random amplitudes within some tunable domains drive the CLS dynamics. The state of a transition tuple ( s k e x t , a k , s k + 1 e x t ) is s k e x t = [ y k , 1 ,   y k , 2 ,   ,   y k 5 , 1 ,   y k 5 , 2 ,   a k 1 , 1 ,   a k 1 , 2 ,     ,   a k 5 , 1 ,   a k 5 , 2 ,   s k , 1 m ,   s k , 2 m ,   ρ k , 1 ,   ρ k , 2   ] T . Transitions from the time steps when ρ k + 1 = ρ k were considered. Approximately 14,000 transitions were formed and used for AIVI learning.
At the same time, to achieve indirect CLS linearization by ORM tracking, level L1 learning requires the selection of a linear ORM model as T O R M ( s ) = d i a g ( e s T s / ( s + 1 ) , e s T s / ( s + 1 ) ) , in the easy-interpretable continuous-time. Before discretizing T O R M ( s ) , note that it acknowledges the dead-time of one sampling period observed for the system, in addition to the relative degree one induced by the zero-order hold. This is read from the discrete-time I/O sample acquisition of the transition collection phase. The discretized t.f. T O R M ( q ) = d i a g ( 0.0952 q 2 1 0.9048 q 1 , 0.0952 q 2 1 0.9048 q 1 ) is then expressed as a state-space model resulting from an observable canonical transformation. The resulting 4th order state-space uses two extra states denoted with ϖ , owing to the time delay of each t.f. from the T O R M ( q ) . The model used is ([45]):
{ [ s k + 1 , 1 m ϖ k + 1 , 1 s k + 1 , 2 m ϖ k + 1 , 2 ] = [ 0.9048 0.0952 0 0 0 0 0 0 0 0 0.9048 0.0952 0 0 0 0 ] [ s k , 1 m ϖ k , 1 s k , 2 m ϖ k , 2 ] + [ 0 0 1 0 0 0 0 1 ] [ ρ k , 1 ρ k , 2 ] , y k m = [ y k , 1 m y k , 2 m ] T = [ 1 , 0 , 1 , 0 ] × [ s k , 1 m ϖ k , 1 s k , 2 m ϖ k , 2 ] T .
A single hidden layer feed-forward controller NN of size 28–15–2 is learned to output a k for the input s k e x , using the AIVI iteration steps, which implies NN training for both the controller and for a Q-function approximator NN, having a similar architecture but size 30–100–1. The CLS with the converged AIVI NN controller after six iterations of the AIVI approach is shown the ORM tracking ability in Figure 5. To check this convergence, the controller resulting after each iteration of the AIVI is tested on a scenario in which sequences of piecewise constant reference inputs drive the CLS. The ORM tracking performance was measured with a regularized finite-length version of V O R M , defined as V t e s t = 1 / N · V O R M N , with N = 2000 . The convergence is illustrated in Figure 6. It occurs relatively quickly, and the best controller is shown in Figure 5.
This concludes the L1 level indirect CLS linearization by matching the behavior of T O R M ( q ) . The subsequent EDILC L2 level learning benefits from this assumption.

6.3. Two Primitives Learned with EDILC (Level L2)

A number of δ = 2 primitives of length N = 400 samples are learned using the EDILC Algorithm 1 with P 0 = Y d , μ = 0.01 , ϵ = 30 , and the learning rate ξ = d i a g ( 299.13 ,   281.35 ) was determined by solving (13) using T O R M ( q ) and an approximate CLS t.f. identified as [45]
T ˜ ( q ) = [ 0.1278 q 2 1 0.77258 q 1 0.1126 q 2 0.02189 q 2 0.02101 q 3 1 1.213 q 1 + 0.2206 q 2 0.0756 q 2 0.05602 q 3 1 0.8129 q 1 + 0.7474 q 2 0.1554 q 2 1 0.4374 q 1 0.422 q 2 ] .
For the first primitive { P ˜ * [ 1 ] , Y [ 1 ] } , the desired trajectory spanning N = 400 time steps ( k = 0 , 399 ¯ ) is y k , 1 d [ 1 ] = χ 1 exp ( χ 2 ( k χ 3 ) 2 ) , ( χ 1 = 2 , χ 2 = 0.002 , χ 3 = 200 ) and y k , 2 d [ 1 ] = 0 , whereas the second primitive { P ˜ * [ 2 ] , Y [ 2 ] } is characterized by the same length and by y k , 1 d [ 2 ] = 0 , y k , 2 d [ 2 ] = y k , 1 d [ 1 ] . That is, for each primitive, one desired output is Gaussian, while the other is zero. The first primitive { P ˜ * [ 1 ] , Y [ 1 ] } learned using EDILC is shown in Figure 7.

6.4. Optimal Tracking Predicted from Primitives (Level L3)

At level L3, it is desired to track y k , 1 d = max { 0.7 , min [ 0.7 , χ 1 sin ( χ 2 k T s ) sin ( χ 3 k T s ) ] } ( χ 1 = 1.1 , χ 2 = 0.6 , χ 3 = 0.02 ) and y k , 2 d = χ 1 sin ( χ 2 k T s ) sin ( χ 3 k T s ) , ( χ 1 = 0.9 , χ 2 = 0.01 , χ 3 = 0.3 ) for k = 0 , Ω 1 ¯ , where Ω = 1600 = 4 × N . Trajectory y k , 1 d is clipped in amplitude with min/max operators to induce an output constraint. Then, the CLS output Y will be constrained because, by the primitive-based mechanism, level L3 ensures that Y = Y d . Then, Y d is divided into four subintervals, and the optimal reference input calculation by the primitive-based approach together with the trajectory tracking takes place on one subinterval of length N at a time. Each subinterval is offset to start the tracking at zero intervals. The parameters in Algorithm 2 are N = 400 , δ = 2 primitive pairs, M = 400 copies of the two primitives are employed, θ [ 400 , 400 ] , to fulfill Theorem 1.
Figure 8 displays the extended, padded, delayed, once-again padded, and concatenated output primitives Y 1 [ θ π e ] , Y 2 [ θ π e ] . Notably, the basis function matrix is Y [ b ] 1600 × 400 . The tracking results for each subinterval are re-offset in the original reference frame and displayed in Figure 9 where, for comparisons, the tracking results obtained with P 1 = Y 1 d , P 2 = Y 2 d are also plotted. It is obvious that tracking is more accurate with the primitive-based approach (red line vs. blue line) than with respect to the case when it is not used (black line vs. blue line). The latter case corresponds to using only the CLS linearized with L1 level, with the reference inputs being set as usually, leaving the control system to reactively ensure trajectory tracking. Primitive-based tracking also better satisfies the constraints (in a soft, not a hard sense), which is enforced on the first axis only.
The tracking MSEs obtained with the primitive-based mechanism for the first and second axes are 0.44·10–3 and 0.31·10–3, respectively. When P 1 = Y 1 d ,   P 2 = Y 2 d , the tracking MSEs for the first and second axis are 6.50·10–3 and 3.75·10–3, respectively. A ten-fold improvement in tracking was realized using the primitive-based approach.
The observed improvement is of no surprise, since the primitive-based tracking solution behaves in an anticipative manner, whereas the tracking based on the L1 control only is “reactive” and only decides based on the provided reference input. The anticipative behavior of the primitive-based solution inherits its character from the level L2 subjected to the ILC concept. It proves that an optimal behavior can be extrapolated successfully to new trajectories to be tracked, without requiring learning by repetitions with ILC.

7. Implementation Issues and Analysis of the Primitive-Based Framework

One of the most important aspects affecting the entire hierarchical learning performance is the CLS linearity assumption achieved at level L1. Therefore, it is critical to ensure that the CLS matches the linear ORM as accurately as possible. While the mismatch does not affect the L2 level too much because the EDILC displays good learning ability even for some nonlinear CLS (the perturbation-based gradient experiment performed in the vicinity of the nominal trajectory where linearity holds), the level L3 operations have proven to be more sensitive to this mismatch because the optimal reference reconstruction relies on the superposition principle. Therefore, all efforts in learning AIVI NN controllers are justified, both in terms of NN complexity and in terms of the required database size and exploration effort. Better parameterization (more specifically, a linear one using polynomials as basis functions, as opposed to a more generic, such as, e.g., a NN) of the Q-function may need fewer transition samples for learning; however, the downside is that it requires more insight about the underlying controlled system. This insight is still needed in terms of relative degree, system order, minimum-phase character, and in matching the system bandwidth with the linear ORM’s bandwidth. Together, these impact the virtual state size (hence the learning space complexity and the function approximators architectural complexity). With careful AIVI design, the NN should be preferred because the underlying training management (e.g., early stopping) allows the more complex NN architectures to use just as much as needed in terms of approximation capacity. Another advantage is automatic feature selection, which is inherently encoded in the NN layers. Yet another aspect of offline AIVI learning is that it does not require online computational power (as in e.g., [46]). The resulting NN controllers take up to a few hours to learn offline; however, they are easily implementable afterwards on hardware with no special computing requirements for typical mechatronic systems. The efforts needed for this L1 level control, also regarded as limitations, are:
a good estimation of the controlled system’s relative degree. A too small degree is equivalent to partial observability and will incur a performance degradation in achieving linearized ORM tracking behavior. A higher degree will add information redundancy which is not a real issue except for the computational costs involving the resulted controller.
some additional required information leading to a degree of subjectiveness in the design: the underlying system’s minimum-phase character, and its bandwidth to be matched with the linear ORM’s bandwidth. This information quantification necessitates some experience from the designer and some insight into the process phenomena.
the controller computational complexity is not seen as an important issue. Indeed, there is more complexity than with a PI/PID controller with few parameters; however, the used NN architectures are shallow, allowing for less than 1ms inference time on a wide range of embedded devices, which would extend the application range of the control. The benefits in terms of nonlinear control performance well balance the costs.
Concerning level L2 learning, it should be used mostly in cases where P 0 = Y d does not render the output Y close to the basis function shape imposed by Y d . Otherwise, the impact on level L3 is not damaging, since essentially just about any shape of Y will eventually have some approximation capacity. The major question at this stage is how to learn primitives to be useful for level L3. This is discussed among the aspects of the next and final learning levels. The L2 level design is dependent only on the initial solution P 0 and the linaerity assumption. The limitations associated with these are
the L2 level CLS requires resetting conditioned which for some system may be more restrictive. ILC solutions for systems without perfect resetting such stochastic resetting exists, but not in a model-free style. However, for most mechatronic systems, this is not an issue.
The initialization for P 0 can be optimized using, e.g., (15), for a neal-optimal approximate model-based initialization. This would require fewer EDILC iterations. On the other hand, model-based initialization based on non-causal filtering showed some oscillations at the beginning and at the end of P 0 hence it was not applied here.
The linearity assumption about the CLS. This must be ensured with the L1 level virtual state-feedback control. There is a certain degree of tolerance, as EDILC was applied before on smooth nonlinear systems and the underlying linear system’s theory showed effectiveness and learnability.
Most aspects concern level L3 implementation details. The discussion about how to learn the proper primitives starts backward, with the new desired trajectory Y d . From Algorithm 2’s steps, Y d undergoes a transformation process (extrapolation plus padding, then concatenation), resulting in Y d [ e ] as a single signal. The same is valid when obtaining Y [ θ π e ] = [ Y 1 [ θ π e ] T , , Y p [ θ π e ] T ] T , except for the additional delay process involved. It is obvious that the final Y [ θ π e ] should resemble a basis function (here, Gaussians were used, also known as radial basis functions), which means that only one output channel should be Gaussian and the rest should be zero. Therefore, the practical rule is that if the CLS has p channels, then a number of p primitives should be learned, each time with a different output channel being a Gaussian, while the other outputs are zero.
Learning Gaussian-like shapes as output primitive trajectories is advantageous because a Gaussian is described with only two design parameters: mean (center) and variance (width). With the function approximation, the width of the Gaussian-like output primitive influences the approximation capacity in relation to the shape of Y d : if Y d is non-smooth, more copies of output primitives should be used when their widths are small. Conversely, when Y d is smooth, a good approximation is achievable with fewer copied, wider output primitives. The Gaussian polarity (pointing upward or downward) is not important as long as the CLS linearity holds, because a negative weighting coefficient suitably reverses polarity and helps with the approximation.
Typically, the number of copies (parameter M) does not significantly affect the approximation quality. The number of delay/advance steps for each copied primitive (parameter θ ) should be chosen from an interval such that the copied and time-shifted output primitives span the entire approximation range of interest for Y d . In summary, the limitations at this particular L3 level include:
correlating the shape of the basis functions with the shape of the desired trajectory to be tracked on each subinterval.
selecting the adequate number of copies for the learned primitives. While more is better in the sense that the approximation quality does not increase past a certain number M of copies, it could increase the computational cost for the underlying regression and reference input recomposition. We still argue that the computations are performed online and manageable.
the tracking has to be of good quality for each subinterval, to avoid large deviations at the end, since this would shift the initial point for trajectory tracking of the next subinterval. This was found to depend largely on the ORM matching (CLS linearization quality) at the level L1.
Concluding, the discussed aspects also reveal the sensitive issues across the proposed framework. At level L1, it is important to ensure exceptional ORM matching for level L3 to accept behavior extrapolation. This is related to AIVI learning, which is strongly tied to exploration quality. This is an open issue in ADP/RL and requires further investigation. At level L2, with an impact on level L3, the basis function shape of the output primitives with respect to the desired trajectory shape is another trade-off and an open issue.

8. Conclusions

The three-level hierarchical cognitive-based learning framework validated here uses several tools from control systems and from machine learning to extrapolate an optimal tracking behavior to new tracking situations by avoiding explicit knowledge of the underlying system dynamics. This framework, inspired by intelligent life, could make progress toward more autonomous and adaptive learning control systems, as required by modern paradigms. It includes all traits of intelligence: memorization and learning from past experiences, generalizing behavior outside of the current knowledge base, and self-improvement in terms of specified performance. The framework’s validation on more complex and different systems will prove its potential and may lead to faster adoption and industrial implementation.

Author Contributions

Conceptualization, M.-B.R. and T.L.; methodology, M.-B.R.; software, M.-B.R. and T.L.; validation M.-B.R. and T.L.; formal analysis, M.-B.R. and T.L.; investigation, M.-B.R. and T.L.; writing—original draft preparation, M.-B.R. and T.L.; writing—review and editing, M.-B.R. and T.L.; visualization, T.L.; supervision, M.-B.R.; project administration, M.-B.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant of the Ministry of Research, Innovation and Digitization, CNCS/CCCDI—UEFISCDI, project number PN-III-P1-1.1-TE-2019-1089, within PNCDI III.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

Acronyms
CLSclosed loop system
CScontrol system
EDILCexperiment-driven Iterative Learning Control
ADPAdaptive Dynamic Programming
IVIIterative Value Iteration
NNneural network
RLReinforcement Learning
ORMOutput Reference Model
NMP non-minimum phase
I/Oinput/output
VSFRTVirtual State-Feedback Reference Tuning
MIMOmultiple input multiple output
SISOsingle input single output
PWMpulse width modulation
PCpersonal computer
MSEmean squared error

Appendix A

Since E ( P ) 2 2 = E T E = E 1 T E 1 + + E p T E p , the gradient J / P is
J P = [ J P 1 J P p ] = 1 N [ i = 1 p E i T E i P 1 i = 1 p E i T E i P p ] .
It is provable by matrix calculus that E i T E i P j = 2 ( T j , i ) T E i , leading to the c.f. gradient
J P = [ J P 1 J P p ] = 2 N [ ( T 1 , 1 ) T ( T 1 , 2 ) T ( T 1 , p ) T ( T 2 , 1 ) T ( T 2 , 2 ) T ( T 2 , p ) T ( T p , 1 ) T ( T p , 2 ) T ( T p , p ) T ] T T [ E 1 E 2 E p ] .
The gradient-based search (8) is then expanded in a mathematical form for each reference input as
[ P 1 j + 1 P 2 j + 1 P p j + 1 ] = [ P 1 j P 2 j P p j ] 2 N ξ   [ ( T 1 , 1 ) T ( T 1 , 2 ) T ( T 1 , p ) T ( T 2 , 1 ) T ( T 2 , 2 ) T ( T 2 , p ) T ( T p , 1 ) T ( T p , 2 ) T ( T p , p ) T ] [ E 1 j E 2 j E p j ] .
If T i , l corresponds to T i , l ( q ) , then ( T i , l ) T corresponds to the t.f. obtained by replacing q with 1/q in T i , l ( q ) . Herein, 1/q must not be mistaken for the t.f. argument q–1 as an alternate notation to q. Replacing q qith 1/q in a causal t.f. leads to a noncausal one. Extending each vector and matrices from (A3) for N , we write in the time domain as
[ ρ k , 1 j + 1 ρ k , 2 j + 1 ρ k , p j + 1 ] = [ ρ k , 1 j ρ k , 2 j ρ k , p j ] 2 N ξ [ T 1 , 1 ( 1 q ) T 1 , 2 ( 1 q ) T 1 , p ( 1 q ) T 2 , 1 ( 1 q ) T 2 , 2 ( 1 q ) T 2 , p ( 1 q ) T p , 1 ( 1 q ) T p , 2 ( 1 q ) T p , p ( 1 q ) ] T T ( 1 q ) [ e k , 1 j e k , 2 j e k , p j ] .
Using the previous equation in the tracking error dynamics across the iterations it follows that
[ e k , 1 j + 1 e k , 2 j + 1 e k , p j + 1 ] = T ( q ) [ ρ k , 1 j + 1 ρ k , 2 j + 1 ρ k , p j + 1 ] [ y k , 1 d y k , 2 d y k , p d ] = T ( q ) [ [ ρ k , 1 j ρ k , 2 j ρ k , p j ] 2 N ξ T T ( 1 q ) [ e k , 1 j e k , 2 j e k , p j ] ] [ y k , 1 d y k , 2 d y k , p d ] = ( I p × p 2 N T ( 1 q ) ξ T T ( 1 q ) ) [ e k , 1 j e k , 2 j e k , p j ] ,
where the frequency domain properties of I p × p 2 N T ( 1 q ) ξ T T ( 1 q ) are used for the convergence analysis.

References

  1. Wolpert, D.M.; Diedrichsen, J.; Flanagan, J.R. Principles of sensorimotor learning. Nat. Rev. Neurosci. 2011, 12, 739–751. [Google Scholar] [CrossRef] [PubMed]
  2. Kawamura, S.; Sakagami, N. Analysis on dynamics of underwater robot manipulators basing on iterative learning control and time-scale transformation. In Proceedings of the IEEE International Conference on Robotics and Automation, Washington, DC, USA, 11–15 May 2002; IEEE: Piscataway, NJ, USA, 2002; Volume 2, pp. 1088–1094. [Google Scholar]
  3. Ijspeert, A.J.; Nakanishi, J.; Schaal, S. Movement imitation with nonlinear dynamical systems in humanoid robots. In Proceedings of the IEEE International Conference on Robotics and Automation, Washington, DC, USA, 11–15 May 2002; IEEE: Piscataway, NJ, USA, 2002; Volume 2, pp. 1398–1403. [Google Scholar]
  4. Hoelzle, D.J.; Alleyne, A.G.; Wagoner Johnson, A.J. Bumpless transfer for a flexible adaptation of iterative learning control. In Proceedings of the American Control Conference, Can Francisco, CA, USA, 29 June–1 July 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 4305–4311. [Google Scholar]
  5. Schöllig, A.; Hehn, M.; Lupashin, S.; D’Andrea, R. Feasiblity of motion primitives for choreographed quadrocopter flight. In Proceedings of the American Control Conference, Can Francisco, CA, USA, 29 June–1 July 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 3843–3849. [Google Scholar]
  6. Grymin, D.J.; Neas, C.B.; Farhood, M. A hierarchical approach for primitive-based motion planning and control of autonomous vehicles. Rob. Auton. Syst. 2014, 62, 214–228. [Google Scholar] [CrossRef]
  7. Wang, H.; Zou, Q. B-spline-decomposition-based approach to multiaxis trajectory tracking: Nanomanipulation example. IEEE Trans. Control Syst. Technol. 2014, 22, 1573–1580. [Google Scholar] [CrossRef]
  8. Radac, M.B.; Precup, R.E.; Petriu, E.M. Model-free primitive-based iterative learning control approach to trajectory tracking of MIMO systems with experimental validation. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 2925–2938. [Google Scholar] [CrossRef] [PubMed]
  9. Li, J.; Li, Z.; Li, X.; Feng, Y.; Hu, Y.; Xu, B. Skill learning strategy based on dynamic motion primitives for human-robot cooperative manipulation. IEEE Trans. Cogn. Dev. Syst. 2021, 13, 105–117. [Google Scholar] [CrossRef]
  10. Kim, Y.L.; Ahn, K.H.; Song, J.B. Reinforcement learning based on movement primitives for contact tasks. Robot. Comput. Integr. Manuf. 2020, 62, 101863. [Google Scholar] [CrossRef]
  11. Camci, E.; Kayacan, E. Learning motion primitives for planning swift maneuvers of quadrotor. Auton. Robots 2019, 43, 1733–1745. [Google Scholar] [CrossRef]
  12. Yang, C.; Chen, C.; He, W.; Cui, R.; Li, Z. Robot learning system based on adaptive neural control and dynamic movement primitives. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 777–787. [Google Scholar] [CrossRef]
  13. Huang, R.; Cheng, H.; Qiu, J.; Zhang, J. Learning physical human-robot interaction with coupled cooperative primitives for a lower exoskeleton. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1566–1574. [Google Scholar] [CrossRef]
  14. Liu, H.; Li, J.; Ge, S. Research on hierarchical control and optimisation learning method of multi-energy microgrid considering multi-agent game. IET Smart Grid 2020, 3, 479–489. [Google Scholar] [CrossRef]
  15. Van, N.D.; Sualeh, M.; Kim, D.; Kim, G.W. A hierarchical control system for autonomous driving towards urban challenges. Appl. Sci. 2020, 10, 3543. [Google Scholar] [CrossRef]
  16. Jiang, W.; Yang, C.; Liu, Z.; Liang, M.; Li, P.; Zhou, G. A hierarchical control structure for distributed energy storage system in DC micro-grid. IEEE Access 2019, 7, 128787–128795. [Google Scholar] [CrossRef]
  17. Merel, J.; Botvinick, M.; Wayne, G. Hierarchical motor control in mammals and machines. Nat. Commun. 2019, 10, 5489. [Google Scholar] [CrossRef] [Green Version]
  18. Wu, B.; Gupta, J.K.; Kochenderfer, M. Model primitives for hierarchical lifelong reinforcement learning. Auton. Agent. Multi. Agent. Syst. 2020, 34, 28. [Google Scholar] [CrossRef]
  19. Chi, R.; Hou, Z.; Jin, S.; Huang, B. An improved data-driven point-to-point ILC using additional on-line control inputs with experimental verification. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 687–696. [Google Scholar] [CrossRef]
  20. Chi, R.; Zhang, H.; Huang, B.; Hou, Z. Quantitative data-driven adaptive iterative learning control: From trajectory tracking to point-to-point tracking. IEEE Trans. Cybern. 2020, 1–15. [Google Scholar] [CrossRef]
  21. Zhang, J.; Meng, D. Convergence analysis of saturated iterative learning control systems with locally lipschitz nonlinearities. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 4025–4035. [Google Scholar] [CrossRef]
  22. Li, X.; Chen, S.L.; Teo, C.S.; Tan, K.K. Data-based tuning of reduced-order inverse model in both disturbance observer and feedforward with application to tray indexing. IEEE Trans. Ind. Electron. 2017, 64, 5492–5501. [Google Scholar] [CrossRef]
  23. Madadi, E.; Söffker, D. Model-free control of unknown nonlinear systems using an iterative learning concept: Theoretical development and experimental validation. Nonlinear Dyn. 2018, 94, 1151–1163. [Google Scholar] [CrossRef]
  24. Zhang, D.; Wang, Z.; Masayoshi, T. Neural-network-based iterative learning control for multiple tasks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4178–4190. [Google Scholar] [CrossRef] [PubMed]
  25. Zhao, H.; Peng, L.; Yu, H. Data driven distributed bipartite consensus tracking for nonlinear multiagent systems via iterative learning control. IEEE Access 2020, 8, 144718–144729. [Google Scholar] [CrossRef]
  26. Fu, H.; Chen, X.; Wang, W.; Wu, M. MRAC for unknown discrete-time nonlinear systems based on supervised neural dynamic programming. Neurocomputing 2020, 384, 130–141. [Google Scholar] [CrossRef]
  27. Wang, W.; Chen, X.; Fu, H.; Wu, M. Data-driven adaptive dynamic programming for partially observable nonzero-sum games via Q-learning method. Int. J. Syst. Sci. 2019, 50, 1338–1352. [Google Scholar] [CrossRef]
  28. Na, J.; Lv, Y.; Zhang, K.; Zhao, J. Adaptive identifier-critic-based optimal tracking control for nonlinear systems with experimental validation. IEEE Trans. Syst. Man Cybern. Syst. 2020, 1–14. [Google Scholar] [CrossRef]
  29. Sardarmehni, T.; Heydari, A. Sub-optimal switching in anti-lock brake systems using approximate dynamic programming. IET Control Theory Appl. 2019, 13, 1413–1424. [Google Scholar] [CrossRef]
  30. Liu, Y.; Zhang, H.; Yu, R.; Xing, Z. H∞ tracking control of discrete-time system with delays via data-based adaptive dynamic programming. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 4078–4085. [Google Scholar] [CrossRef]
  31. Perrusquia, A.; Yu, W. Neural H2 control using continuous-time reinforcement learning. IEEE Trans. Cybern. 2020, 1–10. [Google Scholar] [CrossRef] [PubMed]
  32. Huang, M.; Liu, C.; He, X.; Ma, L.; Lu, Z.; Su, H. Reinforcement learning-based control for nonlinear discrete-time systems with unknown control directions and control constraints. Neurocomputing 2020, 402, 50–65. [Google Scholar] [CrossRef]
  33. Buşoniu, L.; de Bruin, T.; Tolić, D.; Kober, J.; Palunko, I. Reinforcement learning for control: Performance, stability, and deep approximators. Annu. Rev. Control 2018, 46, 8–28. [Google Scholar] [CrossRef]
  34. Chen, C.; Sun, W.; Zhao, G.; Peng, Y. Reinforcement Q-Learning incorporated with internal model method for output feedback tracking control of unknown linear systems. IEEE Access 2020, 8, 134456–134467. [Google Scholar] [CrossRef]
  35. Treesatayapun, C. Knowledge-based reinforcement learning controller with fuzzy-rule network: Experimental validation. Neural Comput. Appl. 2020, 32, 9761–9775. [Google Scholar] [CrossRef]
  36. Lewis, F.L.; Vamvoudakis, K.G. Reinforcement learning for partially observable dynamic processes: Adaptive dynamic programming using measured output data. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2011, 41, 14–25. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, Z.; Liu, D. Data-based controllability and observability analysis of linear discrete-time systems. IEEE Trans. Neural Netw. 2011, 22, 2388–2392. [Google Scholar] [CrossRef]
  38. Ni, Z.; He, H.; Zhong, X. Experimental studies on data-driven heuristic dynamic programming for POMDP. In Frontiers of Intelligent Control and Information Processing; Liu, D., Alippi, C., Zhao, D., Zhang, H., Eds.; World Scientific: Singapore, 2014; Chapter 3; pp. 83–106. [Google Scholar]
  39. De Bruin, T.; Kober, J.; Tuyls, K.; Babuska, R. Integrating state representation learning into deep reinforcement learning. IEEE Robot. Autom. Lett. 2018, 3, 1394–1401. [Google Scholar] [CrossRef]
  40. Yang, Q.; Jagannathan, S. Reinforcement learning controller design for affine nonlinear discrete-time systems using online approximators. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2012, 42, 377–390. [Google Scholar] [CrossRef]
  41. Radac, M.B.; Lala, T. Robust control of unknown observable nonlinear systems solved as a zero-sum game. IEEE Access 2020, 8, 214153–214165. [Google Scholar] [CrossRef]
  42. Radac, M.B.; Borlea, A.I. Virtual state feedback reference tuning and value iteration reinforcement learning for unknown observable systems control. Energies 2021, 14, 1006. [Google Scholar] [CrossRef]
  43. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  44. Radac, M.B.; Precup, R.E. Data-based two-degree-of-freedom iterative control approach to constrained non-linear systems. IET Control Theory Appl. 2015, 9, 1000–1010. [Google Scholar] [CrossRef] [Green Version]
  45. Lala, T.; Radac, M.-B. Learning to extrapolate an optimal tracking control behavior towards new tracking tasks in a hierarchical primitive-based framework. In Proceedings of the 2021 29th Mediterranean Conference on Control and Automation, Virtual Conference, 22–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 421–427. [Google Scholar]
  46. Radac, M.B.; Precup, R.E. Data-driven model-free tracking reinforcement learning control with VRFT-based adaptive actor-critic. Appl. Sci. 2019, 9, 1807. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Extension with padding, delay with padding and final concatenation concepts, in the space of two output primitives. Similar operations are performed for the input primitives.
Figure 1. Extension with padding, delay with padding and final concatenation concepts, in the space of two output primitives. Similar operations are performed for the input primitives.
Mathematics 09 02752 g001
Figure 2. A trajectory consisting of two connected curves 0 0 ¯ and 0 B ¯ ; moving the reference frame to 0′ makes the trajectory 0 B ¯ start in zero.
Figure 2. A trajectory consisting of two connected curves 0 0 ¯ and 0 B ¯ ; moving the reference frame to 0′ makes the trajectory 0 B ¯ start in zero.
Mathematics 09 02752 g002
Figure 3. The hierarchical learning on levels L1–L3. Possible disturbances dk may affect the nonlinear controlled system.
Figure 3. The hierarchical learning on levels L1–L3. Possible disturbances dk may affect the nonlinear controlled system.
Mathematics 09 02752 g003
Figure 4. Conceptual description of the controlled system.
Figure 4. Conceptual description of the controlled system.
Mathematics 09 02752 g004
Figure 5. Tracking ORM results: a k , 1 , y k , 1 , a k , 2 , y k , 2 with AIVI (blue); a k , 1 , y k , 1 , a k , 2 , y k , 2 with PI control (black); y k , 1 m , y k , 2 m (red).
Figure 5. Tracking ORM results: a k , 1 , y k , 1 , a k , 2 , y k , 2 with AIVI (blue); a k , 1 , y k , 1 , a k , 2 , y k , 2 with PI control (black); y k , 1 m , y k , 2 m (red).
Mathematics 09 02752 g005
Figure 6. ORM tracking performance Vtest throughout AIVI, measured on the test scenario for 200 iterations.
Figure 6. ORM tracking performance Vtest throughout AIVI, measured on the test scenario for 200 iterations.
Mathematics 09 02752 g006
Figure 7. Learning results for { P * [ 1 ] , Y[1]}: y k , 1 d [ 1 ] and y k , 2 d [ 1 ] (blue); y k , 1 0 , y k , 2 0 , ρ k , 1 0 , ρ k , 2 0 (black); y k , 1 30 , y k , 2 30 , ρ k , 1 * , ρ k , 2 * (red).
Figure 7. Learning results for { P * [ 1 ] , Y[1]}: y k , 1 d [ 1 ] and y k , 2 d [ 1 ] (blue); y k , 1 0 , y k , 2 0 , ρ k , 1 0 , ρ k , 2 0 (black); y k , 1 30 , y k , 2 30 , ρ k , 1 * , ρ k , 2 * (red).
Mathematics 09 02752 g007
Figure 8. The 400 extended, padded, delayed and finally padded copies of Y 1 [ θ π e ] , Y 2 [ θ π e ] , zoomed in the sample index range [600;1000].
Figure 8. The 400 extended, padded, delayed and finally padded copies of Y 1 [ θ π e ] , Y 2 [ θ π e ] , zoomed in the sample index range [600;1000].
Mathematics 09 02752 g008
Figure 9. The recomposed tracking result on the long desired trajectory: y k , 1 d and y k , 2 d in blue; y k , 1 and y k , 2 (in red) obtained with optimal reference inputs ρ k , 1 * and ρ k , 2 * , respectively; y k , 1 and y k , 2 (in black) are with ρ k , 1 = y k , 1 d , ρ k , 2 = y k , 2 d .
Figure 9. The recomposed tracking result on the long desired trajectory: y k , 1 d and y k , 2 d in blue; y k , 1 and y k , 2 (in red) obtained with optimal reference inputs ρ k , 1 * and ρ k , 2 * , respectively; y k , 1 and y k , 2 (in black) are with ρ k , 1 = y k , 1 d , ρ k , 2 = y k , 2 d .
Mathematics 09 02752 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Radac, M.-B.; Lala, T. Hierarchical Cognitive Control for Unknown Dynamic Systems Tracking. Mathematics 2021, 9, 2752. https://doi.org/10.3390/math9212752

AMA Style

Radac M-B, Lala T. Hierarchical Cognitive Control for Unknown Dynamic Systems Tracking. Mathematics. 2021; 9(21):2752. https://doi.org/10.3390/math9212752

Chicago/Turabian Style

Radac, Mircea-Bogdan, and Timotei Lala. 2021. "Hierarchical Cognitive Control for Unknown Dynamic Systems Tracking" Mathematics 9, no. 21: 2752. https://doi.org/10.3390/math9212752

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop