Next Article in Journal
Existence of Solution, Filippov’s Theorem and Compactness of the Set of Solutions for a Third-Order Differential Inclusion with Three- Point Boundary Conditions
Next Article in Special Issue
Asynchronous Iterations of Parareal Algorithm for Option Pricing Models
Previous Article in Journal
Advanced Expected Tail Loss Measurement and Quantification for the Moroccan All Shares Index Portfolio
Previous Article in Special Issue
A Simple Formula for the Hilbert Metric with Respect to a Sub-Gaussian Cone
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Chaotic Itinerancy in Random Dynamical System Related to Associative Memory Models

by
Ricardo Bioni Liberalquino
1,
Maurizio Monge
1,
Stefano Galatolo
2,* and
Luigi Marangio
3
1
Institute of Mathematics, Universidade Federal do Rio de Janeiro, 21941-901 Rio de Janeiro, Brazil
2
Dipartimento di Matematica, Università di Pisa, 56126 Pisa, Italy
3
Femto-ST Institute, Université de Université Bourgogne Franche-Comté, 21000 Dijon, France
*
Author to whom correspondence should be addressed.
Mathematics 2018, 6(3), 39; https://doi.org/10.3390/math6030039
Submission received: 31 January 2018 / Revised: 26 February 2018 / Accepted: 28 February 2018 / Published: 7 March 2018

Abstract

:
We consider a random dynamical system arising as a model of the behavior of a macrovariable related to a more complicated model of associative memory. This system can be seen as a small (stochastic and deterministic) perturbation of a determinstic system having two weak attractors which are destroyed after the perturbation. We show, with a computer aided proof, that the system has a kind of chaotic itineracy. Typical orbits are globally chaotic, while they spend a relatively long time visiting the attractor’s ruins.

1. Introduction

Chaotic itinerancy is a concept used to refer to a dynamical behavior in which typical orbits visit a sequence of regions of the phase space called “quasi attractors” or “attractor ruins” in some irregular way. Informally speaking, during this itinerancy, the orbits visit a neighborhood of a quasi attractor (the attractor ruin) with a relatively regular and stable motion, for relatively long times and then the trajectory jumps to another quasi attractor of the system after a relatively small chaotic transient. This behavior was observed in several models and experiments related to the dynamics of neural networks and related to neurosciences (see [1]). In this itinerancy, the visit near some attractors was associated to the appearance of some macroscopic aspect of the system, like the emergence of a perception or a memory, while the chaotic iterations from a quasi attractor to another are associated to an intelligent (history dependent with trial and errors) search for the next thought or perception (see [2]). This kind of phenomena was observed in models of the neural behavior which are fast-slow systems or random dynamical systems (sometimes modeling as a random system the fast-slow behavior).
As far as we know the concept did not have a complete mathematical formalization though in [3] some mathematical scenarios are presented, showing some situations in which these phenomena may appear. Often this phenomenon is associated to the presence of some kind of “weak” attractor, as Milnor type attractors (see e.g., [3] or [4]) in the system and to small perturbations allowing typical trajectories to escape the attractor. Chaotic itinerancy was found in many applied contexts, but a systematic treatment of the literature is out of the scope of this paper, for which we invite the reader to consult [1] for a wider introduction to the subject and to its literature.
In this paper, our goal is to investigate and illustrate this concept from a mathematical point of view with some meaningful examples. We consider a simple one-dimensional random map derived from the literature on the subject (see Section 3). This map is a relatively simple discrete time random dynamical system on the interval, suggested by the behavior of certain macrovariables in the neural networks studied in [5], modeling the neocortex with a variant of Hopfield’s asynchronous recurrent neural network presented in [6]. In Hopfield’s network, memories are represented by stable attractors and an unlearning mechanism is suggested in [7] to account for unpinning of these states (see also, e.g., [8]). In the network presented in [5], however, these are replaced by Milnor attractors, which appear due to a combination of symmetrical and asymmetrical couplings and some resetting mechanism. A similar map is also obtained in [9], in the context of the BvP neuron driven by a sinusodial external stimulus. They belong to a family known as Arnold circle maps (named after [10]), which are useful in physiology (see [11] (Equation (3))). The simple model we study shares with the more complicated models from which it is derived the presence of quasi attractors (it can be seen as a stochastic perturbation of a system with Milnor attractors), its undestanding can bring some light on the mathematical undestanding of chaotic itinearacy.
The model we consider is made by a deterministic map T on the circle perturbed by a small additive noise. For a large enough noise, its associated random dynamical system exhibits an everywhere positive stationary density concentrated on a small region (see [12] for an analytical treatment), which can be attributed to the chaotic itinerancy of the neural network.
In the paper, with the help of a computer aided proof, we establish several results about the statistical and geometrical properties of the above system, with the goal to show that “the behavior of this system exhibits a kind of chaotic itineracy”. We show that the system is (exponentially) mixing, hence globally chaotic. We also show a rigorous estimate of the density of probability (and then the frequency) of visits of typical trajectories near the attractors, showing that this is relatively high with respect to the density of probability of visits in other parts of the space. This is done by a computer aided rigorous estimate of the stationary probability density of the system. The computer aided proof is based on the approximation of the transfer operator of the real system by a finite rank operator which is rigorously computed and whose properties are estimated by the computer. The approximation error from the real system to the finite rank one is then managed using an appropriated functional analytic approach developed in [13] for random systems (see also [14,15,16] for applications to deterministic dynamics or iterated functions systems).
The paper is structured as follows. In Section 2, we review basic definitions related to random dynamical systems, in particular, the Perron-Frobenius operator, which will play a major role. In Section 3, we present our example along with an explanation of the method used to study it. The numerical results supported by the previous theoretical sections are presented in Section 4, while in Section 5 we present a conclusion with some additional comments.

2. Random Dynamical Systems

This section follows [17,18]. We denote by ( X , X , p ) a probability space and by ( M , A , μ ) the corresponding space of sequences (over N or Z ), with the product σ -algebra and probability measure. In addition, we denote by f the shift map on M, f ( { x i } i I ) = { x i + 1 } i I , where I = N 0 or Z .
Definition 1.
Let ( N , B ) be a measurable space. Endow M × N with the product σ-algebra A B . A random transformation over f is a measurable transformation of the form
F : M × N M × N , F ( x , y ) = ( f ( x ) , F x ( y ) ) ,
where x F x depends only on the zeroth coordinate of x.
Suppose we have a (bounded, measurable) function ϕ : N R . Given a random orbit { y i } i I , for which we know the value of y 0 = v N , we may ask what is the expected value for ϕ ( y 1 ) , that is, E ( ϕ ( y 1 ) | y 0 = v ) . Since the iterate depends on an outcome x M , which is distributed according to μ , this can be calculated as
U ϕ ( v ) = M ϕ ( F x ( v ) ) d μ ( x ) .
The equation above defines the transition operator associated with the random transformation F as an operator in the space L ( m ) of bounded measurable functions. Dually, we can consider its adjoint transition operator (These operators are related by ϕ d ( U * η ) = ( U ϕ ) d η , for every bounded measurable ϕ : N R (see [18] (lemma 5.3)).) that acts in the space of probability measures η on N, defined by
U * η ( B ) = ( F x ) * η ( B ) d μ ( x ) = η ( F x 1 ( B ) ) d μ ( x ) .
Of particular importance are the fixed points of the operator U * .
Definition 2.
A probability measure η for N is called stationary for the random transformation F if U * η = η .
We recall the deterministic concept of invariant measures.
Definition 3.
If h : C C is a measurable mapping in the measurable space ( C , C ) , we say ν is an invariant measure for h if ν ( h 1 ( E ) ) = ν ( E ) for any measurable E C .
In the one-sided case ( M = X N ), invariant measures and stationary measures for F are related by the following proposition.
Proposition 1
([18] (proposition 5.4)). A probability measure η on N is stationary for F if and only if μ × η is invariant for F.
Another concept from deterministic dynamical systems that can be naturally extended to random dynamical systems is that of ergodicity.
Definition 4.
Suppose η is stationary for F. We say that η is ergodic for F if either
1.
every (bounded measurable) function ϕ that is η-stationary, i.e., U ϕ = ϕ , is constant in some full η-measure set;
2.
every set B that is η-stationary, i.e., whose characteristic function χ B is η-stationary, has full or null η-measure.
In fact, both conditions are equivalent (see [18] (proposition 5.10)). In the one-sided case, the following proposition relates ergodicity of a random dynamical system with the deterministic concept.
Proposition 2.
A stationary measure η is ergodic for F if and only if μ × η is an ergodic F-invariant measure.
Suppose N is a manifols with boundary and consider its Lebesgue measure m. It’s useful to consider the measures which are absolutely continuous and the effect of U * upon their densities. This is possible, for example, if  ( F x ) * m is absolutely continuous for every x X .
Definition 5.
The Perron-Frobenius operator with respect to the random transformation F is the operator (This is an extension to the deterministic case, in which the Koopman operator U f = f F is used instead of the transition operator.)
L : L 1 ( m ) L 1 ( m ) , L f = d U * ( f m ) d m .
We can use the Perron-Frobenius operator to define a mixing property.
Proposition 3.
We say that a stationary measure η = h d m is mixing for F if the Perron-Frobenius operator L satisfies
lim n L n ( f ) g d m = f d m g h d m
for every f L 1 ( m ) and g L ( m ) (This definition is adapted from [19] (Equation (1.5)).).
Since L and U are dual, the mixing condition can be restated as
lim n U n ( g ) f d η = g d η f d η
for every f L 1 ( η ) and g L ( η ) , which is similar to the usual definition of decay of correlations.
If F is mixing, then F is ergodic for η (see [17] (Theorem 4.4.1)). In our example, we shall verify the following stronger condition.
L n | V L 1 ( m ) L 1 ( m ) 0 , V = f L 1 ( m ) : f d m = 0 .
Remark 1.
It suffices to verify that L n | V L 1 ( m ) L 1 ( m ) < 1 for some n N , because L ξ L 1 = 1 .
Proposition 4.
If the Perron-Frobenius operator L of a random dynamical system F satisfies Equation (1), then L admits a unique stationary density h and F is mixing.
Proof. 
L L 1 ( m ) = 1 because L is a Markov operator [17] (Remark 3.2.2). Take a density f L 1 ( m ) . We claim that L n ( f ) h for some density h. On the contrary, there would be ϵ > 0 and a subsequence { L n k ( f ) } k N such that
k N :   L n k + 1 ( f ) L n k ( f ) L 1 ( m ) ϵ .
L n k + 1 n k ( f ) f V because L n k + 1 n k ( f ) is a density, and 2 L | V L 1 ( m ) n k ϵ for every k N , a contradiction.
h is stationary because L is bounded, and unique because g h V implies L n ( g ) h L 1 ( m ) 0 for any density g. Similarly, the mixing property follows from f ( f d m ) h V . ☐
Remark 2.
Any density f must converge exponentially fast to the stationary density h. Precisely, given N and α < 1 such that L N L 1 ( m ) α , the fact that f h V implies that L n ( f ) h L 1 ( m ) C λ n for some C > 0 and λ = α 1 / N < 1 .
In the following, we will use Equation (1) as the definition of mixing property.

3. A Neuroscientifically Motivated Example

In [5], a neural network showing successive memory recall without instruction, inspired by Hopfield’s asynchronous neural network and the structure of the mammalian neocortex, is presented.
A macrovariable, related to the “activity” of the network (see Figure 1, similar plots appeared also in [5,20]) was observed to evolve as a noisy one dimensional map in the case that the network receives no external stimulus (its definition of can be found in [5] (p. 6).). This was regarded in [20] as related to a rule of successive association of memory, exhibiting chaotic dynamics.
This behavior can be modeled as a random dynamical system T ξ , with a deterministic component given by an Arnold circle map (see Figure 2) and a stochastic part given by a random additive noise. The system can be hence defined as
x n + 1 = T ( x n ) + ξ n ( mod 1 ) , where T ( v ) = v + A sin ( 4 π v ) + C ,
for A = 0.08 , C = 0.1 and ξ n an i.i.d. sequence of random variables with a distribution assumed uniform over [ ξ , ξ ] .
The Perron-Frobenius operator (Definition 5) associated to this system is given by L ξ = N ξ L (a proof can be found in [17] (p. 327)), where N ξ is a convolution operator (Equation (3)) and L is the Perron-Frobenius operator of T.
N ξ f ( t ) = ξ 1 ξ / 2 ξ / 2 f ( t τ ) .
It is well known that such an operator has an invariant probability density in L 1 . In the following, we show that L ξ can be rigorously approximated and a mixing property can be proved using the numerical estimates.
In addition, a rigorous approximation of the stationary density of L ξ is obtained, from which we can conclude the existence of a kind of chaotic itinerancy.

Rigorous Approximation

Here we show how we can study the behavior of the Perron-Frobenius operator associated to Equation (2) by approximating it by a finite rank operator. The finite rank approximation we use is known in literature as Ulam’s method. For more details, see [13].
Suppose we are given a partition of S 1 into intervals I δ = { I i } i = 1 n and denote the characteristic function of I i by χ i . An operator L : L 1 ( S 1 ) L 1 ( S 1 ) can be discretized by
L δ : L 1 ( S 1 ) L 1 ( S 1 ) , L δ = π δ L π δ ,
where π δ : L 1 ( S 1 ) L 1 ( S 1 ) is the projection
π δ h ( x ) = i = 1 n E ( h | I i ) χ i .
This operator is completely determined by its restriction to the subspace generated by { χ 1 , , χ n } , and thus may be represented by a matrix in this base, which we call the Ulam matrix. In the following, we assume that δ > 0 and I δ is a partition of [ 0 , 1 ] ( mod 1 ) S 1 with diameter < δ .
For computational purposes, the operator L ξ is discretized as
L δ , ξ = π δ N ξ π δ L π δ .
This is simple to work with because it is the product of the discretized operators π δ N ξ π δ and  π δ L π δ .
Let V = { f L 1 ( m ) : f d m = 0 } and denote by f ξ and f ξ , δ the stationary probability densities for L ξ and L δ , ξ , respectively. Suppose L δ , ξ n α < 1 for some n N . Since
L ξ i | V L 1 L δ , ξ i | V L 1 + L δ , ξ i | V L ξ i | V L 1
and ([13] (Equation (2)))
f ξ f ξ , δ L 1 1 1 α ( L δ , ξ n L ξ n ) f ξ L 1 ,
we search a good estimate of ( L δ , ξ n L ξ n ) f ξ L 1 to prove both mixing of T ξ and give a rigorous estimate of f ξ f ξ , δ L 1 .
Since the calculus of L δ , ξ i | V is computationally complex, an alternative approach is used in [13] (see [22] for a previous application of a similar idea to deterministic dynamics). First, we use a coarser version of the operator, L δ contr , ξ , where δ contr is a multiple of δ . Then, we determine n contr N and constants C i , contr , for i < n contr , and α contr < 1 in order that
L δ contr , ξ i C i , contr , L δ contr , ξ n contr | V α contr .
Finally, the following lemma from [13] is used.
Lemma 1.
Let L γ , ξ i | V L 1 C i ( γ ) ; let σ be a linear operator such that σ 2 = σ , σ L 1 1 , and σ π γ = π γ σ = π γ ; let Λ = σ N ξ σ L . Then we have
( L γ , ξ n Λ n ) N ξ L 1 γ ξ · 2 i = 0 n 1 C i ( γ ) + 1
It is applied to two cases.
  • γ = δ contr , σ = π δ and Λ = L δ , ξ implies
    ( L δ contr , ξ n L δ , ξ n ) N ξ L 1 δ ξ · 2 i = 0 n 1 C i , contr + 1 .
    This is used to obtain n N , α < 1 and C i , i < n , such that
    L δ , ξ i | V C i , L δ , ξ n | V α .
  • γ = δ , σ = Id and Λ = L ξ implies
    L ξ n + 1 | V L 1 L δ , ξ n L ξ | V L 1 + ( L ξ n L δ , ξ n ) L ξ L 1 α + δ ξ 2 i = 0 n 1 C i ( δ ) + 1 .
    By Remark 1, we conclude that the mixing condition is satisfied whenever
    λ = α + δ ξ 2 i = 0 n 1 C i ( δ ) + 1 < 1 .
We remark that a simple estimate to Equation (5) is given by ([13] (Equation (4)))
f ξ f ξ , δ L 1 1 + 2 i = 0 n 1 C i 2 ( 1 α ) δ ξ 1 Var ( ρ ) .
The analysis of data obtained from the numerical approximation f ˜ of f δ , ξ , in particular its variance, allows the algorithm in [13] to improve greatly this bound, using interval arithmetic. In Table 1, the value l1apriori obtained from the first estimate Equation (9) and the best estimate l1err are compared.

4. Results

We verified mixing and calculated the stationary density for the one dimensional system Equation (2) using the numerical tools from the compinv-meas project (see [13]), which implements the ideas presented in Section 3.1. The data obtained is summarized in Table 1.
For an explanation of the values calculated, refer to Section 3.1. In the column l1apriori, we have the estimate Equation (9) for the approximation error of the stationary density in L 1 and in l1err, the improved estimate as in [13] (Section 3.2.5).
In every case, we used the following sizes of the partition.
In Figure 3, stationary densities obtained with this method are shown.
We also studied the system Equation (2) in the case that A = 0.07 for the same range of noises (Table 2). In Figure 4, stationary densities obtained in this case are shown. We note that the same kind of “chaotic itinerancy” obtained in the main case is observed.

5. Conclusions

We have shown how the numerical approach developed in [13] can be used to study dynamical properties for a one dimensional random dynamical system of interest in the areas of physiology and neural networks.
In particular, we established mixing of the system and a rigorous estimate of its stationary density, which allowed us to observe that the trajectories concentrate in certain “weakly attracting” and “low chaotic” regions of the space, in concordance with the concept of chaotic itinerancy. The concept itself still did not have a complete mathematical formalization, and deeper understanding of the systems where it was found is important to extract its characterizing mathematical aspects.
The work we have done is only preliminary, to obtain some initial rigorous evidence of the chaotic itineracy in the system. Further investigations are important to understand the phenomenon more deeply. Firstly, it would be important to understand more precisely the nature of the phenomenon: rigorously computing Lyapunov exponents and other chaos indicators. It would also be important to investigate the robustness of the behavior of the system under various kinds of perturbations, including the zero noise limit. Another important direction is to refine the model to adapt it better to the experimental data shown in Figure 1, with a noise intensity which depends on the point.

Author Contributions

All authors contributed to this article equally.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ichiro Tsuda (2013) Chaotic Itinerancy. Scholarpedia, 8(1):4459. Available online: http://www.scholarpedia.org/article/Chaotic_itinerancy (accessed on 27 February 2018).
  2. Tsuda, I. Chaotic itineracy and its roles in cognitive neurodynamics. Curr. Opin. Neurobiol. 2015, 31, 67–71. [Google Scholar] [CrossRef] [PubMed]
  3. Tsuda, I. Hypotheses on the functional roles of chaotic transitory dynamics. Chaos 2009, 19, 015113. [Google Scholar] [CrossRef] [PubMed]
  4. John, W. Milnor (2006) Attractor. Scholarpedia, 1(11):1815. Available online: http://www.scholarpedia.org/article/Attractor (accessed on 27 February 2018).
  5. Tsuda, I.; Koerner, E.; Shimizu, H. Memory dynamics in asynchronous neural networks. Prog. Theor. Phys. 1987, 78, 51–71. [Google Scholar] [CrossRef]
  6. Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed]
  7. Hopfield, J.J.; Feinstein, D.I.; Palmer, R.G. ‘Unlearning’ has a stabilizing effect in collective memories. Nature 1983, 304, 158–159. [Google Scholar] [CrossRef] [PubMed]
  8. Hoshino, O.; Usuba, N.; Kashimori, Y.; Kambara, T. Role of itinerancy among attractors as dynamical map in distributed coding scheme. Neural Netw. 1997, 10, 1375–1390. [Google Scholar] [CrossRef]
  9. Zeller, M.; Bauer, M.; Martienssen, W. Neural dynamics modeled by one-dimensional circle maps. Chaos Solitons Fractals 1995, 5, 885–893. [Google Scholar] [CrossRef]
  10. Arnold, V.I. Small denominators. I. Mappings of the circumference onto itself. Am. Math. Soc. Transl. Ser. 2 1965, 46, 213–284. [Google Scholar]
  11. Glass, L. Synchronization and rhythmic processes in physiology. Nature 2001, 410, 277–284. [Google Scholar] [CrossRef] [PubMed]
  12. Zmarrou, H.; Humburg, A.J. Dynamics and bifurcations of random circle diffeomorphism. Discrete Contin. Dyn. Syst. B 2008, 10, 719–731. [Google Scholar]
  13. Galatolo, S.; Monge, M.; Nisoli, I. Existence of Noise Induced Order, a Computer Aided Proof. Available online: http://arxiv.org/pdf/1702.07024v2.pdf (accessed on 27 February 2018).
  14. Galatolo, S.; Nisoli, I. An elementary approach to rigorous approximation of invariant measures. SIAM J. Appl. Dyn. Syst. 2014, 13, 958–985. [Google Scholar] [CrossRef]
  15. Galatolo, S.; Nisoli, I. Rigorous computation of invariant measures and fractal dimension for maps with contracting fibers: 2D Lorenz-like maps. Ergod. Theory Dyn. Syst. 2016, 36, 1865–1891. [Google Scholar] [CrossRef]
  16. Galatolo, S.; Monge, M.; Nisoli, I. Rigorous approximation of stationary measures and convergence to equilibrium for iterated function systems. J. Phys. A 2016, 49, 274001. [Google Scholar] [CrossRef]
  17. Lasota, A.; Mackey, M. Chaos, Fractals, and Noise. Stochastic Aspects of Dynamics, 2nd ed.; Springer-Verlag: New York, NY, USA, 1994. [Google Scholar]
  18. Viana, M. Lectures on Lyapunov Exponents; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  19. Baladi, V. Decay of correlations. In Proceedings of Symposia in Pure Mathematics; American Mathematical Society: Providence, RI, USA, 2001; Volume 69. [Google Scholar]
  20. Tsuda, I. Dynamic link of memory—Chaotic memory map in nonequilibrium neural networks. Neural Netw. 1992, 5, 313–326. [Google Scholar] [CrossRef]
  21. Tsuda, I.; Chubu University. Personal communication, 2016.
  22. Galatolo, S.; Nisoli, I.; Saussol, B. An elementary way to rigorously estimate convergence to equilibrium and escape rates. J. Comput. Dyn. 2015, 2, 51–64. [Google Scholar] [CrossRef]
Figure 1. Data plot suggesting the model we studied ([21]).
Figure 1. Data plot suggesting the model we studied ([21]).
Mathematics 06 00039 g001
Figure 2. Deterministic component of Equation (2).
Figure 2. Deterministic component of Equation (2).
Mathematics 06 00039 g002
Figure 3. Approximated stationary densities f ξ , δ for T ξ , with δ = 2 19 and A = 0.08 . (a) ξ = 0.732 × 10 1 ; (b) ξ = 0.488 × 10 1 ; (c) ξ = 0.305 × 10 1 ; (d) ξ = 0.214 × 10 1 .
Figure 3. Approximated stationary densities f ξ , δ for T ξ , with δ = 2 19 and A = 0.08 . (a) ξ = 0.732 × 10 1 ; (b) ξ = 0.488 × 10 1 ; (c) ξ = 0.305 × 10 1 ; (d) ξ = 0.214 × 10 1 .
Mathematics 06 00039 g003
Figure 4. Approximated stationary densities f ξ , δ for T ξ , with δ = 2 19 and A = 0.07 . (a) ξ = 0.732 × 10 1 ; (b) ξ = 0.488 × 10 1 ; (c) ξ = 0.305 × 10 1 ; (d) ξ = 0.214 × 10 1 .
Figure 4. Approximated stationary densities f ξ , δ for T ξ , with δ = 2 19 and A = 0.07 . (a) ξ = 0.732 × 10 1 ; (b) ξ = 0.488 × 10 1 ; (c) ξ = 0.305 × 10 1 ; (d) ξ = 0.214 × 10 1 .
Mathematics 06 00039 g004
Table 1. Summary of the L 1 bounds on the approximation error obtained for the range of noises ξ , where n contr and α contr are chosen in order to satisfy Equation (6) so that the values α and C i obtained through Lemma 1 attempt to minimize the error l1err obtained through the algorithm in [13].
Table 1. Summary of the L 1 bounds on the approximation error obtained for the range of noises ξ , where n contr and α contr are chosen in order to satisfy Equation (6) so that the values α and C i obtained through Lemma 1 attempt to minimize the error l1err obtained through the algorithm in [13].
ξ n contr α contr α C i l1aprioril1err
0.732 × 10 1 126 0.027 0.05 56.64 0.313 × 10 2 0.715 × 10 4
0.610 × 10 1 167 0.034 0.067 78.66 0.530 × 10 2 0.105 × 10 3
0.488 × 10 1 231 0.051 0.1 120.56 0.106 × 10 1 0.184 × 10 3
0.427 × 10 1 278 0.068 0.14 156.45 0.163 × 10 1 0.268 × 10 3
0.366 × 10 1 350 0.087 0.19 213.17 0.273 × 10 1 0.432 × 10 3
0.305 × 10 1 453 0.12 0.26 307.03 0.523 × 10 1 0.813 × 10 3
0.275 × 10 1 532 0.14 0.32 380.64 0.776 × 10 1 0.122 × 10 2
0.244 × 10 1 596 0.19 0.41 467.70 0.124 0.202 × 10 2
Here δ = 2 19 is used to calculate the invariant density. δ contr = 2 14 : Used to find the estimates in Equation (6). δ est = 2 12 : Used to estimate the L 1 error of the invariant density.
Table 2. Summary of the L 1 bounds on the approximation error obtained for the range of noises ξ , for the system Equation (2) with the alternative value A = 0.07 .
Table 2. Summary of the L 1 bounds on the approximation error obtained for the range of noises ξ , for the system Equation (2) with the alternative value A = 0.07 .
ξ n contr α contr α C i l1aprioril1err
0.732 × 10 1 183 0.03 0.059 83.57 0.466 × 10 2 0.255 × 10 4
0.610 × 10 1 237 0.046 0.089 119.31 0.822 × 10 2 0.282 × 10 4
0.488 × 10 1 332 0.069 0.14 186.80 0.170 × 10 1 0.323 × 10 4
0.427 × 10 1 406 0.087 0.18 244.95 0.267 × 10 1 0.358 × 10 4
0.366 × 10 1 494 0.12 0.25 330.89 0.459 × 10 1 0.419 × 10 4
0.305 × 10 1 500 0.3 0.46 419.92 0.974 × 10 1 0.646 × 10 4
0.275 × 10 1 596 0.32 0.52 517.97 0.151 0.807 × 10 4
0.244 × 10 1 600 0.49 0.73 573.04 0.326 0.189 × 10 3

Share and Cite

MDPI and ACS Style

Bioni Liberalquino, R.; Monge, M.; Galatolo, S.; Marangio, L. Chaotic Itinerancy in Random Dynamical System Related to Associative Memory Models. Mathematics 2018, 6, 39. https://doi.org/10.3390/math6030039

AMA Style

Bioni Liberalquino R, Monge M, Galatolo S, Marangio L. Chaotic Itinerancy in Random Dynamical System Related to Associative Memory Models. Mathematics. 2018; 6(3):39. https://doi.org/10.3390/math6030039

Chicago/Turabian Style

Bioni Liberalquino, Ricardo, Maurizio Monge, Stefano Galatolo, and Luigi Marangio. 2018. "Chaotic Itinerancy in Random Dynamical System Related to Associative Memory Models" Mathematics 6, no. 3: 39. https://doi.org/10.3390/math6030039

APA Style

Bioni Liberalquino, R., Monge, M., Galatolo, S., & Marangio, L. (2018). Chaotic Itinerancy in Random Dynamical System Related to Associative Memory Models. Mathematics, 6(3), 39. https://doi.org/10.3390/math6030039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop