Next Article in Journal
A Readout Circuit for MEMS Gas Sensor
Next Article in Special Issue
Development of an Assessment Model for the Effect of the Replacement of Minimal Artificial Ossicles on Hearing in the Inner Ear
Previous Article in Journal
A Novel Intelligent Rebound Hammer System Based on Internet of Things
Previous Article in Special Issue
A 7.6-nW 1-kS/s 10-Bit SAR ADC for Biomedical Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Architecture Optimization of a Non-Linear Autoregressive Neural Networks for Mackey-Glass Time Series Prediction Using Discrete Mycorrhiza Optimization Algorithm

Tijuana Institute of Technology, TecNM, Tijuana 22379, Mexico
*
Author to whom correspondence should be addressed.
Micromachines 2023, 14(1), 149; https://doi.org/10.3390/mi14010149
Submission received: 1 December 2022 / Revised: 30 December 2022 / Accepted: 30 December 2022 / Published: 6 January 2023

Abstract

:
Recurrent Neural Networks (RNN) are basically used for applications with time series and sequential data and are currently being used in embedded devices. However, one of their drawbacks is that RNNs have a high computational cost and require the use of a significant amount of memory space. Therefore, computer equipment with a large processing capacity and memory is required. In this article, we experiment with Nonlinear Autoregressive Neural Networks (NARNN), which are a type of RNN, and we use the Discrete Mycorrhizal Optimization Algorithm (DMOA) in the optimization of the NARNN architecture. We used the Mackey-Glass chaotic time series (MG) to test the proposed approach, and very good results were obtained. In addition, some comparisons were made with other methods that used the MG and other types of Neural Networks such as Backpropagation and ANFIS, also obtaining good results. The proposed algorithm can be applied to robots, microsystems, sensors, devices, MEMS, microfluidics, piezoelectricity, motors, biosensors, 3D printing, etc.

1. Introduction

Optimization is not limited to applied mathematics, engineering, medicine, economics, computer science, operations research or any other science, but has become a fundamental tool in all fields, where constantly developing new algorithms and theoretical methods have allowed it to evolve in all directions, with a particular focus on artificial intelligence, such as deep learning, machine learning, computer vision, fuzzy logic systems, and quantum computing [1,2].
Optimization has grown steadily over the past 50 years. Modern society not only lives in a highly competitive environment, but is also forced to plan for growth in a sustainable manner and be concerned about resource conservation. Therefore, it is essential to optimally plan, design, operate and manage resources and assets. The first approach is to optimize each operation separately. However, the current trend is toward an integrated approach: synthesis and design, design and control, production planning, scheduling and control [3].
Theoretically, optimization has evolved to provide general solutions to linear, non-linear, unbounded and constrained network optimization problems. These optimization problems are called mathematical programming problems and are divided into two different categories: linear and nonlinear programming problems. Biologically derived genetic algorithms and simulated annealing are two equally powerful methods that have emerged in recent years. The development of computer technology has provided users with a variety of optimization codes with varying degrees of rigor and complexity. It is also possible to extend the capabilities of an existing method by integrating the features of two or more optimization methods to achieve more efficient optimization methodologies [4]; current optimization methods that can solve specific problems are still being developed, as we do not yet have a method that can solve them all, such as explained by the No Free Lunch (NFL) Algorithm [5], although the research trend is moving in that direction.
RNNs are a special class of neural network characterized by their inherent self-connectivity [6], and their variants are used in many contexts where temporal dependence of data is an important latent feature in model design [7]. The most important applications of RNNs include time series prediction [8], sequence transduction [9], language modeling [10,11,12,13], speech recognition [14], word embedding learning [15], sound modeling [16], handwriting recognition [17,18], and image generation [19]. A common variant of RNN called long short-term memory [20] is used in many of these studies.
One of the main advantages of this method with respect to others is that in general the NARNN-DMOA method is much easier to implement with better results, and with lower computation costs. Other methods use very robust Ensemble Neural Network architectures of several layers and of more than 2000 neurons and different architectures of Interval Type-2 Fuzzy Logic Systems (IT2FLSs), in addition to optimization algorithms such as PSO Genetic Algorithms [21,22,23], which implies a high computational cost.
The algorithm can be applied, as we previously mentioned, to robots, microsystems, sensors, devices, etc., in the optimization of the parameters of their models that are being experimented upon. The proposed algorithm can be used in the optimization of the architecture of a neural network or in the parameters of the membership functions of a fuzzy logic system; as we have seen in other articles [24,25,26,27], this type of experimentation with the DMOA is the subject of a future work that we plan to undertake in due course
The main contribution of this research is to use the DMOA algorithm to optimize the architecture of the NARNN neural network using the MG chaotic time series, which has not previously been done in the current literature.
The structure of this paper is as follows: (1) brief introduction of Optimization and Recurrent Neural Networks (RNNs), (2) we include a brief description of Nonlinear Autoregressive Neural Networks (NARNNN), (3) presentation of the Discrete Mycorrhiza Optimization Algorithm (DMOA) inspired by the symbiosis of plant roots and MN, (4) proposed method using the NARNN, the new DMOA algorithm and Mackey Glass chaotic time series, (5) results obtained from this research, such as statistical data, hypothesis testing and comparison of the DMOA-NARNN method with other methods, (6) in-depth discussion of the results and comparison of the error with other methods, and (7) conclusions of the obtained results.

2. Nonlinear Autoregressive Neural Networks

An Artificial Neural Network (ANN) is a type of neural network represented by a mathematical model inspired by the neural connections of the human brain. It is an intelligent system capable of recognizing time series patterns and nonlinear features.
Therefore, it is widely used to model nonlinear dynamic time series [28]. ANN incorporates artificial neurons to process information. It consists of single neurons connected to a network via weighted links. Each input is multiplied by a weight calculated by a mathematical function that determines the activation of the neurons. Another activation function calculates the output of the artificial neuron based on a certain threshold [29].
The output of a neuron can be written as Equation (1):
y = f b + i w i x i
where b is the bias of the neuron, the bias input to the neuron algorithm is an offset value that helps the signal exceed the threshold of the activation function, f is the activation function, wi is the weight, xi is the input, and y is the output.
Several types of ANNs have been presented in the literature, including Multilayer Perceptron (MLP), in which neurons are grouped into an input layer, one or more hidden layers, and an output layer. These also include RNNs such as Layer Recurrent Networks [30], Time Delay Neural Networks (TDNN) [31], and NARNN [32]. In RNNs, the output of a dynamic system depends not only on the current inputs, but also on the history of inputs and states of the system. The NARNN is a recurrent dynamic network based on a linear autoregressive model with feedback connections, and consists of several network layers.
Humans do not start their thinking from scratch every second. As we read, we understand each word based on our understanding of the previous words. We never start thinking from scratch every time we do; our thoughts have permanence. A traditional ANN cannot do this, and it seems like a major shortcoming. For example, imagine that you want to classify what kind of event is happening at each point in a movie. It is not clear how a traditional ANN could use its reasoning about earlier events in the movie to inform later events, and RNN address this problem. They are networks with loops in them which allows information to persist.
An RNN is a type of artificial neural network that uses sequential or time series da-ta. These deep learning algorithms are commonly used for ordinal or temporal problems, such as language translation, natural language processing (NLP) [33,34], speech recognition, and image captioning [35]. They are distinguished by their "memory" because they take information from previous inputs to influence the current input and output. While traditional deep neural networks assume that inputs and outputs are independent of each other, the output of recurrent neural networks depends on previous elements within the sequence.
NARNNs are a type of RNN with memory and feedback capabilities. The output of each point is based on the result of the dynamic synthesis of the system before the current time. It has great advantages for modeling and simulating dynamic changes in time series [36]. Typical NARNNs mainly consist of an input layer, a hidden layer, an output layer and an input delay function, the basic structure of which is shown in Figure 1.
In Figure 1, y(t) is the output of the NARNN, 1..19 represents the delay order, w is the joint weight and b is the threshold of NARNNs. The model of NARNN networks can be expressed as in Equation (2), where d is the delay order and f are a nonlinear function, where the future values depend only on the previous values d of the output signal.
From the equation, it can be seen that the value of y(t) is determined by the values of y(t − 1), …, y(t − d), which indicates that based on the continuity of data development, the model uses past values to estimate the current value [37,38].
The prediction method of the NARNN model adopts the recursive prediction method. The main purpose of this prediction method is to reproduce the predicted value one step ahead.
The future values of the time series y(t) are predicted only from the past values of this series. This type of prediction is called Nonlinear Autoregression (NAR) and can be written as Equation (2):
y t = f y t 1 , y t d
This model can be used to predict financial instruments, but it does not use additional sequences [39].
Looking at Figure 2, NARNN represents the entire neural network. Figure 3 “Unrolled” represents the individual layers, or time steps, of the NARNN network. Each layer corresponds to a single piece of data [40,41].
Predicting a sequence of values in a time series is also known as multi-pass fore-casting. Closed-loop networks can perform multi-step forecasting. When external feedback is missing, closed-loop networks can still make predictions using internal feedback. In NARNN prediction, the future values of a time series are predicted only from the past values of that series.
The current literature provides a history of very extensive research on the use of NARNNs in the following areas:
  • The use of NARNN in medical devices such as continuous glucose monitors and drug delivery pumps that are often combined with closed-loop systems to treat chronic diseases, for error detection and correction due to their predictive capabilities [42].
  • The use of NARNNs as Chinese e-commerce sales forecasting to develop purchasing and inventory strategies for EC companies [43], to support management decisions [44], the effects of air pollution on respiratory morbidity and mortality [45], the relationship between time series in the economy [46], to model and forecast the prevalence of COVID-19 in Egypt. [47], etc.

3. Discrete Mycorrhiza Optimization Algorithm

Most of the world’s plant species are associated with mycorrhizal fungi in nature; this association involves the interaction of fungal hyphae on plant roots. Hyphae extend from the roots into the soil, where they absorb nutrients and transport them through the mycelium to the colonized roots [48]. Some hyphae connect host plants in what is known as a Mycorrhizal Network (MN). The MN is subway and is difficult to understand. As a result, plant and ecosystem ecologists have largely overlooked the role of MNs in plant community and ecosystem dynamics [49].
It is clear that most MN are present and provide nutrition to many plant species. This has important implications for plant competition for soil nutrients, seedling formation, plant succession and plant community and ecosystem dynamics [50].
Plant mycorrhizal associations have large-scale consequences throughout the eco-system [51,52]. The origins of plant-fungal symbiosis are ancient and have been proposed as a mechanism to facilitate soil colonization by plants 400 Mya [53,54]. Mycorrhizal symbiosis is a many-to-many relationship: plants tend to form symbioses with a diverse set of fungal species and, similarly, fungal species tend to be able to colonize plants of different species [55].
In Figure 4 we can see that through the MN resources such as carbon (CO2) from plants to fungi and water, phosphorus, nitrogen and other nutrients from fungi to plants are exchanged, in addition to an exchange of information through chemical signals when the habitat feels threatened by fire, floods, pests, or predators. It should be noted that this exchange of resources can be between plants of the same species or of different species. Figure 5 shows the symbiosis between plants and the fungal network and how the carbon in the form of sugars flows from the plants to the MN and how the MN fixes the nutrients in the roots of the plants.
The Nobel optimization algorithm DMOA is inspired by the nature of the Mycorrhiza Network (MN) and plant roots with this intimate interaction between these two organisms (plant roots and the network of MN fungi), a symbiosis is generated and it has been discovered that in this relationship [56,57,58,59,60]:
  • There is a communication between plants, which may or may not be of the same species, through a network of fungi (MN).
  • There is an exchange of resources between plants through the fungal network (MN).
  • There is a defensive behavior against predators that can be insects or animals, for the survival of the whole habitat (plants and fungi).
  • The colonization of a forest through a fungal network (MN) thrives much more than a forest where there is no exchange of information and resources.
The launch and publication of the DMOA algorithm has just been carried out in 2022 [61].
Figure 6 describes the flowchart of the DMOA algorithm: we initialize the parameters such as dimensions, epochs, number of iterations, etc., and we also initialize the two populations of plants and mycorrhizae; with these populations we find the best fitness of plants and mycorrhizae, while with these results we use the biological operators. The first operator is represented by the Lotka-Volterra System of Discrete Equations (LVSDE) Cooperative Model [62], whose result has inference on the other two models represented by LVSDE, Defense and Competitive [63,64], and in this frequency we evaluate the fitness to determine if it is better than the previous one and we update the same as the populations, if not we continue with the next iteration and continue the calculation with the biological operators. If the stop condition is fulfilled we obtain the last solution before evaluation and the algorithm ends.

4. Proposed Method

The proposed method is to use the Discrete Mycorrhiza Optimization Algorithm (DMOA) to optimize the architecture of the Nonlinear Autoregressive Neural Network (NARNN), and as input data we use the Mackey-Glass chaotic time series. In Figure 7 and Algorithm 1 we can find the DMOA-NARNN flowchart and DMOA-NARNN pseudocode, respectively. The DMOA algorithm is explained in Figure 6 in the previous section, in this flowchart we include the optimization of the NARNN, evaluating its results by means of the RMSE, until we manage to find the minimum error of that architecture through the iterations and the populations of the DMOA algorithm (Algorithm 1).
Algorithm 1 DMOA-NARNN Pseudocode. Discrete Mycorrhiza Optimization Algorithm (DMOA)
  •   Objective min or max f(x), x = (x1, x2, …, xd)
  •   Define parameters (a, b, c, d, e, f, x, y)
  •   Initialize a population of n plants and mycorrhiza with random solutions
  •   Find the best solution fit in the initial population
  •   while (t < maxIter)
  •       for i = 1:n (for n plants and Mycorrhiza population)
  •        X p = a b s F i t A
  •        X m = a b s F i t B
  •   end for
  •    a = m i n o r X p
  •    d = m i n o r X m
  •   Apply (LV-Cooperative Model)
  •    x i t + 1 = a x i b x i y i 1 g x i
  •    y i t + 1 = d y i + e x i y i 1 + h y i
  •   if x i < y i
  •      x t = x i
  •   else
  •      x t = y i
  •   end if
  •   rand ([1 2])
  •    if (rand = 1)
  •      Apply (LV-Predator-Prey Model)
  •       x i t + 1 = a x i 1 x i b x i y i
  •       y i t + 1 = d x i y i g y i
  •   else
  •      Apply (LV- Competitive Model)
  •       x i t + 1 = a x i b x i y i 1 + g x i
  •       y i t + 1 = d y i e x i y i 1 + h y i
  •   end if
  •   Evaluate new solutions.
  •   NARNN-Architecture
  •   Evaluate Error
  •   Error minor?
  •   Update NARNN-Architecture.
  •   Find the current best NARNN-Architecture solution.
  • end while
Difference equations often describe the evolution of a particular phenomenon over time. For example, if a given population has discrete generations, the size of (n + 1) 1st generation x(n + 1) is a function of the nth generation x(n). This relationship is expressed by Equation (3):
x n + 1 = f x n
We can look at this issue from another perspective. You can generate a sequence from the point x0, Equation (4):
x 0 ,   f x 0 ,   f f x 0 ,   f f f x 0 ,
f(x0) is called the first iterate of x0 under f.
Discrete models driven by difference equations are more suitable than continuous models when reproductive generations last only one breeding season (no overlapping generations) [65,66].
An example would be a population that reproduces seasonally, that is, once a year. If we wanted to determine how the population size changes over many years, we could collect data to estimate the population size at the same time each year (say, shortly after the breeding season ends). We know that between the times at which we estimate population size, some individuals will die and that during the breeding season many new individuals will be born, but we ignore changes in population size from day to day, or week to week, and look only at how population size changes from year to year. Thus, when we build a mathematical model of this population, it is reasonable that the model only predicts the population size for each year shortly after the breeding season. In this case, the underlying variable, time, is represented in the mathematical model as increasing in discrete one-year increments.
The LVSDE Equations (5)–(10), have many uses in applied science. These models were first developed in mathematical biology, after which research spread to other fields [67,68,69,70,71].
Discrete Equations (5) and (6) Cooperative Model (Resource-Exchange), for both species, where parameters a, b, d, e, g, and h are positive constants, xi and yi represent the initial conditions of the population for both species and are positive real numbers [72].
The biological operators are represented by LVSDE, the mathematical description of the Discrete Equations (7) and (8) Defense Model (Predator-Prey), where the parameters a, b, d and g are positive constants, xi and yi represent the initial population conditions for both species and are positive real numbers [73,74].
Discrete Equations (9) and (10) Competitive Model (Colonization), for two species, where the parameters a, b, d, e, g, and h are positive constants, xi and yi are the populations for each of the species respectively and are positive real numbers. Each of the parameters of the above equations is described in Table 1, [74].
x i t + 1 = a x i b x i y i 1 g x i
y i t + 1 = d y i + e x i y i 1 + h y i
x i t + 1 = a x i 1 x i b x i y i
y i t + 1 = d x i y i g y i
x i t + 1 = a x i b x i y i 1 + g x i
y i t + 1 = d y i e x i y i 1 + h y i
Table 1 contains the parameters used in all the experiments performed in this research, both those of the DMOA algorithm and those of the NARNN neural network.
The theory of Differential Equations, as well as that of Equations by Differences, can be found in Youssef N. Raffoul. Qualitative Theory of Volterra Difference Equations [75], Sigrun Bodine et al., Asymptotic Integration of Differential and Difference Equations [76], Takashi Honda et al., Operator Theoretic Phenomena of the Markov Operators which are Induced by Stochastic Difference Equations [77], Ronald E. Mickens, Difference Equations Theory, Applications and Advanced Topics [78], and Konrad Kitzing, et al., A Hilbert Space Approach to Difference Equations [79].
The metric for measuring error is RMSE (Root Mean Square Error) or root mean square deviation, which is one of the most commonly used measures for evaluating the quality of predictions. It shows how far predictions fall from measured true values using Euclidean distance Equation (11), where n is the number of data points, yi is the ith measurement and ŷi is the expected prediction [80,81].
R M S E = i = 1 n y ^ i y i 2

Mackey-Glass

Chaotic and random time series are both disordered and unpredictable. In extreme cases, the data are so mixed up that those consecutive values seem unrelated to each other. Such disorder would normally eliminate the ability to predict future values from past data.
The Mackey-Glass chaotic time series Equation (12) is a nonlinear differential equation of time delay, and this equation is widely used in the modeling of natural phenomena to make comparisons between different forecasting techniques and regression models [82,83,84], where a = 0.1, b = 0.2, and τ = 17 are real numbers, t is the time, and with this setting the series produces chaotic behavior, and we can compare the forecasting performance of DMOA-NARNN with other models in the literature.
y ˙ t = b y t + c y t τ 1 + y 10 t τ

5. Results

This section shows the results of the experiments performed in the research involving the Non-Optimized and Optimized results of the method.
Table 2 presents 10 different non-optimized NARNN architectures using only the Mackey-Glass chaotic time series; in the table the columns are represented by: N—Experiment Number, Experiment Name, S—Sample size, T—Training, V—Validation, P—Prediction, HL—Hidden Layers of the NARNN, E—Number of experiment and RMSE (Root Mean Square Error), while the best architecture of the non-optimized NARNN is found in experiment number 4, with the RMSE of 0.1670.
In Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, the y axes represent the input values (Validation-Training) and output values of the samples (Prediction-Error), the x axis represents the number of samples in time, Name is the name of the experiment, Samples is the total number of samples in the experiment, Training is the number of samples for training, Error is the minimum error obtained in the experiment, and HL represents the number of neurons in the hidden layers.
Figure 8 shows the behavior of the data for 1000 samples of the NARNN403, obtaining an RMSE of 0.2307, with the reference data at the top of the figure.
Figure 9 and Figure 10 show the data behavior for 1000 samples of the NARNN404 and NARNN405, obtaining an RMSE of 0.167 and 0.2488, respectively, with the reference data at the top of each figure.
Table 3 shows the results of 39 NARNN architectures optimized with the DMOA algorithm using the Mackey-Glass chaotic time series, in the table the columns are represented by: N - Experiment Number, Experiment Name, S—Sample size, T—Training, V—Validation, P—Prediction, HL—Hidden Layers of the NARNN, I—Number of iterations, Tt—total time of the experiments in seconds, T—time in which the best result was found and RMSE (Root Mean Square Error). The best architecture of the non-optimized NARNN is found in experiment number 31, with the RMSE of 0.0023.
Figure 11, Figure 12 and Figure 13 show the data behavior for 700, 700 and 1000 samples of the NARNN053, NARNN302 and NARNN303, obtaining an RMSE of 0.0044, 0.0023 and 0.0033, respectively, with the reference data at the top of each figure.
As for the complexity of the DMOA algorithm, it is a linear order algorithm that uses the discrete equations of Lotka-Volterra Equations (5)–(10), and in the search to find the global minimum it performs iterations and in each cycle it compares the best previous local minimum with the lowest current minimum and updates the value in the case that this is the case. As for the times, Table 3 shows the times Tt which represents the total time (seconds) of the experiment and T (seconds) the time in which the DMOA algorithm found the lowest local minimum; in terms of its efficiency the algorithm took 1235 s, about 21 min, to find the lowest minimum 0.0023, which seems to us a short time compared to the times used by the method [22] of up to 3 h and a half, the method [21], its experiments took up to 81 h to find the lowest minimum and as for the method [23] it does not provide the times of its experiments.

5.1. Statistical Data

Table 4 shows 30 experiments with eight non-optimized NARNNN architectures. Each column represents the total number of samples and the number of training samples used for each architecture (700 × 300), and at the end of the table we can find the results of the total sum, mean and standard deviation for each column.
Table 5 shows 30 experiments with eight optimized NARNNN architectures; each column represents the total number of samples and the number of training samples used by each architecture (700 × 300), and at the end of the table we can find the results of the total sum, mean and standard deviation for each column.

5.2. Hypothesis Test

Equation (13) represents Hypothesis Testing, Null Hypothesis Equation (14) and Alternative Hypothesis Equation (15), with which comparisons were made between the non-optimized and optimized experiments of the method proposed here.
z = x ¯ 1 x ¯ 2 D 0 σ 1 2 n 1 + σ 2 2 n 2
H 0 : μ 1 μ 2
H a : μ 1 < μ 2   c l a i m
where x ¯ 1 is the Mean of sample 1, x ¯ 2 Mean of sample 2, σ 1 Standard Deviation of sample 1, σ 2 Standard Deviation of sample 2, n 1 Number of sample data 1, n 2 Number of sample data 2, μ 1 μ 2 = D 0 and μ 1 μ 2 = D 0 .
Significance Level α = 0.05, Confidence Level = 95%, Confidence Level = 1 − α; 1−0.05 = 0.95 o 95%, Since the p-value is less than 0.01, the null hypothesis is rejected.
Table 6 and Table 7 show the results of the hypothesis testing done on the non-optimized and optimized methods shown above; of the eight different architectures, the test results show that in only six were the optimized NARNNs better, and the non-optimized NARNNs were better in two.
In Table 6, N and Name represent the number and name of the experiment, respectively. Error is the minimum error found, HL are the Hidden Layers of neural network (1, 2, 3), and N is the number of neurons in each HL. In Table 7, the samples are represented by Total number of samples, T is the training samples, V is the validation samples, P represents the prediction, and p-value represents the results of the hypothesis test.

5.3. Comparisone with Other Methods

Table 8 shows the comparison with other methods that performed experimentation with the chaotic Mackey-Glass time series, and it can be seen from the table that the lowest error belongs to the optimized NARNN-302.
In Table 8, case number 1, the method is the Optimization of the Fuzzy Integrators in Ensembles of ANFIS Model for Time Series Prediction [21], where the authors use the Mackey-Glass chaotic time series, with genetic optimization of Type-1 Fuzzy Logic System (T1FLS) and Interval Type-2 Fuzzy Logic System (IT2FLS) integrators in Ensemble of ANFIS models and evaluate the results through Root Mean Square Error (RMSE). ANFIS is a hybrid model of a neural network implementation of a TSK (Takagi-Sugeno-Kang) fuzzy inference system. ANFIS applies a hybrid algorithm which integrates BP (Backpropagation) and LSE (least square estimation) algorithms, and thus it has a fast learning speed.
Case number 2 refers to the method using Particle Swarm Optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange [22]. In this case, the authors propose an ensemble neural network model with type-2 fuzzy logic for the integration of responses; in addition, the particle swarm optimization method determines the number of modules of the ensemble neural network, the number of layers and number of neurons per layer, and thus the best architecture of the ensemble neural network is obtained. Once this architecture is obtained, the results of the modules with type-1 and type-2 fuzzy logic systems are added, the inputs to the fuzzy system are the responses according to the number of modules of the network, and this is the number of inputs of the fuzzy system.
Case number 3 refers to the Application of Interval Type-2 Fuzzy Neural Networks (IT2FNN) in non-linear identification and time series prediction (MG) [23]. The authors propose IT2FNN models that combine the uncertainty management advantage of type-2 fuzzy sets with the learning capabilities of neural networks. One of the main ideas of this approach is that the proposed IT2FNN architectures can obtain similar or better outputs than type-2 interval fuzzy systems using the Karnik and Mendel (KM) algorithm, but with lower computational cost, which is one of the main disadvantages of KM mentioned in many papers in the literature. Cases 4 and 5 have already been explained earlier in this article.
By making a brief description of the techniques of the different methods above, we can observe the complexity of their designs using optimization algorithms such as PSO and GAs as optimizers, robust Ensemble Neural Networks, T1FLS and IT2FLS, in comparison with our method that uses the optimization algorithm DMOA and NARNNN, which are neural networks with short memory, and according to the results are made precisely for the prediction of time series. In a future work we plan to perform experiments with the RNN LSTM networks, which have short- and long-term memories.

6. Discussion of Results

The use of metaheuristics in the optimization of methods is a constant in all research work in artificial intelligence, and in this work the DMOA algorithm was used to optimize the architecture of the NARNN neural network using the MG chaotic series as input data. We also performed experiments without optimizing the NARNN network, while with the optimization we performed experiments with 39 different architectures, and without optimization we performed experiments with 10 different architectures. When we performed the optimization we found an extremely fast algorithm that found the right architecture with very satisfactory results. Of the 39 different optimized architectures, the one that gave us the best results was number 31 (narAll303) Table 3, a NARNN network with three hidden layers of 6, 7, and 5 neurons, respectively. With this architecture we performed 3000 experiments with a total time of 5353 s and in the second 1235 we obtained the best result of 0.0023 (error). Of the 10 experiments without optimization, with architecture number 4 (narAll404) Table 2, a NARNN network with two hidden layers of 9 and 2 neurons, respectively, we also performed 3000 experiments with this architecture and obtained the best result of 0.1670 (error). We performed eight hypothesis tests under equal conditions with these results and found that in five tests the NARNN architectures optimized with the DMOA algorithm were better and in three tests the non-optimized architectures were better, as shown in Table 6. We also performed error comparisons with three other different methods of which the DMOA-NARNN was better, as shown in Table 8.

7. Conclusions

A total of 49 different architectures were designed, of which 10 non-optimized and 39 were optimized by the DMOA algorithm, 30,000 experiments were performed with the non-optimized architectures, and approximately 110,000 experiments were performed with the optimized architectures. A total of 700, 1000 and 1500 samples were generated with the MG chaotic time series, of which between 300 and 1000 were used for training, between 300 and 900 were used for validation in different combinations, and between 300 and 900 points were generated as prediction points, as can be seen in Table 2 and Table 3. The design of the NARNN architectures were two and three hidden layers, with neurons in the range of 2–9, and the graphs of the most representative results of the non-optimized and optimized NARNNs are presented in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.
The optimization of the NARNN network with the DMOA algorithm obtained good results, better than without optimizing the network, and better than the other methods with which it was compared, although not all of the optimized architectures were better in the hypothesis test (only five of them were), the results of the error were much better, as can be seen in Table 7. In the comparison with other methods, the results were also better, as demonstrated in Table 8. We were also able to verify that the DMOA optimization algorithm is fast and efficient, which was really the reason for this research. We wish to continue investigating the efficiency of the algorithm in the optimization of architectures with other types of neural networks, also in Fuzzy Logic Systems Type-1 and Type-2, and also to do the same with the optimization algorithm CMOA (Continuous Mycorrhiza Optimization Algorithm). In addition, the proposed algorithm can be applied to robots, microsystems, sensors, devices, MEMS, microfluidics, piezoelectricity, motors, biosensors, 3D printing, etc.
We also intend to conduct further research and experimentation with the DMOA method and other time series. We will also consider the DMOA and the LSTM (Long Short-Term Memory) Neural Regression Network for Mackey-Glass time series, weather and financial forecasting, and we are interested in hybridizing the method with Interval Type-2 Fuzzy Logic System (IT2FLS), and Generalized Type-2 Fuzzy Logic System (GT2FLS).

Author Contributions

Conceptualization, F.V. and O.C.; methodology, F.V. and P.M.; software, H.C.-O.; validation, H.C.-O. and F.V.; formal analysis, H.C.-O.; investigation, O.C. and P.M.; resources, O.C.; writing—original draft preparation, F.V. and O.C.; writing—review and editing, H.C.-O. and P.M.; visualization, F.V.; supervision, F.V. and O.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank Tecnologico Nacional de Mexico (TecNM) for the support and Consejo Nacional de Ciencia y Tecnologia for the Ph.D. scholarship provided to Hector Carreon-Ortiz.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Diwekar, U.M. Introduction to Applied Optimization; Springer Nature Switzerland AG: Cham, Switzerland, 2020; pp. 1–3. [Google Scholar]
  2. Ghaemi, M.B.; Gharakhanlu, N.; Rassias, T.M.; Saadati, R. Advances in Matrix Inequalities; Springer Nature Switzerland AG: Cham, Switzerland, 2021; pp. 3–4. [Google Scholar]
  3. Lange, K. Optimization Second Edition; Springer Text in Statistics; Biomathematics, Human Genetics, Statistics University of California; Springer: New York, NY, USA, 2013; pp. 1–3. [Google Scholar] [CrossRef]
  4. Kochenderfer, M.J.; Wheeler, T.A. Algorithms for Optimization; The MIT Press Cambridge: London, UK, 2019; pp. 1–4. [Google Scholar]
  5. Adam, S.P.; Alexandropoulos, S.N.; Pardalos, P.M.; Vrahatis, M.N. No Free Lunch Theorem: A Review. In Approximation and Optimization; Springer Optimization and Its Applications; Demetriou, I., Pardalos, P., Eds.; Springer: Cham, Switzerland, 2019; Volume 145. [Google Scholar]
  6. Bianchi, F.M.; Maiorino, E.; Kampffmeyer, M.C.; Rizzi, A.; Jenssen, R. An overview and comparative analysis of Recurrent Neural Networks for Short Term Load Forecasting. arXiv 2017, arXiv:1705.04378. [Google Scholar]
  7. Schäfer, A.M.; Zimmermann, H.G. Recurrent Neural Networks Are Universal Approximators; Artificial Neural Networks—ICANN 2006. ICANN 2006. Lecture Notes in Computer Science; Kollias, S.D., Stafylopatis, A., Duch, W., Oja, E., Eds.; Springer: Berlin, Heidelberg, 2006; Volume 4131. [Google Scholar]
  8. Brownlee, J. Deep Learning for Time Series Forecasting Predict the Future with MLPs, CNNs and LSTMs in Python; Machine Learning Mastery: San Juan, PR, USA, 2019; pp. 6–7. [Google Scholar]
  9. Graves, A. Sequence transduction with recurrent neural networks. arXiv 2012, arXiv:1211.3711. [Google Scholar]
  10. Graves, A. Generating sequences with recurrent neural networks. arXiv 2013, arXiv:1308.0850. [Google Scholar]
  11. Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training Recurrent Neural Networks. In Proceedings of the 30th International Conference on Machine Learning, ICML 2013, JMLR.org, Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  12. Mikolov, T. Statistical Language Models Based on Neural Networks. Ph.D. Thesis, Brno University of Technology, Brno-střed, Czechia, 2012. [Google Scholar]
  13. Sutskever, I.; Martens, J.; Hinton, G. Generating Text with Recurrent Neural Networks. In Proceedings of the 28th International Conference on Machine Learning. ICML 2011, Bellevue, WA, USA, 28 June–2 July 2011. [Google Scholar]
  14. Graves, A. Practical variational inference for neural networks. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2011; pp. 2348–2356. [Google Scholar]
  15. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2013; pp. 3111–3119. [Google Scholar]
  16. Oord, A.; Dieleman, S.; Zen, H.; Vinyals, K.S.O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  17. Graves, A.; Schmidhuber, J. Offline handwriting recognition with multidimensional recurrent neural networks. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2009; pp. 545–552. [Google Scholar]
  18. Graves, A.; Fernández, S.; Liwicki, M.; Bunke, H.; Schmidhuber, J. Unconstrained On-line Handwriting Recognition with Recurrent Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; pp. 577–584. [Google Scholar]
  19. Gregor, K.; Danihelka, I.; Graves, A.; Rezende, D.; Wierstra, D. DRAW: A recurrent neural network for image generation. In Proceedings of the 32nd International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015. [Google Scholar]
  20. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  21. Soto, J.; Melin, P. Optimization of the Fuzzy Integrators in Ensembles of ANFIS Model for Time Series Prediction: The case of Mackey-Glass. In Proceedings of the 2015 Conference of the International Fuzzy Systems Association and the European Society for Fuzzy Logic and Technology (IFSA-EUSFLAT-15), Gijón, Spain, 30 June 2015; pp. 994–999. [Google Scholar]
  22. Pulido, M.; Melin, P.; Castillo, O. Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange. Inf. Sci. 2014, 280, 188–204. [Google Scholar] [CrossRef]
  23. Castillo, O.; Castro, J.R.; Melin, P.; Rodríguez-Díaz, A. Application of interval type-2 fuzzy neural networks in non-linear identification and time series prediction. Soft Comput. 2013, 18, 1213–1224. [Google Scholar] [CrossRef]
  24. Amador-Angulo, L.; Castillo, O. Amador-Angulo, L.; Castillo, O. A Fuzzy Bee Colony Optimization Algorithm Using an Interval Type-2 Fuzzy Logic System for Trajectory Control of a Mobile Robot. In Mexican International Conference on Artificial Intelligence; Springer: Cham, Switzerland, 2015. [Google Scholar]
  25. Zangeneh, M.; Aghajari, E.; Forouzanfar, M. A Review on Optimization of Fuzzy Controller Parameters in Robotic Applications. IETE J. Res. 2020, 1–10. [Google Scholar] [CrossRef]
  26. Peraza, C.; Ochoa, P.; Castillo, O.; Geem, Z.W. Interval-Type 3 Fuzzy Differential Evolution for Designing an Interval-Type 3 Fuzzy Controller of a Unicycle Mobile Robot. Mathematics 2022, 10, 3533. [Google Scholar] [CrossRef]
  27. Jiang, Y.; Yin, S.; Dong, J.; Kaynak, O. A Review on Soft Sensors for Monitoring, Control and Optimization of Industrial Processes. IEEE Sens. J. 2020, 21, 12868–12881. [Google Scholar] [CrossRef]
  28. Bradley, E.; Kantz, H. Nonlinear time-series analysis revisited. Chaos Interdiscip. J. Nonlinear Sci. 2015, 25, 097610. [Google Scholar] [CrossRef] [Green Version]
  29. Benmouiza, K.; Cheknane, A. Forecasting hourly global solar radiation using hybrid k-means and nonlinear autoregressive neural network models. Energy Convers. Manag. 2013, 75, 561–569. [Google Scholar] [CrossRef]
  30. Long, D.; Zhang, R.; Mao, Y. Recurrent Neural Networks With Finite Memory Length. IEEE Access. 2019, 7, 12511–12520. [Google Scholar] [CrossRef]
  31. Ji, W.; Chan, C. Prediction of hourly solar radiation using a novel hybrid model of ARMA and TDNN. Solar Energy 2011, 85, 808–817. [Google Scholar] [CrossRef]
  32. Taherdangkoo, R.; Tatomir, A.; Taherdangkoo, M.; Qiu, P.; Sauter, M. Nonlinear Autoregressive Neural Networks to Predict Hydraulic Fracturing Fluid Leakage into Shallow Groundwater. Water 2020, 12, 841. [Google Scholar] [CrossRef] [Green Version]
  33. Kumar, A.; Irsoy, O.; Su, J.; Bradbury, J.; English, R.; Pierce, B.; Ondruska, P.; Gulrajani, I.; Socher, R. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing. In Proceedings of the International conference on machine learning, Palo Alto, CA, USA, 6 July–11 July 2015; p. 97. [Google Scholar]
  34. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent Trends in Deep Learning Based Natural Language Processing. IEEE Comput. Intell. Mag. 2018, 13, 55–75. [Google Scholar] [CrossRef]
  35. Kalimuthu, M.; Mogadala, A.; Mosbach, M.; Klakow, D. Fusion Models for Improved Image Captioning. In ICPR International Workshops and Challenges, ICPR 2021, Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; Volume 12666. [Google Scholar]
  36. Yassin, I.M.; Zabidi, A.; Salleh, M.K.M.; Khalid, N.E.A. Malaysian tourism interest forecasting using nonlinear auto regressive (NAR) model. In Proceedings of the 3rd International Conference on System Engineering and Technology, New York, NY, USA, 19–20 August 2013; pp. 32–36. [Google Scholar]
  37. Raturi, R.; Sargsyan, H. A Nonlinear Autoregressive Scheme for Time Series Prediction via Artificial Neural Networks. J. Comput. Commun. 2018, 6, 14–23. [Google Scholar] [CrossRef] [Green Version]
  38. Ahmed, A.; Khalid, M. A Nonlinear Autoregressive Neural Network Model for Short-Term Wind Forecasting. In Proceedings of the 2017 9th IEEE-GCC Conference and Exhibition (GCCCE), Manama, Bahrain, 8–11 May 2017. [Google Scholar]
  39. MATLAB 2022b. Deep Learning Toolbox Reference; The MathWorks, Inc.: Natick, MA, USA, 2022; pp. 2200–2203. [Google Scholar]
  40. Padilla, C.; Hashemi, R.; Mahmood, N.; Latva-aho, M. A Nonlinear Autoregressive Neural Network for Interference Prediction and Resource Allocation in URLLC Scenarios. In Proceedings of the 2021 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 20–22 October 2021. [Google Scholar]
  41. Adedeji, P.A.; Akinlabi, S.A.; Ajayi, O.O.; Madushele, N. Non-Linear Autoregressive Neural Network (NARNET) with SSA filtering for a university Campus Energy Consumption Forecast. Procedia Manuf. 2019, 33, 176–183. [Google Scholar] [CrossRef]
  42. Olney, B.; Mahmud, S.; Karam, R. Efficient Nonlinear Autoregressive Neural Network Architecture for Real-Time Biomedical Applications. In Proceedings of the 2022 IEEE 4th International Conference on Artificial Intelligence Circuits and Systems (AICAS), Incheon, Republic of Korea, 13–15 June 2022; pp. 411–414. [Google Scholar]
  43. Li, M.; Ji, S.; Liu, G. Forecasting of Chinese E-Commerce Sales: An Empirical Comparison of ARIMA, Nonlinear Autoregressive Neural Network, and a Combined ARIMA-NARNN Model. Math. Probl. Eng. Vol. 2018, 2018, 1–12. [Google Scholar] [CrossRef]
  44. Kummong, R.; Supratid, S. Long-term forecasting system using wavelet – nonlinear autoregressive neural network conjunction model. J. Model. Manag. 2019, 14, 948–971. [Google Scholar] [CrossRef]
  45. Davood, N.K.; Goudarzi, G.R.; Taghizadeh, R.; Asumadu-Sakyi, A.B.; Fehresti-Sani, M. Long-term effects of outdoor air pollution on mortality and morbidity–prediction using nonlinear autoregressive and artificial neural networks models. Atmos. Pollut. Res. 2021, 12, 46–56. [Google Scholar]
  46. Domaschenko, D.; Nikulin, E. Forecasting time series of the market indicators based on a nonlinear autoregressive neural network. Stat. Econ. Vol. 2017, 2017, 4–9. [Google Scholar] [CrossRef]
  47. Saba, A.I.; Elsheikh, A.H. Forecasting the prevalence of COVID-19 outbreak in Egypt using nonlinear autoregressive artificial neural networks. Process. Saf. Environ. Prot. 2020, 141, 1–8. [Google Scholar] [CrossRef] [PubMed]
  48. Newman, E.I. Mycorrhizal links between plants: Their functioning and ecological significance. Adv. Ecol. Res. 1988, 18, 243–270. [Google Scholar]
  49. Bahram, M.; Põlme, S.; Kõljalg, U.; Tedersoo, L. A single European aspen (Populus tremula) tree individual may potentially harbour dozens of Cenococcum geophilum ITS genotypes and hundreds of species of ectomycorrhizal fungi. FEMS Microbiol. Ecol. 2010, 75, 313–320. [Google Scholar] [CrossRef]
  50. Schimel, J.P.; Bennett, J. Nitrogen mineralization: Challenges of a changing paradigm. Ecology 2004, 85, 591–602. [Google Scholar] [CrossRef]
  51. Averill, C.; Turner, B.L.; Finzi, A.C. Mycorrhiza-mediated competition between plants and decomposers drives soil carbon storage. Nature 2014, 505, 543–545. [Google Scholar] [CrossRef] [PubMed]
  52. Dickie, I.A.; Koele, N.; Blum, J.D.; Gleason, J.D.; McGlone, M.S. Mycorrhizas in changing ecosystems, Botany 2014, 92, 149–160. [Google Scholar] [CrossRef]
  53. Redecker, D.; Kodner, R.; Graham, L.E. Glomalean Fungi from the Ordovician. Science 2000, 289, 1920–1921. [Google Scholar] [CrossRef] [Green Version]
  54. Humphreys, C.P.; Franks, P.J.; Rees, M.; Bidartondo, M.I.; Leake, J.R.; Beerling, D.J. Mutualistic mycorrhiza-like symbiosis in the most ancient group of land plants. Nat. Commun. 2010, 1, 103. [Google Scholar] [CrossRef] [Green Version]
  55. Lang, C.; Seven, J.; Polle, A. Host preferences and differential contributions of deciduous tree species shape mycorrhizal species richness in a mixed Central European forest. Mycorrhiza 2010, 21, 297–308. [Google Scholar] [CrossRef] [Green Version]
  56. Simard, S.W. Mycorrhizal Networks Facilitate Tree Communication, Learning, and Memory. In Memory and Learning in Plants; Signaling and Communication in Plants; Baluska, F., Gagliano, M., Witzany, G., Eds.; Springer: Cham, Switzerland, 2018. [Google Scholar]
  57. Castro-Delgado, A.L.; Elizondo-Mesén, S.; Valladares-Cruz, Y.; Rivera-Méndez, W. Wood Wide Web: Communication through the mycorrhizal network. Tecnol. Marcha J. 2020, 33, 114–125. [Google Scholar]
  58. Beiler, K.J.; Simard, S.W.; Durall, D.M. Topology of tree-mycorrhizal fungus interaction networks in xeric and mesic Douglas-fir forests. J. Ecol. 2015, 103, 616–628. [Google Scholar] [CrossRef]
  59. Simard, S.W.; Asay, A.; Beiler, K.; Bingham, M.; Deslippe, J.; He, X.; Philip, L.; Song, Y.; Teste, F. Resource Transfer Between Plants Through Ectomycorrhizal Fungal Networks. In Mycorrhizal Networks. Ecological Studies; Horton, T., Ed.; Springer: Dordrecht, The Netherlands, 2015; Volume 224. [Google Scholar]
  60. Gorzelak, M.A.; Asay, A.K.; Pickles, B.J.; Simard, S.W. Inter-plant communication through mycorrhizal networks mediates complex adaptive behaviour in plant communities. AoB Plants 2015, 7, plv050. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Carreon, H.; Valdez, F.; Castillo, O. A New Discrete Mycorrhiza Optimization Nature-Inspired Algorithm. Axioms 2022, 11, 391. [Google Scholar] [CrossRef]
  62. Liu, P.; Elaydi, S.N. Discrete Competitive and Cooperative Models of Lotka–Volterra Type. J. Comput. Anal. Appl. 2001, 3, 53–73. [Google Scholar]
  63. Muhammadhaji, A.; Halik, A.; Li, H. Dynamics in a ratio-dependent Lotka–Volterra competitive-competitive-cooperative system with feedback controls and delays. Adv. Differ. Equ. 2021, 230, 1–14. [Google Scholar] [CrossRef]
  64. Din, Q. Dynamics of a discrete Lotka-Volterra model. Adv. Differ. Equ. 2013, 2013, 95. [Google Scholar] [CrossRef]
  65. Liu, X. A note on the existence of periodic solutions in discrete predator–prey models. Appl. Math. Model. 2010, 34, 2477–2483. [Google Scholar] [CrossRef]
  66. Zhou, Z.; Zou, X. Stable periodic solutions in a discrete periodic logistic equation. Appl. Math. Lett. 2003, 16, 165–171. [Google Scholar] [CrossRef] [Green Version]
  67. Krabs, W. A General Predator-Prey Model. Math. Comput. Model. Dyn. Syst. 2003, 9, 387–401. [Google Scholar] [CrossRef]
  68. Allen, L.J.S. An Introduction to Mathematical Biology; Pearson Prentice Hall: New York, NY, USA, 2007; pp. 89–140. [Google Scholar]
  69. Brauer, F.; Castillo-Chavez, C. Mathematical Models in Population Biology and Epidemiology, 2nd ed.; Springer: New York, NY, USA; Dordrecht, The Netherlands; Heidelberg, Germany, 2012; pp. 123–134. [Google Scholar]
  70. Müller, J.; Kuttler, C. Methods and Models in Mathematical Biology, Deterministic and Stochastic Approaches. In Lecture Notes on Mathematical Modelling in the Life Sciences; Springer-Verlag: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  71. Voroshilova, A.; Wafubwa, J. Discrete Competitive Lotka–Volterra Model with Controllable Phase Volume. Systems 2020, 8, 17. [Google Scholar] [CrossRef]
  72. Saha, P.; Bairagi, N.; Biswas, M. On the Dynamics of a Discrete Predator-Prey Model. In Trends in Biomathematics: Modeling, 337 Optimization and Computational Problems; Mondaini, R., Ed.; Springer: Berlin, Germany, 2018; pp. 219–232. [Google Scholar]
  73. Zhao, M.; Xuan, Z.; Li, C. Dynamics of a discrete-time predator-prey system. In Advances in Difference Equations 2016; Springer Open: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  74. Chou, C.S.; Friedman, A. Introduction to Mathematical Biology, Modeling, Analysis, and Simulations. In Springer Undergraduate Texts in Mathematics and Technology; Springer International Publishing: Cham, Switzerland, 2016. [Google Scholar]
  75. Raffoul, Y.N. Qualitative Theory of Volterra Difference Equations; Springer Nature Switzerland AG: Cham, Switzerland, 2018; pp. 55–89. [Google Scholar]
  76. Bodine, S.; Lutz, D.A. Asymptotic Integration of Differential and Difference Equations; Springer International Publishing: Cham, Switzerland, 2015; pp. 237–282. [Google Scholar]
  77. Honda, T.; Iwata, Y. Operator Theoretic Phenomena of the Markov Operators which are Induced by Stochastic Difference Equations. In Advances in Difference Equations and Discrete Dynamical Systems. ICDEA 2016; Springer Proceedings in Mathematics & Statistics; Elaydi, S., Hamaya, Y., Matsunaga, H., Pötzsche, C., Eds.; Springer: Singapore, 2017; Volume 212. [Google Scholar]
  78. Mickens, R.E. Difference Equations Theory, Applications and Advanced Topics; CRC Press: Boca Raton, FL, USA, 2015; pp. 83–115. [Google Scholar]
  79. Kitzing, K.; Picard, R.; Siegmund, S.; Trostorff, S.; Waurick, M. A Hilbert Space Approach to Difference Equations. In Difference Equations, Discrete Dynamical Systems and Applications, ICDEA 2017; Springer Proceedings in Mathematics & Statistics; Elaydi, S., Pötzsche, C., Sasu, A., Eds.; Springer: Cham, Switzerland, 2019; Volume 287, pp. 285–307. [Google Scholar]
  80. Castro, J.R.; Castillo, O.; Melin, P.; Rodríguez-Díaz, A. Building Fuzzy Inference Systems with a New Interval Type-2 Fuzzy Logic Toolbox. In Transactions on Computational Science I; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  81. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?– Arguments against avoiding RMSE in the literature. Geoscientific Model Development. 2014, 7, 1247–1250. [Google Scholar] [CrossRef] [Green Version]
  82. Saeed, W.; Ghazali, R. Chaotic Time Series Forecasting Using Higher Order Neural Networks. Int. J. Adv. Sci. Eng. Inf. Technol. 2016, 6, 624–629. [Google Scholar]
  83. Martínez-García, J.A.; González-Zapata, A.M.; Rechy-Ramírez, E.J.; Tlelo-Cuautle, E. On the prediction of chaotic time series using neural networks. Chaos Theory Appl. 2022, 4, 94–103. [Google Scholar]
  84. López-Caraballo, C.H.; Salfate, I.; A Lazzús, J.; Rojas, P.; Rivera, M.; Palma-Chilla, L. Mackey-Glass noisy chaotic time series prediction by a swarm-optimized neural network. J. Physics: Conf. Ser. 2016, 720, 012002. [Google Scholar] [CrossRef]
Figure 1. Standard NARNN schematic structure of the neural networks.
Figure 1. Standard NARNN schematic structure of the neural networks.
Micromachines 14 00149 g001
Figure 2. NARNN have loops.
Figure 2. NARNN have loops.
Micromachines 14 00149 g002
Figure 3. Unrolled NARNN.
Figure 3. Unrolled NARNN.
Micromachines 14 00149 g003
Figure 4. Signaling and resource exchange through MN.
Figure 4. Signaling and resource exchange through MN.
Micromachines 14 00149 g004
Figure 5. Symbiosis between plant roots and MN.
Figure 5. Symbiosis between plant roots and MN.
Micromachines 14 00149 g005
Figure 6. DMOA Flowchart.
Figure 6. DMOA Flowchart.
Micromachines 14 00149 g006
Figure 7. DMOA-NARNN Flowchart.
Figure 7. DMOA-NARNN Flowchart.
Micromachines 14 00149 g007
Figure 8. Performance of NARNN403 for 1000 samples.
Figure 8. Performance of NARNN403 for 1000 samples.
Micromachines 14 00149 g008
Figure 9. Performance of NARNN404 for 1000 samples.
Figure 9. Performance of NARNN404 for 1000 samples.
Micromachines 14 00149 g009
Figure 10. Performance of NARNN405 for 1000 samples.
Figure 10. Performance of NARNN405 for 1000 samples.
Micromachines 14 00149 g010
Figure 11. Performance of NARNN053 for 700 samples.
Figure 11. Performance of NARNN053 for 700 samples.
Micromachines 14 00149 g011
Figure 12. Performance of NARNN302 for 700 samples.
Figure 12. Performance of NARNN302 for 700 samples.
Micromachines 14 00149 g012
Figure 13. Performance of NARNN303 for 1000 samples.
Figure 13. Performance of NARNN303 for 1000 samples.
Micromachines 14 00149 g013
Table 1. DMOA-NARNN Parameters.
Table 1. DMOA-NARNN Parameters.
ParameterDescriptionValue
DMOA—Parameters:
x i t + 1 Population x at time t
y i t + 1 Population y at time t
x i Grow rates of populations x at time t
y i Grow rates of populations y at time t
ttime
aPopulation growth rate x0.01
bInfluence of population x on itself0.02
gInfluence of population y on population x0.06
dPopulation growth rate y0
eInfluence of population x on population y1.7
hInfluence of population y on itself0.09
xInitial population in x0.0002
yInitial population in y0.0006
In the absence of population x = 0, In the absence of population y = 0
a, b, c, d, e and f—are positive constants
PopulationPopulation size20
PopulationsNumber of populations2
DimensionsDimensions size30, 50, 100
EpochsNumber of epochs30
IterationsIteration’s size30, 50, 100, 500
NARNN—Parameters:
hHidden Layers2, 3
nNeurons2–10
Vector time delay01:06:19
Table 2. Results of non-optimized experiments.
Table 2. Results of non-optimized experiments.
NNameSTVPHLERMSE
1NARNN401700300400400[1 1]30000.2777
2NARNN402700400300300[1 1]30000.1683
3NARNN4031000400600600[3 1]30000.2307
4NARNN4041000500500500[9 2]30000.1670
5NARNN4051000600400400[2 2]30000.2488
6NARNN4061500600900900[4 1 9]30000.2550
7NARNN4071500700800800[8 5 2]30000.2158
8NARNN4081500800700700[1 2 1]30000.4001
9NARNN4091500900600600[2 3 2]30000.2810
10NARNN41015001000500500[6 6 8]30000.1712
Table 3. Results of optimized experiments.
Table 3. Results of optimized experiments.
NNameSTVPHLITtTRMSE
1NARNN041700300400400[5 6]900 0.0114
2NARNN042700300400400[5 6]1000 0.0054
3NARNN043700300400400[5 6]1000 0.1012
4NARNN053700300400400[6 6]500 4050.0044
5NARNN055700300400400[6 1]100 40.0202
6NARNN056700300400400[6 6]500 4860.0067
7NARNN057700300400400[4 3]1000 8070.0075
8NARNN058700300400400[8 7]2500 1540.0131
9NARNN058r700300400400[6 5]2500 750.0202
10NARNN059700300400400[5 4]1000 4450.0061
11NARNN060700300400400[6 2]1000 2590.0081
12NARNN061700300400400[5 6]1000 4960.0044
13NARNN062700400300300[5 6]5000 6880.0024
14NARNN201700400300300[9 4]5000294814760.0024
15NARNN2021000600400400[7 5]5000 19050.0035
16NARNN2031000500500500[4 7]1000 6810.0084
17NARNN2041000400600600[8 3]1000 9170.0144
18NARNN2051000400600600[7 8]5000284927410.0076
19NARNN2061000500500500[8 5]5000326225170.0059
20NARNN2071000600400400[6 2]5000365918220.0047
21NARNN2081500600900900[7 9]5000366616860.0187
22NARNN2091500700800800[7 7]5000403926650.0104
23NARNN210p1500800700700[7 3]6000575052260.0122
24NARNN2111500900600600[7 3]3000543916580.0136
25NARNN2121500900600600[5 6]300064796060.0055
26NARNN2131500800700700[5 6]3000126854700.0080
27NARNN214p1500700800800[5 6]5000618347390.0157
28NARNN2151500600900900[5 6]3000284819100.0230
29NARNN31115001000500500[6 1]3000691919170.0047
30NARNN301700300400400[7 8 3]2000137113130.0052
31NARNN302700400300300[6 7 5]3000535312350.0023
32NARNN3031000600400400[9 2 8]3000234820030.0033
33NARNN3041000500500500[4 5 4]3000507213340.0040
34NARNN3051000400600600[8 7 1]3000443524140.0070
35NARNN3061500600900900[5 5 1]3000561023090.0851
36NARNN3071500700800800[8 4 1]3000315116640.0245
37NARNN3081500800700700[8 1 1]30003161760.0127
38NARNN3091500900600600[8 2 7]3000317515020.0098
39NARNN31015001000500500[8 3 8]3000333310500.0074
Table 4. 30 Experiments with non-optimized NARNN.
Table 4. 30 Experiments with non-optimized NARNN.
NoNon Optimized
2 Hidden Layers3 Hidden Layers
700 × 300700 × 4001000 × 4001000 × 5001000 × 6001500 × 6001500 × 7001500 × 800
13.58 × 10−13.17 × 10−12.84 × 10−12.45 × 10−13.01 × 10−19.88 × 10−12.78 × 10−17.27 × 10−1
22.97 × 10−12.39 × 10−12.81 × 10−12.32 × 10−12.96 × 10−13.41 × 10−12.72 × 10−15.39 × 10−1
32.84 × 10−12.35 × 10−12.77 × 10−12.26 × 10−12.90 × 10−13.39 × 10−12.44 × 10−14.67 × 10−1
42.81 × 10−12.23 × 10−12.66 × 10−12.25 × 10−12.55 × 10−13.38 × 10−12.36 × 10−14.35 × 10−1
52.68 × 10−12.22 × 10−12.60 × 10−12.22 × 10−12.49 × 10−13.27 × 10−12.34 × 10−14.27 × 10−1
62.66 × 10−12.22 × 10−12.47 × 10−12.06 × 10−12.47 × 10−13.19 × 10−12.33 × 10−14.23 × 10−1
72.65 × 10−12.21 × 10−12.39 × 10−12.03 × 10−12.44 × 10−13.16 × 10−12.31 × 10−14.18 × 10−1
82.62 × 10−12.19 × 10−12.39 × 10−11.83 × 10−12.44 × 10−13.07 × 10−12.30 × 10−14.18 × 10−1
92.62 × 10−12.16 × 10−12.34 × 10−11.75 × 10−12.43 × 10−13.00 × 10−12.30 × 10−14.08 × 10−1
102.52 × 10−12.13 × 10−12.33 × 10−11.74 × 10−12.40 × 10−12.92 × 10−12.27 × 10−14.02 × 10−1
112.51 × 10−12.12 × 10−12.32 × 10−11.73 × 10−12.36 × 10−12.73 × 10−12.27 × 10−13.98 × 10−1
122.50 × 10−12.11 × 10−12.31 × 10−11.72 × 10−12.34 × 10−12.73 × 10−12.21 × 10−13.94 × 10−1
132.50 × 10−12.10 × 10−12.31 × 10−11.71 × 10−12.33 × 10−12.69 × 10−12.19 × 10−13.91 × 10−1
142.49 × 10−12.09 × 10−12.29 × 10−11.69 × 10−12.31 × 10−12.63 × 10−12.16 × 10−13.89 × 10−1
152.49 × 10−12.08 × 10−12.29 × 10−11.69 × 10−12.15 × 10−12.60 × 10−12.14 × 10−12.53 × 10−1
162.43 × 10−12.07 × 10−12.29 × 10−11.68 × 10−12.11 × 10−12.47 × 10−12.04 × 10−12.51 × 10−1
172.43 × 10−12.00 × 10−12.28 × 10−11.67 × 10−12.10 × 10−12.33 × 10−11.97 × 10−12.42 × 10−1
182.43 × 10−11.95 × 10−12.27 × 10−11.66 × 10−12.08 × 10−12.29 × 10−11.95 × 10−12.38 × 10−1
192.36 × 10−11.93 × 10−12.23 × 10−11.66 × 10−12.07 × 10−12.29 × 10−11.87 × 10−12.33 × 10−1
202.36 × 10−11.89 × 10−12.17 × 10−11.66 × 10−12.05 × 10−12.27 × 10−11.86 × 10−12.27 × 10−1
212.36 × 10−11.89 × 10−12.13 × 10−11.54 × 10−12.02 × 10−12.12 × 10−11.86 × 10−12.25 × 10−1
222.33 × 10−11.89 × 10−12.01 × 10−11.52 × 10−12.02 × 10−12.06 × 10−11.82 × 10−12.25 × 10−1
232.33 × 10−11.88 × 10−11.95 × 10−11.35 × 10−12.02 × 10−12.04 × 10−11.75 × 10−12.24 × 10−1
242.28 × 10−11.82 × 10−11.90 × 10−11.30 × 10−12.01 × 10−12.03 × 10−11.72 × 10−12.23 × 10−1
252.26 × 10−11.78 × 10−11.85 × 10−11.19 × 10−12.00 × 10−11.98 × 10−11.69 × 10−12.22 × 10−1
262.25 × 10−11.78 × 10−11.85 × 10−11.19 × 10−11.97 × 10−11.96 × 10−11.67 × 10−12.19 × 10−1
272.24 × 10−11.77 × 10−11.74 × 10−11.14 × 10−11.74 × 10−11.96 × 10−11.59 × 10−12.16 × 10−1
282.23 × 10−11.70 × 10−11.60 × 10−11.09 × 10−11.62 × 10−11.95 × 10−11.58 × 10−12.15 × 10−1
292.10 × 10−11.70 × 10−11.45 × 10−11.01 × 10−11.32 × 10−11.92 × 10−11.57 × 10−12.14 × 10−1
301.96 × 10−11.65 × 10−15.20 × 10−26.12 × 10−27.29 × 10−21.61 × 10−11.50 × 10−11.94 × 10−1
Sum:7.48 × 10+06.15 × 10+06.53 × 10+04.97 × 10+06.54 × 10+08.34 × 10+06.16 × 10+09.86 × 10+0
Mean:2.49 × 10−12.05 × 10−12.18 × 10−11.66 × 10−12.18 × 10−12.78 × 10−12.05 × 10−13.29 × 10−1
SD:3.00 × 10−22.91 × 10−24.59 × 10−24.25 × 10−24.58 × 10−21.44 × 10−13.40 × 10−21.27 × 10−1
Table 5. 30 Experiments with optimized NARNN.
Table 5. 30 Experiments with optimized NARNN.
NoOptimized
2 Hidden Layers3 Hidden Layers
700 × 300700 × 4001000 × 4001000 × 5001000 × 6001500 × 6001500 × 7001500 × 800
11.72 × 10−14.62 × 10−22.63 × 10−11.97 × 10−12.71 × 10−13.88 × 10−12.85 × 10−12.93 × 10−1
21.52 × 10−14.44 × 10−22.34 × 10−11.97 × 10−11.92 × 10−13.07 × 10−12.67 × 10−12.67 × 10−1
31.35 × 10−14.21 × 10−22.32 × 10−11.92 × 10−11.92 × 10−13.06 × 10−12.53 × 10−12.66 × 10−1
41.34 × 10−14.07 × 10−22.31 × 10−11.91 × 10−11.91 × 10−12.97 × 10−12.43 × 10−12.36 × 10−1
51.18 × 10−13.98 × 10−22.30 × 10−11.87 × 10−11.91 × 10−12.94 × 10−12.39 × 10−12.29 × 10−1
61.09 × 10−13.73 × 10−22.28 × 10−11.86 × 10−11.89 × 10−12.90 × 10−12.36 × 10−12.28 × 10−1
71.07 × 10−13.59 × 10−22.27 × 10−11.86 × 10−11.89 × 10−12.86 × 10−12.25 × 10−12.27 × 10−1
89.68 × 10−23.40 × 10−22.26 × 10−11.83 × 10−11.89 × 10−12.74 × 10−12.22 × 10−12.24 × 10−1
99.67 × 10−23.26 × 10−22.22 × 10−11.80 × 10−11.86 × 10−12.74 × 10−12.21 × 10−12.23 × 10−1
109.61 × 10−23.08 × 10−22.20 × 10−11.75 × 10−11.83 × 10−12.72 × 10−12.19 × 10−12.17 × 10−1
119.54 × 10−23.06 × 10−22.18 × 10−11.74 × 10−11.74 × 10−12.70 × 10−12.17 × 10−12.13 × 10−1
129.23 × 10−22.19 × 10−22.18 × 10−11.73 × 10−11.69 × 10−12.66 × 10−12.16 × 10−12.09 × 10−1
139.01 × 10−22.18 × 10−22.16 × 10−11.72 × 10−11.69 × 10−12.62 × 10−12.14 × 10−12.08 × 10−1
148.57 × 10−22.14 × 10−22.12 × 10−11.72 × 10−11.57 × 10−12.53 × 10−12.04 × 10−12.07 × 10−1
158.11 × 10−21.91 × 10−22.09 × 10−11.72 × 10−11.50 × 10−12.53 × 10−11.78 × 10−12.05 × 10−1
167.47 × 10−21.91 × 10−22.08 × 10−11.71 × 10−11.40 × 10−12.51 × 10−11.74 × 10−12.02 × 10−1
177.13 × 10−21.79 × 10−21.89 × 10−11.69 × 10−11.35 × 10−12.44 × 10−11.73 × 10−11.99 × 10−1
186.11 × 10−21.77 × 10−21.89 × 10−11.69 × 10−11.33 × 10−12.37 × 10−11.71 × 10−11.98 × 10−1
195.99 × 10−21.31 × 10−21.85 × 10−11.67 × 10−11.32 × 10−12.33 × 10−11.58 × 10−11.98 × 10−1
205.76 × 10−21.25 × 10−21.78 × 10−11.66 × 10−11.32 × 10−12.28 × 10−11.53 × 10−11.90 × 10−1
215.43 × 10−21.12 × 10−21.78 × 10−11.66 × 10−11.29 × 10−12.27 × 10−11.53 × 10−11.78 × 10−1
225.35 × 10−21.02 × 10−21.77 × 10−11.64 × 10−11.29 × 10−12.24 × 10−11.52 × 10−11.66 × 10−1
233.96 × 10−29.31 × 10−31.76 × 10−11.45 × 10−11.28 × 10−12.21 × 10−11.50 × 10−11.65 × 10−1
243.58 × 10−29.11 × 10−31.74 × 10−11.42 × 10−11.28 × 10−12.12 × 10−11.45 × 10−11.61 × 10−1
252.82 × 10−26.99 × 10−31.71 × 10−11.37 × 10−11.27 × 10−11.98 × 10−11.44 × 10−11.58 × 10−1
262.27 × 10−25.98 × 10−31.71 × 10−11.30 × 10−11.26 × 10−11.98 × 10−11.42 × 10−11.58 × 10−1
272.13 × 10−25.63 × 10−31.54 × 10−11.16 × 10−11.26 × 10−11.92 × 10−11.41 × 10−19.36 × 10−2
281.89 × 10−24.96 × 10−31.19 × 10−11.15 × 10−11.24 × 10−11.91 × 10−11.38 × 10−16.64 × 10−2
291.84 × 10−24.21 × 10−31.03 × 10−18.25 × 10−21.23 × 10−11.88 × 10−11.37 × 10−15.36 × 10−2
301.14 × 10−23.52 × 10−35.66 × 10−21.51 × 10−21.11 × 10−11.75 × 10−11.35 × 10−14.84 × 10−2
Sum:2.29 × 10+06.50 × 10−15.81 × 10+04.79 × 10+04.71 × 10+07.51 × 10+05.70 × 10+05.69 × 10+0
Mean:7.63 × 10−22.17 × 10−21.94 × 10−11.60 × 10−11.57 × 10−12.50 × 10−11.90 × 10−11.90 × 10−1
SD:4.21 × 10−21.38 × 10−24.35 × 10−23.81 × 10−23.52 × 10−24.60 × 10−24.47 × 10−25.91 × 10−2
Table 6. Data for the eight non-optimized and optimized architectures.
Table 6. Data for the eight non-optimized and optimized architectures.
NARNN—Non-OPTIMIZEDNARNN—OPTIMIZED
NNameErrorHL and NNameErrorHL and N
123123
1NARNN4010.277711 NARNN0530.005266
2NARNN4020.168311 NARNN0620.002356
3NARNN4030.230731 NARNN2050.007078
4NARNN4040.167092 NARNN2060.004085
5NARNN4050.248822 NARNN2070.003362
6NARNN4060.2550419NARNN3060.0851551
7NARNN4070.2158852NARNN3070.0245841
8NARNN4080.4001121NARNN3080.0127811
Table 7. Hypothesis test results of the eight non-optimized and optimized NARNNs.
Table 7. Hypothesis test results of the eight non-optimized and optimized NARNNs.
SamplesResults
Non-OptimizedOptimized
TotalTVPp-Value
700300400400 8.25 × 10−26
700400300300 6.00 × 10−38
1000400600600 4.23 × 10−2
10005005005005.59 × 10−1
1000600400400 3.13 × 10−7
15006009009003.25 × 10−1
15007008008001.46 × 10−1
1500800700700 1.14 × 10−6
Table 8. Error comparison with three different methods and the non-optimized and optimized NARNN.
Table 8. Error comparison with three different methods and the non-optimized and optimized NARNN.
NExperiment DescriptionErrorSerieRef
1Genetic Algorithm—Ensemble ANFIS—T1FLS—IT2FLS 0.0219MG[21]
2Ensemble Neural Network Architecture 0.008945MSE[22]
2 Modules, 2 Hidden Layers, 2116 and 2128 Neurons respectively, PSO Optimized
3SNR(dB)ANFISIT2FNN-0IT2FNN-1IT2FNN-2IT2FNN-30.0028MG[23]
300.02250.01060.00790.00450.0028
4NARNN-DMOA (No-Optimized) 0.167MG
5NARNN-DMOA (Optimized) 0.0023MG
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Carreon-Ortiz, H.; Valdez, F.; Melin, P.; Castillo, O. Architecture Optimization of a Non-Linear Autoregressive Neural Networks for Mackey-Glass Time Series Prediction Using Discrete Mycorrhiza Optimization Algorithm. Micromachines 2023, 14, 149. https://doi.org/10.3390/mi14010149

AMA Style

Carreon-Ortiz H, Valdez F, Melin P, Castillo O. Architecture Optimization of a Non-Linear Autoregressive Neural Networks for Mackey-Glass Time Series Prediction Using Discrete Mycorrhiza Optimization Algorithm. Micromachines. 2023; 14(1):149. https://doi.org/10.3390/mi14010149

Chicago/Turabian Style

Carreon-Ortiz, Hector, Fevrier Valdez, Patricia Melin, and Oscar Castillo. 2023. "Architecture Optimization of a Non-Linear Autoregressive Neural Networks for Mackey-Glass Time Series Prediction Using Discrete Mycorrhiza Optimization Algorithm" Micromachines 14, no. 1: 149. https://doi.org/10.3390/mi14010149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop