Next Article in Journal
Prediction of Post-Intubation Tachycardia Using Machine-Learning Models
Next Article in Special Issue
Neuronless Knowledge Processing in Forests
Previous Article in Journal
A Consistent Methodology to Evaluate Temperature and Heat Wave Future Projections for Cities: A Case Study for Lisbon
Previous Article in Special Issue
Targeted Sentiment Classification Based on Attentional Encoding and Graph Convolutional Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Natural Brain-Inspired Intelligence for Non-Gaussian and Nonlinear Environments with Finite Memory

Dept of Electrical and Computer Engineering, McMaster University, Hamilton, ON L8S 4L8, Canada
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(3), 1150; https://doi.org/10.3390/app10031150
Submission received: 30 November 2019 / Revised: 4 February 2020 / Accepted: 4 February 2020 / Published: 8 February 2020

Abstract

:
The cyber processing layer of smart systems based on a cognitive dynamic system (CDS) can be a good solution for better decision making and situation understanding in non-Gaussian and nonlinear environments (NGNLE). The NGNLE situation understanding means deciding between certain known situations in NGNLE to understand the current state condition. Here, we report on a cognitive decision-making (CDM) system inspired by the human brain decision-making. The simple low-complexity algorithmic design of the proposed CDM system can make it suitable for real-time applications. A case study of the implementation of the CDS on a long-haul fiber-optic orthogonal frequency division multiplexing (OFDM) link was performed. An improvement in Q-factor of ~7 dB and an enhancement in data rate efficiency ~43% were achieved using the proposed algorithms. Furthermore, an extra 20% data rate enhancement was obtained by guaranteeing to keep the CDM error automatically under the system threshold. The proposed system can be extended as a general software-based platform for brain-inspired decision making in smart systems in the presence of nonlinearity and non-Gaussian characteristics. Therefore, it can easily upgrade the conventional systems to a smart one for autonomic CDM applications.

1. Introduction

The cognitive dynamic system (CDS) is inspired by the human brain neuroscience model that is built on the principles of cognition based on perception-action cycle (PAC), memory, attention, intelligence, and language [1,2,3]. The basic model of a CDS is provided in Figure 1. Also, Figure 1 shows three main subsystems of CDS: (i) perceptor for sensing the environment, (ii) the feedback channel for sending internal rewards, and (iii) the executive for performing actions on the environment. The CDS cycles through these subsystems which are known as the perception-action cycle. More details about CDS and PAC can be found in [2]. The CDS can be considered as enhanced artificial intelligence (AI). A CDS creates internal rewards and uses it to take some actions, while AI takes actions based on the rewards [2]. The CDS was proposed as an alternative to artificial intelligence in most AI applications [2,3].
In [2,3], the CDS was used for cognitive radar applications to provide optimal target detection and intelligent signal processing. In [2,3,4], the CDS was applied for cognitive radio applications, such as dynamic spectrum management in wireless communication networks [4]. The CDS was discussed for theory and applications of cognitive control [3,5]. In [6], the general concept of CDS was introduced for risk control in physical systems. Specific applications of CDS for risk control were presented for vehicular radar transmit waveform selection [7], mitigating the cyber-attack in a smart grid [8], detecting the cyber-attack in smart grid [9], and to mitigate vehicle to vehicle (V2V) jamming [10]. Also, CDS was applied as the brain of complex network [11]. A prospective block diagram of a CDS for smart homes was presented in [12]. In [13], the application of CDS for cybersecurity applications was discussed.
To date, the algorithms of CDS proposed in literature are for linear and Gaussian environments (LGE) [2,3,4,5,6,7,8,9,10,11,12,13]. Here, LGE means that the outputs of the environment linearly depend on the inputs and they have a Gaussian distribution. Therefore, the typical CDS in [2,3,4,5,6,7,8,9,10,11,12,13] uses the Kalman filter. Usually, the Kalman filter, can be applied to linear environments only. The simplified version of Shannon theory is used in [2,3,4,5,6,7,8,9,10,11,12,13] for a Gaussian environment for calculating the entropic state at time n. A typical CDS can use the entropic state at time n as the internal reward. However, Shannon’s entropy equation cannot be simplified for non-Gaussian environment. Therefore, algorithms proposed in the literature for a typical CDS cannot be applied for non-Gaussian and nonlinear environments (NGNLE). Here, NGNLE means that the outputs of the environment are not linearly dependent on the input and do not have a Gaussian distribution.
We should mention here that many AI applications [14,15,16,17,18,19,20] are for NGNLEs. For example, most data obtained or measured from health conditions, education, and social sciences are often not normally distributed [14]. Some examples of non-Gaussian distribution for health conditions, education, and social sciences are prostate cancer modeling [15,16], psychometrics [17], and labor income [18], respectively. The same holds true for a long-haul fiber-optic link as a NGNLE [19,20].
The computational complexity of CDS algorithms in [2,3,4,5,6,7,8,9,10,11,12,13] is high in a complex environment. Also, in an NGNLE, the complexity of CDS algorithms in [2,3,4,5,6,7,8,9,10,11,12,13] is higher than those for a LGE. In this paper, we develop a CDS for NGNLE with simple, straightforward, and faster PAC algorithms.
In [20,21,22], the CDS was proposed for smart fiber optic communication systems as an example of smart systems using cognitive decision-making (CDM). Here, a CDS is used to control the quality of service in long-haul fiber-optic links. The proposed CDS can take some actions such as changing the data rate so that the bit error rate (BER) is under the predefined threshold. In [20], a CDS with a simple executive was presented. The simple executive cannot predict the outcome of the actions before applying actions on the environment. Besides, the simple executive cannot control the modeling configuration of the perceptor or predict the outcome of actions using a virtual environment. In [23], a CDS is presented for NGNLE applications using advanced executive that can predict the outcome of the actions. However, the perceptor in [20,23] can only extract the environment is model without memory. In [24], a CDS was presented for real-time health situation understanding for smart e-Health home applications. Reference [24] uses a perceptor that can extract the model using the adaptive decision-making tree. However, in [24], the executive cannot predict the outcome of actions using virtual actions before applying them on a NGNLE.
In this paper, a general-purpose algorithm for the CDS for a CDM system in a NGNLE with finite memory is presented. The proposed CDS uses an advanced executive that can predict the outcome of multiple actions before applying an action to the environment. In addition, the advanced executive can change the modeling configuration of the perceptor through internal commands. Also, the perceptor can extract the model of a NGNLE with finite memory. In addition, a case study of new CDS and an algorithm are presented for a long-haul fiber-optic link as example of NGNLE. It is demonstrated that this new design can provide ~43% data-rate efficiency enhancement and 7 dB Q-factor improvement. This CDS can replace the typical soft decision forward error correction (SD-FEC) in a fiber optic link and increase the data rate efficiency. To ensure that we have a reliable communications link, the SD-FEC uses 20% overhead (OH) in the optical communications system [25]. However, the typical function of the CDS is to keep the CDM error under the system threshold. A comparison between the proposed method and some related works [20,21,23] for our long-haul fiber optic communications case study is given in Table 1.
The paper is organized as follows. In Section 2, the NGNLE with finite memory is discussed. In Section 3, the architectural structure of CDS and the proposed algorithms CDM for a general NGNLE are presented. In Section 4, we discuss the simulation results for the case study of a long-haul fiber-optic orthogonal frequency division multiplexing (OFDM) link. Finally, in Section 5, we summarize the main contributions of our work.

2. Non-Gaussian and Nonlinear Environment (NGNLE) with Finite Memory

As mentioned before, a NGNLE means that most of the observables (outputs) from the measurement are not linear functions of unknown NGNLE inputs. Also, the observables do not have a Gaussian distribution. In most NGNLE applications, we can measure the observables as the output. However, since the current situation of NGNLE is unknown, we need to decide between M discrete known situations.
Suppose X n = [ X 1 n , X 2 n , , X S n ] is a vector of variables as a function of discrete time n corresponding to S known discrete states or situations (See Figure 2). For example, in communication systems that uses QAM-4, there are four possible symbols, X 1 n , X 2 n , X 3 n   a n d   X 4 n   ( 00 ,   01   , 10 ,   11 ) which vary as a function of time. The receiver may select X 2 out of four possible symbols as the symbol that was transmitted at time n. The relation between observables and a NGNLE situation can be defined as follows:
Y n m = g n ( X 1 n , X 2 n , , X S n , w n ) ,
Y n m = [ Y n m , Y n m + 1 , , Y n ] .
Here, Y n m is set of observables that is extracted from measured signals from the sensors (e.g., [26,27]) from discrete time ( n m ) to n . Also, m is the current focus level that can be used for the reasoning with acceptable complexity. As mentioned before, X s n corresponds to the current NGNLE condition or situation, which we want to estimate. In (1), Y n m is known by measurements or another information gathering technique, but, X s n is unknown. Also, in (1), w n represents noise with an arbitrary probability density function (PDF), which could be non-Gaussian. The g n ( ) is a nonlinear mapping of X n and w n to Y n m . One example of such a system can be seen in a long-haul orthogonal frequency division multiplexing fiber optic link. In the next section, we briefly describe the OFDM fiber optic link.

2.1. OFDM-Based Fiber Optic System

OFDM has received significant attention from the researchers in optical communication systems [28,29,30] because high data rate fiber optic links are the backbone of modern communications such as the internet, TV, or medical data transmissions. Figure 3 shows the conventional OFDM-based fiber optic system. The binary data is mapped onto symbols using a symbol mapper that modulate the subcarriers using the inverse fast Fourier transform (IFFT). Usually, a cyclic prefix (CP) can be used in the guard intervals (GI) between OFDM frames to improve distortions due to chromatic dispersion (CD) [30].
The baseband OFDM data in the digital domain is converted into the optical domain using an optical IQ modulator. After propagating through a long-haul fiber optic link, the coherent receiver [31,32] converts the data into the digital domain. The fast Fourier transform (FFT) demultiplexes the data in orthogonal subcarriers, and then, the signal passes through a linear distortion and phase noise mitigation (LDPNM) block. The LDPNM can adequately compensate for CD and linear distortions. But nonlinear distortions in a fiber optic link cannot be efficiently removed by conventional nonlinear mitigation methods in practical applications. These nonlinear distortions can be deterministic such as self-phase modulation (SPM), cross-phase modulation (XPM), and four-wave mixing (FWM) between subcarriers or stochastic nonlinear impairments such as the interaction between amplified spontaneous emission (ASE) and Kerr effect leading to Gordon–Mollenauer phase noise [33].

2.2. Long-Haul Fiber Optic OFDM Link an Example of NGNLE

Typically, in the analysis of long-haul fiber optic systems, the fiber optic channel is assumed to be nonlinear and the noise is assumed to be Gaussian to simplify the problem [34]. Under the Gaussian noise assumption, it follows that the variance of nonlinear noise should be proportional to the total distance L. However, in [34], it was shown that the variance of nonlinear noise scales as L x where x ( 1 ,   2 ) depending on the dispersion map. Therefore, the Gaussian noise assumption is not accurate for OFDM systems in the presence of nonlinearity. In this paper, we investigate the performance of a long-haul fiber optic OFDM link that is an example of a NGNLE. Figure 4 shows an OFDM system with a CDS. In Figure 4, we assumed that the CDS is placed in the receiver, but the CDS can also send commands to the transmitter, such as changing the data-rate or the transmitted signal power. In addition, the intelligent subsystem of the CDS can be considered as a subsystem of the data symbol decision part of the optical transceiver.
We should mention here the conventional long-haul fiber optic system uses soft decision (SD) and hard-decision (HD) forward error correction (FEC). This leads to additional overhead to the data rate that reduces the system’s efficiency. Typically, the SD-FEC is used to guarantee the decision-making error before the HD-FEC is under the HD threshold. However, when the CDS is used for decision making as proposed in this paper, it can guarantee that the decision-making error is always under the HD-FEC threshold. Hence, the SD-FEC may no longer be needed when the CDS is used, and therefore, the additional overhead due to SD-FEC could be removed. Figure 3 shows the conventional OFDM-based fiber optic system [28] and Figure 4 shows the upgraded system with the proposed CDS.

3. Why A Cognitive Dynamic System?

In this section, we briefly discuss different machine learning methods. Then, we discuss why CDS is selected among popular machine learning schemes. In this paper, AI uses machine learning approaches to creates intelligent machines. AI created by machine learning is different from symbolic rule-based AI. Therefore, instead of programming predefined rules, AI using machine learning can learn from datasets, examples, and experiences. However, machine learning-based AI requires access to large and reliable datasets, examples, and experiences to provide reliable outcomes. This is an important weakness of machine learning based-AI approaches (including CDS) in comparison to the rule-based AI. Also, this means that the training time is generally longer than rule-based AI. In machine learning-based AI, a machine learning algorithm extracts the model from the dataset. Then, the model can be used for prediction. Also, the algorithm can learn to optimize models based on datasets and policies, for example, in a special task such as providing pre-FEC BER under 0.01 with acceptable modeling complexity.

3.1. Machine Learning Approaches

Machine learning is an interdisciplinary field in computer science, mathematics, and statistics. Generally, there are many machine learning approaches such as supervised learning, reinforcement learning, semi-supervised learning, unsupervised learning, and transfer learning. We will only focus on the first two main types supervised learning and reinforcement learning.

3.1.1. Supervised Learning

In machine learning, a very popular approach for practical applications such as predicting the length of stay in hospital, transmitted symbols in communication systems, radar target detection, or health condition is supervised learning (SL). SL can find patterns inside data. In general, the SL algorithm can learn how to create a classifier for predicting the output variable y for a given input variable x (see Figure 5a). The SL algorithm extracts a mapping function f where y = f(x). An algorithm with a set of data {x1, x2, …, xn} with the corresponding output label {y1, y2, …, yn} builds the classifier.
Supervised learning can be divided into two main branches: (1) learning by type of prediction and (2) learning by type of modeling. The type of prediction problems can be divided into regression or classification. For predicting continuous output data, regression learning methods can be more suitable. For the prediction of output class, classification algorithms such as linear regression, naive Bayes, decision trees, or support vector machines (SVM) are better choices [35]. For example, a child’s height prediction is improved with linear regression. However, the decision tree or naive Bayes are better for binary diagnostic test predictions. This type of modeling can be the extraction of a discriminative model such as decision trees and SVM algorithms. These algorithms can extract the decision boundary within the data based on the learning goal. Also, machine learning methods such as naive Bayes or Bayesian approaches, can learn the probability distributions of the data.
In summary, SL trains an algorithm on a labeled database to predict the correct outputs for the unseen inputs. SL can be applied to problems with input/output. In addition, SL can be used for prediction and classification, such as image recognition and filtering spam emails. In this paper we have used the Bayesian approach in CDS perceptor, which is a type of modeling that can extract the posteriori of the data.

3.1.2. Reinforcement Learning

Figure 5b shows a schematic diagram of reinforcement learning (RL). RL maps a decision-making problem into the interaction of a computer agent with a dynamic environment by trial and error [36]. The computer agent attempts to reach the best reward based on feedback received from the dynamic environment when it searches the state and action spaces. For example, in fiber optic communication systems, the RL algorithm will try to improve the model parameters based on iteratively simulating the state (transmitting symbol). Then, after applying the action (e.g., increasing or decreasing data rate and modeling accuracy), the agent obtains the feedback reward from the environment (real transmitted data in transmitter). The RL algorithm can then converge to a model that may generate optimal actions [37].
In summary, RL learns through trial and error from interaction with a dynamic environment such as learning to play a game or a movie/video recommendation system. There are states and actions in RL. Typically, no database is required for RL and it can find that the action for optimizing the reward. But, RL requires that a model be extracted using examples and experiences. RL receives the reward/punishment from the dynamic environment. In this paper, the executive uses RL for applying cognitive actions on a NGNLE (more details provided in the Section 3.2).

3.2. Proposed Cognitive Dynamic System

Figure 6 shows the conceptual implementation of our proposed CDS for the NGNLE (see Figure 7 later). By combining SL and RL as two main approaches of machine learning, CDS can be considered as an enhanced AI. The perceptor of the CDS can extract the model using a SL algorithm. Using the extracted model, the perceptor can generate the internal reward and predict the dynamic environment outcome [3]. In this paper, the dynamic environment is the NGNLE with finite memory (e.g., a long haul OFDM fiber optic link). The perceptor sends the internal reward to the executive through the feedback channel. RL in the executive uses the internal reward in the current PAC to find an action that can decrease/increase the internal reward for the next PAC. But, the proposed CDS can predict the outcome of the actions using a virtual NGNLE to make sure that the action can optimize the internal reward within predefined policy requirements.
In summary, unlike the typical RL, the executive of the CDS can use the extracted model by the perceptor to apply a cognitive action on the dynamic environment. The internal reward gives the CDS self-awareness, self-consciousness, and independence from the dynamic environment. Briefly, the CDS has the “conscience” about the action. For extracting the internal reward, the SL algorithm should be applied in the perceptor. Also, the prediction of cognitive actions outcome is inspired from human brain imagination and risk assessment before doing any action in the real word. However, in a typical RL, the agent applies the actions blindly to receive feedback from the environment. Therefore, the CDS is a more appropriate choice rather than a typical RL in intelligent machine applications.

4. Design of CDS Architecture for Finite Memory NGLNE

The detailed CDS architecture for decision-making is shown in Figure 7. The CDS has two main subsystems: (i) the perceptor and (ii) the executive, with a feedback channel linking them. Through interacting with the environment and shunt cycles, a PAC is created to improve the decision-making accuracy or predictions. The use of local and global feedback loops makes CDS different from artificial intelligence [3]. In this section, we first describe the perceptor and executive using related algorithms when the executive takes actions on the environment, which is followed by changing the focus level when the CDS sends the command to the perceptor as post adaptive actions. The CDS can operate in four modes: (i) pre-adaptive actions mode; (ii) environmental actions mode; (iii) changing the focus level; and (iv) steady state mode. The executive functions will be different depending on the CDS mode, and we will describe them in the following sections. The steady-state mode means that all is well and no more actions are required. Pre-adaptive mode actions are related to finding the first actions to apply on the environment (a typical CDS makes the first action randomly). Mode (ii) is related to actions that can be done on the environment to improve the decision-making accuracy. Also, for improving decision-making accuracy, the CDS can increase the focus level of modeling in the perceptor and learning in the executive in mode (iii). In addition, the CDS can decrease the focus level for the new desired complexity threshold in the mode (iii).
Figure 7 shows that the CDS can use a default database. However, it can be replaced using other techniques in some applications, such as in a long-haul nonlinear fiber optic system (See Section 4 and Figure 4) in which the database is replaced with the estimated data after forward error correction. The perceptor is responsible for knowledge extraction. The knowledge is used to reason and predict the unknown situations for decision-making. The knowledge in the perceptor is represented as a set of concepts and the relationships with the environment situations as described in Section 2. The extracted knowledge can be refined if the CDS increases the focus level. In a specific situation or on user request, the CDS switches from steady state to the pre-adaptive mode. Also, the CDS can dynamically update knowledge in the perceptor and actions library in the executive. Next, the detailed algorithmic descriptions of environmental actions mode (ii) and changing the focus level mode (iii), are presented.

4.1. Perceptor

Conventionally, the perceptor of a CDS should have Bayesian filtering [3]. In this paper, we use the Bayesian equation and the decision-making tree approach for extracting the posterior in decision-making (i.e., P ( X n | Y n m ) ), which is more straightforward than the Bayesian filtering approach. In addition, the multilayer Bayesian generative model using the decision trees concept is a typical method in machine learning. Here, we define m focus levels as the depth of the decision-making tree and i is the arbitrary decision-making node level between zero to m depth levels. This is because we need to design low-complexity and real-time algorithms for NGNLE applications. Therefore, the main sub-blocks of the perceptor are: (i) Bayesian generative model, (ii) previous model selection, and (iii) Bayesian equation and decision-making. The multilayer Bayesian generative model for different focus levels implementation is now presented.

Bayesian Generative Model Extraction

Suppose that there exists no relevant model in a model library. The three-layered Bayesian modeling block extracts the statistical model of the system (see [20,23] for further details) using a decision-making tree. The three-layered Bayesian modeling consists of three layers for an arbitrary focus level of m, as explained in (a), (b), and (c) below:
(a) Layer I:
In a NGNLE, some observables could have any value, and hence, a large number of discretized cells are required for saving vacant spaces. For example, the output of the fiber optic link is a complex variable Y n k , m , where n is the discrete time, k is the PAC number, and m is the current focus level. Therefore, we define the probability box (PB) percentage. The data normalized in this way is sent to Layer II as the Y ¯ n k , m for further processing. For the 2D observables, this box can be determined as the x min k , m and x max k , m for the minimum and maximum of the probability box in the x-axis together with y min k , m and y max k , m for minimum and maximum of the box in the y-axis. The internal reward is used to select the borders of the probabilistic box adaptively as the internal commands by the executive. More details provided on the PB parameters can be found in references [20,21,22,23].
(b) Layer II:
In Layer II, the values related to observables are discretized with discretization steps Δ x i k and Δ y i k , which can be two-dimensional if the observables are complex numbers (e.g., see [20,21,22,23]) and 0 i m . For simplicity, we define a precision factor ( P F i k ) for each focus level as (assume Δ x i k = Δ y i k ):
P F i k = 10 Δ x i k = 10 Δ y i k ,          0 i m ,
P F k , m = [ P F 0 k , P F 1 k , , P F i k , ,   P F m k ] ,
F i k = N x k , i N y k , i ,        0   i m ,
N x k , i = 10 ( x max k , i x min k , i ) P F i k ,
N y k , i = 10 ( y max k , i y min k , i ) P F i k ,
F m t o t a l , k = f = 0 m F i k ,          m { 0 , 1 , 2 , , M } , and ,    F m t o t a l ,   k C o m p l e x i t y    t h r e s h o l d .
Here, N x k , i and N y k , i are the number of discretized cells in the x-axis and y-axis, respectively. Furthermore, F i k corresponds to the number of decision-making tree branches (see Equations (3) and (5)). It should be noted that the required storage memory increases exponentially with increasing focus levels (see Equation (8)). In addition, the perceptor requires a larger database to extract a reliable model. Therefore, the complexity threshold is predefined based on the available storage memory for saving the model on hard disk or loading the model into RAM. For the desired complexity threshold, the CDS can update the current focus level m, P F k , m and probability box using Equation (8). Then, Layer II will send the discretized observables Y ^ n k , m and P F k , m to Layer III (see Equations (2) and (8)).
Y ^ n k , m = [ Y ^ n m k ( P F m k ) , Y ^ n m + 1 k ( P F m 1 k ) , ,    Y ^ n f k ( P F m f k ) , , Y ^ n k ( P F 0 ) ] .
(c) Layer III:
In this layer, the system estimates the probability of measured observables Y n k , m = [ Y n m k , Y n m + 1 k , , Y n f k , , Y n k ] for a given X n k (usually, this is known from the database), i.e., P ( Y n k , m | X n k ) by approximating it as the probability of Y ^ n k , m (see Equation (9)) for a given X ^ n k , using the Monte–Carlo method. The extracted model in this layer can be called the Bayesian model. This model is saved in the model library for future use. The executive will use the extracted statistical model for prediction by receiving them from the internal feedback to evaluate the actions virtually before applying them on the environment.

4.2. Bayesian Equation and Maximum Posteriori Selection

The posterior can be calculated using the Bayesian equation as follows:
q k ( X ^ n k , Y ^ n k , m )   = P ( X ^ n k | Y ^ n k , m ) = P ( Y ^ n k , m | X ^ n k ) P ( X ^ n k ) P ( Y ^ n k , m ) ,
b k   ( Y ^ n k , m , X ^ n k ) = P ( Y ^ n k , m | X ^ n k ) .
Here, k denotes the PAC number and n means the CDM at time n. The estimated Bayesian generative model P ( Y ^ n k , m | X ^ n k ) and the evidence P ( Y ^ n k , m ) are extracted in Layer III. The probability P ( X ^ n k ) can be estimated using the database.

4.2.1. Maximum Posteriori Selection and Assurance Factor

The CDS can save the estimated posterior P ( X ^ n k | Y ^ n k , m ) in the perceptor library after a one-time calculation. We can use the posterior P ( X ^ n k | Y ^ n k , m ) to select X ¯ n k that has the maximum probability in each discretized cell as:
X ¯ n k = arg max s = 1 , 2 , , S { P ( X ^ n k = X s | Y ^ n k , m ) } ,
where, s = 1 , 2 , , S and S is the total number of decision-making situations (see Section 2 and Equation (1)).

4.2.2. Feedback Channel, Assurance Factor, and Internal Rewards

Typically, the CDS uses the Kalman filter and Shannon theory [3] to calculate the entropic state at time n. The technique based on Kalman filter and Shannon theory is applied to a linear and Gaussian environment. However, in a NGNLE, the complexity of using this algorithm is higher than in the LGE. For our application, the algorithm should be simple, straightforward, and fast. Thus, we would like to avoid using extra logarithms, complicated functions, and integrals required in the entropic state calculation associated with Kalman filter. Therefore, here, the internal reward is inspired by the human decision-making approach with lower computational cost. Most of the time, we make the wrong decision when we assure about something whose occurrence is less than 100% (e.g., gambling). Similarly, the CDS makes a decision using Equation (12), and selects the maximum probability using the maximum a posteriori probability (MAP) rule. Therefore, the CDM error is proportional to the assurance of the decision. Consequently, we define the average assurance factor (AF) as:
A F n k = o =   n L n P ( X ¯ o k | Y ^ o k , m ) L ,        n L m
Here, L is an arbitrary time discrete interval to calculate the assurance factor. Also, we calculate the incremental of the assurance factor in the kth PAC using:
Δ k n = A F n k A F n k 1
where the A F n k 1 and A F n k are the assurance factors that can be calculated at the (k−1) and kth PAC, respectively. The internal reward, denoted as the r w n k , can be defined as the arbitrary function of A F n k and Δ k n and from the function:
r w n k = h k ( A F n k ,    Δ k n )
where the h k ( ) is an arbitrary function. The h k ( ) can be defined based on the environment type (e.g., the human body or a fiber optic channel) or the perceptor configurations such as desired computational cost or accuracy. The perceptor sends r w n k to the executive by the global feedback channel. In addition, the perceptor sends the estimated Bayesian model and evidence through the internal feedback channel (see Figure 7). Also, the perceptor receives the modeling configurations such as P F k , m (see Equation (4)) through internal feedforward channel that are internal commands from the executive. Here, P F m k is the precision factor at the kth PAC and focus level m. Therefore, the executive uses this information and it can find the proper action for applying on NGNLE. Also, for easy reference, we provide a list of the important notations and their meanings in Table 2.
Now, we provide Algorithm 1 to compute environment actions through the global feedback. In this Algorithm, we include several heading comments, so that it is easy to follow.
Algorithm 1. Environmental actions through the global feedback
Input: The observables Y n k , m , focus level m
Output: The c k + 1 as the final action in environmental actions mode
   Initialization:
   Pre-adaptive actions:
   Turn steady state mode off, environmental actions on
    c 1 f i r s t   e n v i r o m e n t a l   a c t i o n   f r o m   C
   Apply c 1 to the environment (such as fiber optic channel)
   Threshold(M), P F 1 , m = [ P F 0 1 , P F 1 1 , ,   P F m 1 ]
   1: for k = 1 to K then
   2:  Take the observable Y n k , m
   3:   if the model is not available then
   4:   if k>1 then
   5:   Predict P ( X ^ n k | Y ^ n k , m ) using b ( k 1 ) ( X ^ n k , Y ^ n k , m )
   6:   Estimate X ¯ n k by MAP rule
   7:   End if
   8:   Extract the P ( Y ^ n k , m | X ^ n k )   and   P ( Y ^ n k , m )
   9:    Calculate the posterior P ( X ^ n k | Y ^ n k , m )
   10:    Estimate X ¯ n k by decision making
   11:  Else if model is available then
   12:   Load model, evidence and posterior from perceptor library
   13:   Estimate X ¯ n k by MAP rule
   14:  End if
   15:   Send P F k , m , P ( Y ^ n k , m | X ^ n k )   and   P ( Y ^ n k , m ) to the Executive
     Internal reward
   16:  Calculate r w n k and send it to executive
     Planning
   17:   r w n ( k + 1 ) , 0 r w n k
   18:  Localize the set of all close actions to ck
     Learning
   19: If actions belong to environmental set then
   20:  for t = 1 to T then
   21:    Apply c k + 1 t virtually ( c k + 1 t C )
   22:    Calculate q k ( X n ( k + 1 ) , t , Y ^ n ( k + 1 ) , m , t )
   23:    Estimate X ¯ n ( k + 1 ) , t
   24:    Calculate r w n ( k + 1 ) , t
   25:    if r w n ( k + 1 ) , t r w n ( k + 1 ) , t , and   A F n k T h r e s h o l d ( m ) then
   26:      t arg min t = 0 , 1 , 2 , , T ( r w n ( k + 1 ) , t )
   27:      c k + 1 c k + 1 t
   28:     break for
   29:    End if
   30:   End for
   31:   if t 1 then
   32:    apply c k + 1 to the environment
   33:   else
       Policy updating
   34:    Update T h r e s h o l d   ( m + 1 )
       Planning update
   35:    Switch actions type to internal commands
   36:    Update P F k , m + 1 = [ P F 0 k , P F 1 k , ,   P F m + 1 k ]
   37:    if F m + 1 t o t a l , k c o m p l e x i t y   t h r e s h o l d then
   38:      m m + 1
   39:     Switch actions type to environmental actions
   40:    else
   41:     Turn on steady-state mode
   42:    End if
   43:  End if
   44: End for

4.3. Executive

The executive is an essential part of any CDS. It is responsible for improving the decision-making accuracy. This can be done by applying action on the NGNLE. For example, the executive can activate the actuators in the smart home [38,39] or sends the internal commands to the perceptor for changing the modeling configurations such as P F k , m . The executive provides non-monotonic reasoning to the CDS by using the internal reward and changing its focus level. The executive designed here has three parts (see Figure 7): Planner (it consists of the actions library also), policy, and learning using prediction of the outcome of the virtual action [21,23]. Next, we describe the planner and policy followed by learning using prediction.

4.3.1. Planner and Policy

The planner extracts the set of prospective actions that are already saved in CDS actions library. Also, the planner selects the first action in the first PAC using pre-adaptive actions. In addition, the planner updates the actions type and sends internal commands to update P F k , m (see Equation (4)) through the internal feedback and feedforward channels (i.e., the shunt cycle).
In this paper, the policy determines the desired goals that the CDS should achieve using the PAC. For example, the goal can be the trade-off between the desired CDM accuracy and the computational cost. Here, for simplicity, we define the CDS goal as the T h r e s h o l d m which means the CDS decision making accuracy goal in focus level m for the desired complexity threshold. For example, the T h r e s h o l d m = 0 can be 10 percent acceptable decision-making error. The T h r e s h o l d m = 1 can also be five percent acceptable decision-making error, but the cost is increased modeling complexity.

4.3.2. Learning Using Prediction

(a) Executive actions
The actions space in the executive can be classified as the environmental actions and internal CDS commands. Therefore, we define the actions space as follows:
C = { c k | c k = E n v i r o n m e n t a l   a c t i o n s ,   or I n t e r n a l   c o m m a n d s = { Increase   m , m < M Decrease   m , m 1 } ,
where ck is the action in the current PAC number k and C is set of all possible actions in the actions space.
The posteriori due to the virtual environmental action of c k + 1 t (see Table 1) can be predicted virtually using:
r a t i o k + 1 , t = max s = 1 , 2 , , S { s t d ( Y ^ n k , m | X ^ n k = X s ) s t d ( Y ^ n k 1 , m | X ^ n k 1 = X s ) } ,
Y ^ n ( k + 1 ) , t , m = f ( Y ^ n k , m ,    r a t i o k + 1 , t , c k , c k 1 , t , m ) ,
P ( X ^ n ( k + 1 ) , t | Y ^ n ( k + 1 ) , t , m ) = P ( Y ^ n ( k + 1 ) , t , m | X ^ n ( k + 1 ) , t ) P ( X ^ n ( k + 1 ) , t ) P ( Y ^ n ( k + 1 ) , t , m ) , o r q k   ( X ^ n ( k + 1 ) , t | Y ^ n ( k + 1 ) , t , m ) = P ( X ^ n ( k + 1 ) , t | Y ^ n ( k + 1 ) , t , m ) .
Here, t { 1 , 2 , , T } and T is a total number of desired virtual actions that the kth model is still valid for predicting the Tth virtual action. Also, in Equation (19), q k ( ) and b k ( ) (see Equations (10) and (11)) are received through the internal feedback from the perceptor. Therefore, the predicted assurance factor can be calculated as:
A F n ( k + 1 ) , t = o =   n L n P ( X ^ o ( k + 1 ) , t | Y ^ o ( k + 1 ) , t , m ) L ,          n L m ,
and
Δ k + 1 , t n = A F n k + 1 , t A F n k + 1 , t 1 .
Then, the predicted internal rewards for the desired virtual action c k + 1 t can be calculated using Equation (15) as follows:
r w n ( k + 1 ) , t = { h ( k + 1 ) ( A F n ( k + 1 ) , t ,    Δ k + 1 , t n ) , t { 1 , 2 , , T } r w n k t = 0 .
Therefore, we can find the action c k + 1 T that brings the minimum internal reward as:
T =   arg min t { 0 , 1 , 2 , , T } ( r w n ( k + 1 ) , t ) .
As a result, the actions to be applied on the environment can be selected as:
c k + 1 =   c k + 1 T ,         f o r    T 1 .
Therefore, c k + 1 is the best action to be applied on the environment to improve the decision-making performance. Algorithm 1 shows the outline of the main processes for the global PAC of our CDS. In some applications, the policy may update the T h r e s h o l d m for more accuracy (or even higher complexity with acceptable accuracy).
(b) Internal Commands
As mentioned before, some actions apply on the environment and the CDS receives the internal reward through the global PAC using the main feedback channel. However, the actions can be the internal commands such as the increase or decrease the focus level, which is a trade-off between the computational cost and CDM accuracy (see Equations (4), (8), and (16)).

5. CDS Implementation Case Study: OFDM Long-Haul Fiber Optic Communications

Figure 8 shows the CDS architecture block diagram for an OFDM long-haul fiber optic link. As mentioned before, we select the fiber optic link as an example of a NGNLE. In general, X ^ n k can be supplied by a database during a training period. However, the estimated X ^ n k  after FEC can be used instead of the database shown in Figure 7 in fiber optic link for model extraction. The bit error rate after forward error correction (FEC) is between 10−15 and 10−12 [34] when the BER before FEC is less than the FEC-threshold. In addition, the decision-making error can be calculated as the BER. Therefore, the BER can be estimated in two ways. First, if the BER is less than the FEC threshold, then the BER after the FEC can be measured and the BER before FEC can be found using extrinsic information transfer (EXIT) chart [25]. If the BER is more than the FEC threshold, we can estimate the BER using 16 or 32 training frames that are used for linear distortion mitigation. The channel estimation and compensation of linear impairments is done using the linear adaptive channel equalizer that uses 16 training frames to mitigate linear distortion [40]. Therefore, the prospective internal reward can be calculated as follows (see Equation (22) and [20,21,22,23]):
r w n k , t = B E R k + t n ( R t k R r e f ) 4 ,
R t k = R r e f + t × d ,
where d is 4 Gb/s discretization step for data rate and R r e f is an available reference data rate that is set equal to ( R 1 1 d ) . Also, the implementation of Equations (17)–(19) for a fiber optic communication system is presented in [21,23].
The actions space for this specific case study is (see Equation (16)):
C = { c k | c k = d a t a   r a t e    R k > R r e f ,   or I n t e r n a l   c o m m a n d s = { m = 1 , m = 0 }

5.1. Simulations Parameters

We illustrate the modeling process by performing a numerical simulation of an OFDM system (see Figure 8). Signal propagation in an optical fiber is described by the nonlinear Schrödinger equation (NLSE), which is solved using the standard split-step Fourier scheme [34]. The output of the linear equalizer is passed to the perceptor (see Figure 4 and Figure 8). X n k represents the transmitted symbols which takes on the values from the following set X n k { ± 3 ± 3 j , ± 1 ± 1 j , ± 1 ± 3 j , ± 3 ± 1 j } with equal probability and Y n k represents the corresponding received symbols after linear equalization.
The simulation parameters of the fiber optic system are presented in Table 3. The CDS parameters are as follows: P F i = 0 k = 2 , P F i = 1 k = 5 , and 96% probability box. Also, the T h r e s h o l d m = 0 which is the FEC threshold in focus level zero is 1.03 × 10−2 (with 14.3% OH for the HD) [41]. Typically, to guarantee reliable communication over a fiber optic link, ~20% SD-FEC overhead is used [25]. However, the CDS can replace the SD-FEC, and with this replacement, the data rate can be enhanced by ~20%. This is because, the CDS (as presented in Algorithm 1) can keep the BER under FEC threshold before HD.

5.2. Simulation Results and Discussions

5.2.1. Pre-Adaptive Actions

The BER of a conventional fiber optic communications system (i.e., without CDS, see Figure 3) is shown in Figure 9. We consider 1.03 × 10−2 (with 14.3% OH) as the FEC threshold or Threshold (m = 0) for HD [25] (see Table 3). Thus, we can see from Figure 9 that the maximum achievable data rate is 208 Gb/s. The CDS uses 208 Gb/s for j = 1 (i.e., first data rate, R 1 1 = 208 Gb/s). Also, we assume data rate variation by discretization step d = 4 Gb/sand a step of 1 dB for the launch power. Therefore, the data rate for PAC k = 0 is Rref = 204 Gb/s (see Equation (26)). The optimum launch power can be also found using the pre-adaptive method this is already used in the conventional system.

5.2.2. Results for the System with the CDS Focus Level One

The global PAC procedure can be summarized using the results shown in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 as follows:
  • First PAC (focus level zero): CDS reduces the date rate to find the first data rate for which the BER is under the FEC-threshold (see Figure 9 and Section 4.2.1). Therefore, 208 Gb/s is applied to the fiber optic link as the first action (c1). Because the BER is under the FEC threshold, the perceptor can extract the statistical model for 208 Gb/s. The perceptor can improve the BER using the extracted model (see Figure 9, c1:208 Gb/s). Note that the minimum BER for a conventional system at 208 Gb/s is 8.0 × 10 3 (see Figure 9) whereas this is reduced to 5.5 × 10 4 when the CDS is used (see Figure 10, c1:208 Gb/s). The perceptor sends the extracted model to the executive using the internal feedback link.
  • Second PAC (focus level zero): The executive predicts the virtual internal rewards and BER for virtual action of c2 (see Figure 8 and Figure 11 as well as Algorithm 1). Then, the executive applies 236 Gb/s as the c2, because the virtual BER for 240 Gb/s at t = 8 is more than the FEC-threshold of 10 2 (see Figure 11b, c2: prediction). Also, 236 Gb/s has the minimum virtual internal rewards (See Figure 11a, c2: prediction) and at this data rate, the BER is under FEC threshold (See Figure 11b, c2: prediction). After applying c2 by the CDS (i.e., transmitter sends the data at 236 Gb/s), the perceptor uses the 208 Gb/s model (see Figure 8 and Algorithm 1, previous model selection) to predict the BER for 236 Gb/s. However, the BER is more than FEC-threshold (see Figure 10, c2: 236 Gb/s prediction, topmost curve). Note that we used the previous model selection in this case study. By using previous model selection, the system does not need to transmit the training frames for model extraction.
  • Third PAC (focus level zero): The BER is more than the FEC-threshold for c2. Therefore, the executive reduces the data rate by d = 4 Gb/s (see Figure 11). As a result, executive applies 232 Gb/s as c3 in third PAC number. Then, the perceptor uses 208 Gb/s model for BER estimation of 232 Gb/s (See Figure 10, c3: 232 Gb/s prediction) and it can bring the BER for 232 Gb/s under the FEC-threshold. Therefore, the perceptor can extract the model for 232 Gb/s. The extracted model results in further BER improvement for 232 Gb/s (see Figure 10, c3: 232 Gb/s).
  • Fourth PAC (focus level zero): The executive applies the 236 Gb/s (See Figure 11, c4: prediction) as c4 to the environment (i.e., fiber optic link) because the predicted BER for 240 Gb/s is more than the FEC-threshold (see Figure 11b, c4: prediction) and virtual internal reward increases for 240 Gb/s (see Figure 11a, c4: prediction). Figure 10 (c4: 236 Gb/s prediction) shows that the perceptor provides the BER for 236 Gb/s under FEC-threshold using the model for 232 Gb/s (provided that the model for 208 Gb/s was used and BER was higher than FEC threshold in second PAC). Like the third PAC, the BER can improve further using the exact extracted model (see Figure 10, c4: 236 Gb/s).
  • Fifth PAC (focus level zero and focus level one): The executive finds that the predicted BER for 240 Gb/s is just less than the FEC threshold (see Figure 11b, c5: prediction), but the virtual internal reward increases in comparison to 236 Gb/s (see Figure 11a, c5: prediction). Therefore, the executive updates policy and FEC threshold or T h r e s h o l d m = 1 = 4.7 × 10 3 . Then, the executive sends the new policy ( T h r e s h o l d m = 1 = 4.7 × 10 3 ) and the internal command (increasing focus level) to the perceptor. Using, the model extracted in focus level one, the perceptor provides the BER under updated FEC threshold (see Figure 12, c5: 236 Gb/s and Figure 14b). Also, the internal rewards decreased correspondingly (see Figure 14a).
  • Sixth PAC (focus level one): The executive predicts the virtual internal rewards and the BER for virtual action of c6 using extended model in focus level 1 (see Figure 12, Figure 13, and Algorithm 1). The executive applies 252 Gb/s as the c6, because the virtual BER for 256 Gb/s at t = 5 is more than the updated FEC-threshold (see Figure 13b, c6: prediction) and the 252 Gb/s has the minimum virtual internal rewards (See Figure 13a, c6: prediction) with virtual BER under updated FEC threshold (See Figure 13b, c6: prediction). After applying c6 by the CDS, the perceptor uses 236 Gb/s extracted model (see Figure 12, c6: 252 Gb/s prediction) to predict the BER for 252 Gb/s. Figure 12 shows that the perceptor provides the BER for 252 Gb/s under the updated FEC-threshold using the extracted model for 236 Gb/s. Similar to previous PAC, the BER can be improved further using the exact extracted model (see Figure 12, c6: 252 Gb/s).
  • Seventh, eighth and ninth PAC (focus level one): Similar to sixth PAC, the CDS can provide 260 Gb/s, 272 Gb/s, and 280 Gb/s for the c7, c8, and c9, respectively.
  • 10th PAC (focus level one): The executive finds that the predicted BER for 284 Gb/s is more than the updated FEC threshold (see Figure 13b). As a result, the executive turns on steady state for 280 Gb/s. The CDS does not increase the focus level, because the number of cells in the focus level two is more than the desired complexity threshold (see Table 3 for F m = 2 t o t a l , k = 10 ).
From Figure 12 and Figure 14b, we see that the CDS provides the minimum BER of 4.7 × 10 3 before HD-FEC, when the data rate is 280 Gb/s (see Figure 12 and Figure 14b). Based on [25], the required OH can be reduced to 6.25%. As a result, an extra ~24 Gb/s in 280 Gb/s data rate can be used for data transmission instead of assigning it for overhead.

5.2.3. CDS Learning Curve

The CDS learning curve is shown in Figure 14a and corresponding BER is shown in Figure 14b. In Figure 14, the PACs 1, 2, 3, and 4 correspond to focus level 0 actions which correspond to the data rate of 208, 236, 232, and 236 Gb/s, respectively (see Algorithm 1). The PACs 5, 6, 7, 8, and 9 are the CDS focus level 1 actions as data rates of 236, 252, 260, 272, and 280 Gb/s, respectively.
The results for the CDS at steady state are summarized in Table 4. The smart fiber optic link that used the focus level concept enhances the data rate efficiency by ~43% in comparison to conventional fiber optic link. This is the summation of ~35% data rate enhancement (i.e., enhancement of the data rate R1 = 208 Gb/s to R9 = 280 Gb/s) and 8% of HD overhead reduction (~24 Gb/s can be used for data carrying, instead of overhead redundancy) reduction, which is achieved by lowering the FEC threshold from 1.02 × 10 2 to 4.7 × 10 3 . The final Q-factor improvement is 7 dB, which is a significant improvement for the long-haul fiber optic link. The reason for such significant improvement is that the fiber optic channel using memory and the focus level concept can model the fiber optic channel more accurately in comparison to a memoryless fiber optic link channel. Under the same conditions, the performance improvement using the digital backpropagation (DBP) method is typically 1–2 dB [42] and the computational cost and complexity is significantly higher compared to our proposed system. In addition, the DBP method does not provide data rate enhancement.

5.2.4. CDS Complexity for Proof-Concept Case Study on Long-Haul OFDM Fiber Optic Link

In order to calculate the computational cost, we distinguish two types of CDS modes: (i) Modeling extraction mode and (ii) Steady-state mode. The complexity of a CDS varies depending on the perceptor mode. When the CDS operates in the BER improvement mode, the BER converges after N OFDM frames (in simulation N = 28,000) and each frame has D data-carrying subcarriers (From Table 1, it is 126). To create the histogram (i.e., to calculate P ( X n | Y n m ) ) based on the discretized data, we need N × D simple additions (just counting the samples within the tiny surface). To calculate P ( X n | Y n m ) , we need S real multiplications where S is the order of QAM (in our example, S = 16), then, we save the maximum probability for the related QAM alphabet. For simplicity, we assume that the minimum and maximum of the probability box in all focus levels are y min k , m = x min k , m = 6 and y max k , m = x max k , m = 6 , respectively for this proof-of-concept example. As a result, the storage requirement depends on the precision factor. For example, with a P F k , m = 0 of 2 in focus level zero, we need to store a matrix of dimension N x k , 0 × N y k , 0 = 60 × 60 (see, Equations (6) and (7), max ( F m = 0 t o t a l , k ) = 3600 ). Also, the storage requirement in focus level one with a P F k , m = 0 of 2 and a P F k , m = 1 of 5, we need to store the matrix which has a dimension of N x k , 1 × N y k , 1 × N x k , 0 × N y k , 0 = 12 by 12 by 60 by 60 (see, Equations (6) and (7), max ( F m = 1 t o t a l , k ) = 518 , 400 ). These matrices can be stored in a hard disk and can be loaded into RAM when needed. Therefore, the complexity threshold (here, it is 106) is the predefined maximum available storage for saving in a hard disk or loading into RAM, when needed. When the CDS operates in the steady-state mode, there is no cost associated with model extraction (i.e., to calculate P ( X n | Y n m ) ). Therefore, no multiplication or addition is needed in steady state.
For validating, we need to process 28,000 frames (see Table 3) using digital back propagation (DBP), and this cannot be done using typical processors. Therefore, we assumed the DBP processing the 28,000 frames as ~55 blocks of 512 frames. The computational cost associated with DBP for processing 512 OFDM frames is as follows. In each propagation step, J log 2 ( J ) complex multiplications are required where J is the number of samples in time domain. If there are B propagation steps, the complex multiplication is O ( B × J log 2 J   ) [43,44]. For example, for a 20-span system, assuming 10 steps per span and with J = 16,384, we need 45,875,200 complex multiplications! Besides, this computation needs to be done continuously for each block of data. Three complex vectors of length J (the total size of 98,304) needs to be stored. In contrast, most of the computational complexity of the CDS is present only when there is a fluctuation in the fiber optic channel. Under steady-state, the computational cost is minimal.
It is useful to compare the complexity of the proposed method in this paper with the DBP method for Q-factor enhancement using the simulation parameters in Table 3 and processing 28,000 OFDM frames. We run both algorithms on a Microsoft Surface-Pro with Intel ® Core ™ i706650U CPU @ 2.20 GHz 2.21 GHz, 16 GB RAM, system type 64-bit Operating System x64-Based processor using MATLAB. For the DBP algorithm [42], the runtime is 2465.8 s, that is ~41 min and 6s. However, this runtime for 9 PACs with the proposed CDS is 792s (~13 min and 12s). In addition, the CDS running time is reduced to 100s for the CDM in steady-state mode. In contrast, the DBP run time will always remain similar for any continuous data stream. We should mention that the CDS can process all 28,000 frames as one block. However, processing 28,000 frames using DBP method is not feasible using Microsoft Surface-Pro with Intel ® Core ™ i706650U CPU @ 2.20GHz 2.21 GHz, 16 GB RAM. Therefore, we break the 28,000 frames into ~55 blocks, and the DBP run time of ~41 min was achieved in this way. However, we should mention that processing received 28,000 frames in less than 100s in the steady-state mode makes it a practical solution for almost real-time high-quality data transmission such as real-time high-quality video conferencing.

6. Conclusions

In recent years, some efforts were made to develop intelligent machines using cognitive dynamic system for artificial intelligence applications. However, currently, CDS high complexity algorithms available are only applied to LGE. These algorithms cannot be applied for a non-Gaussian and nonlinear environment with finite memory. Therefore, we proposed novel CDS algorithms that could be applied on a NGNLE with finite memory. We redesigned the CDS in detail to make it suitable for NGNLE cognitive decision making applications. Also, the system architecture and algorithms are presented. The proposed algorithm is applied on a long-haul nonlinear fiber-optic link as a case study. By upgrading the conventional link with the CDS, the applicable data-rate increased from 208 Gb/s to 280 Gb/s. Also, the CDS Q-factor improved by ~7 dB (the optimum BER = 4.4 × 10 3 for 280 Gb/s with CDS). The data rate enhancement is ~35% without considering the HD overhead reduction. However, the CDS brings the BER under the new FEC threshold (= 4.7 × 10 3 ), we do not need 8% of the HD-FEC overhead. Hence, the net data rate enhancement is ~43%. Typically, SD-FEC with as overhead of ~20% is used to guarantee that the BER is under the HD-FEC threshold. An additional advantage of the CDS is that SD-FEC overhead is no longer needed since the CDS tries to keep the BER under the HD-FEC threshold automatically.
Finally, the presented algorithms in this paper are simple and fast because it used the Bayesian equation for calculating the posterior that has low complexity for the calculation. Therefore, the low complexity algorithms make the proposed CDS a perfect candidate for the NGNLE applications such as long-haul fiber optic links or in healthcare.

Author Contributions

All authors contribute equally for writing this paper. All authors have read and agreed to the published version of the manuscript.

Acknowledgments

We would thank Natural Sciences and Engineering Research Council (NSERC) and Canada Research Chair Program (CRCP) for supporting this research work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fuster, J.M. Cortex and Mind: Unifying Cognition; Oxford University: Oxford, UK, 2003. [Google Scholar]
  2. Haykin, S. Cognitive Dynamic Systems: Perception-Action Cycle, Radar, and Radio; Cambridge University: Cambridge, UK, 2012. [Google Scholar]
  3. Haykin, S. Cognitive Dynamic Systems: Radar, Control, and Radio. Proc. IEEE 2012, 100, 2095–2103. [Google Scholar] [CrossRef]
  4. Khozeimeh, F.; Haykin, S. Brain-Inspired Dynamic Spectrum Management for Cognitive Radio Ad Hoc Networks. IEEE Trans. Wirel. Commun. 2012, 11, 3509–3517. [Google Scholar] [CrossRef]
  5. Fatemi, M.; Haykin, S. Cognitive Control: Theory and Application. IEEE Access 2014, 2, 698–710. [Google Scholar] [CrossRef]
  6. Haykin, S. The Cognitive Dynamic System for Risk Control. Proc. IEEE 2017, 105, 1470–1473. [Google Scholar] [CrossRef]
  7. Feng, S.; Haykin, S. Cognitive Risk Control for Transmit-Waveform Selection in Vehicular Radar Systems. IEEE Trans. Veh. Technol. 2018, 67, 9542–9556. [Google Scholar] [CrossRef]
  8. Oozeer, M.I.; Haykin, S. Cognitive Risk Control for Mitigating Cyber-Attack in Smart Grid. IEEE Access 2019, 7, 125806–125826. [Google Scholar] [CrossRef]
  9. Oozeer, M.I.; Haykin, S. Cognitive Dynamic System for Control and Cyber-Attack Detection in Smart Grid. IEEE Access 2019, 7, 78320–78335. [Google Scholar] [CrossRef]
  10. Feng, S.; Haykin, S. Cognitive Risk Control for Anti-Jamming V2V Communications in Autonomous Vehicle Networks. IEEE Trans. Veh. Technol. 2019, 68, 9920–9934. [Google Scholar] [CrossRef]
  11. Fatemi, M.; Setoodeh, P.; Haykin, S. Observability of stochastic complex networks under the supervision of cognitive dynamic systems. J. Complex Netw. 2017, 5, 433–460. [Google Scholar] [CrossRef] [Green Version]
  12. Feng, S.; Setoodeh, P.; Haykin, S. Smart Home: Cognitive Interactive People-Centric Internet of Things. IEEE Commun. Mag. 2017, 55, 34–39. [Google Scholar] [CrossRef]
  13. Haykin, S. Artificial Intelligence Communicates with Cognitive Dynamic System for Cybersecurity. IEEE Trans. Cogn. Commun. Netw. 2019, 5, 463–475. [Google Scholar] [CrossRef]
  14. Bono, R.; Blanca, M.J.; Arnau, J.; Gómez-Benito, J. Non-normal Distributions Commonly Used in Health, Education, and Social Sciences: A Systematic Review. Front. Psychol. 2017, 8, 1602. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Si, Y.; Liu, R.B. Diagnostic Performance of Monoexponential DWI Versus Diffusion Kurtosis Imaging in Prostate Cancer: A Systematic Review and Meta-Analysis. Am. J. Roentgenol. 2018, 211, 358–368. [Google Scholar] [CrossRef] [PubMed]
  16. Maurer, M.H.; Heverhagen, J.T. Diffusion weighted imaging of the prostate-principles, application, and advances. Transl. Androl. Urol. 2017, 6, 490–498. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Micceri, T. The unicorn, the normal curve, and other improbable creatures. Psychol. Bull. 1989, 105, 156–166. [Google Scholar] [CrossRef]
  18. Diaz-Serrano, L. Labor income uncertainty, skewness and homeownership: A panel data study for Germany and Spain. J. Urban Econ. 2005, 58, 156–176. [Google Scholar] [CrossRef]
  19. Shahi, S.N.; Kumar, S.; Liang, X. Analytical modeling of cross-phase modulation in coherent fiber-optic system. Opt. Express 2014, 22, 1426–1439. [Google Scholar] [CrossRef]
  20. Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Brain Inspired Dynamic System for the Quality of Service Control Over the Long-Haul Nonlinear Fiber-Optic Link. Sensors 2019, 19, 2175. [Google Scholar] [CrossRef] [Green Version]
  21. Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Cognitive decision making for the long-haul fiber optic communication systems. In Proceedings of the 2019 16th Canadian Workshop on Information Theory (CWIT), Hamilton, ON, Canada, 2–5 June 2019; pp. 1–5. [Google Scholar]
  22. Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Smart long-haul fiber optic communication systems using brain-like intelligence. In Proceedings of the 2019 16th Canadian Workshop on Information Theory (CWIT), Hamilton, ON, Canada, 2–5 June 2019; pp. 1–6. [Google Scholar]
  23. Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Brain-Inspired Cognitive Decision Making for Nonlinear and Non-Gaussian Environments. IEEE Access 2019, 7, 180910–180922. [Google Scholar] [CrossRef]
  24. Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Brain-Inspired Intelligence for Real-Time Health Situation Understanding in Smart e-Health Home Applications. IEEE Access 2019, 7, 180106–180126. [Google Scholar] [CrossRef]
  25. Zhang, L.M.; Kschischang, F.R. Staircase Codes With 6% to 33% Overhead. J. Lightwave Technol. 2014, 32, 1999–2002. [Google Scholar] [CrossRef] [Green Version]
  26. Majumder, S.; Mondal, T.; Deen, M.J. Wearable sensors for remote health monitoring. Sensors 2017, 17, 130. [Google Scholar] [CrossRef] [PubMed]
  27. Majumder, S.; Mondal, T.; Deen, M.J. Smartphone sensors for health monitoring and diagnosis. Sensors 2019, 19, 2164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Jansen, S.L.; Morita, I.; Takeda, N.; Tanaka, H. 20 Gb/s OFDM transmission over 4160-km SSMF enabled by RF-pilot tone phase noise compensation. In Proceedings of the PDP15, NFOEC, Anaheim, CA, USA, 25–29 March 2007. [Google Scholar]
  29. Du, A.; Schmidt, B.; Lowery, A. Efficient digital backpropagation for PDM-CO-OFDM optical transmission systems. In Proceedings of the OFC/NFOEC, OTuE2, San Diego, CA, USA, 21–25 March 2010. [Google Scholar]
  30. Peng, W.; Takahashi, H.; Morita, I.; Tsuritani, T. Per-symbol-based DBP approach for PDM-CO-OFDM transmission systems. Opt. Express 2013, 21, 1547–1554. [Google Scholar] [CrossRef] [PubMed]
  31. Bandyopadhyay, A.; Deen, M.J.; Tarof, L.E.; Clark, W. A simplified approach to time-domain modeling of avalanche photodiodes. IEEE J. Quantum Electron. 1998, 34, 691–699. [Google Scholar] [CrossRef]
  32. Deen, M.J.; Basu, P.K. Silicon Photonics: Fundamentals and Devices; John Wiley & Sons: New York, NY, USA, 2010. [Google Scholar]
  33. Gordon, J.P.; Mollenauer, L.F. Phase noise in photonic communications systems using linear amplifiers. Opt. Lett. 1990, 15, 1351–1353. [Google Scholar] [CrossRef]
  34. Skidin, A.S.; Sidelnikov, O.S.; Fedoruk, M.P.; Turitsyn, S.K. Mitigation of nonlinear transmission effects for OFDM 16-QAM optical signal using adaptive modulation. Opt. Express 2016, 24, 30296–30308. [Google Scholar] [CrossRef] [Green Version]
  35. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  36. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
  37. Raghu, A.; Komorowski, M.; Celi, L.A.; Szolovits, P.; Ghassemi, M. Continuous state-space models for optimal sepsis treatment-a deep reinforcement learning approach. In Proceedings of the Machine Learning for Healthcare (MLHC) 2017, Boston, MA, USA, 18–9 August 2017; pp. 1–17. [Google Scholar]
  38. Majumder, S.; Aghayi, E.; Noferesti, M.; Memarzadeh-Tehran, H.; Mondal, T.; Pang, Z.; Deen, M.J. Smart homes for elderly healthcare—Recent advances and research challenges. Sensors 2017, 17, 2496. [Google Scholar] [CrossRef] [Green Version]
  39. Deen, M.J. Information and communications technologies for elderly ubiquitous healthcare in a smart home. Pers. Ubiquitous Comput. 2015, 19, 573–599. [Google Scholar] [CrossRef]
  40. Kumar, S.; Deen, M.J. Fiber Optic Communications: Fundamentals and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2014. [Google Scholar]
  41. Mizuochi, T.; Miyata, Y.; Kubo, K.; Sugihara, T.; Onohara, K.; Yoshida, H. Progress in soft-decision FEC. In Proceedings of the 2011 Optical Fiber Communication Conference and Exposition and the National Fiber Optic Engineers Conference, Los Angeles, CA, USA, 6–10 March 2011; pp. 1–3. [Google Scholar]
  42. Liang, X.; Kumar, S. Correlated digital backpropagation based on perturbation theory. Opt. Express 2015, 23, 14655–14665. [Google Scholar] [CrossRef] [PubMed]
  43. Ip, E.; Kahn, J.M. Compensation of dispersion and nonlinear impairments using digital backpropagation. J. Lightwave Technol. 2008, 26, 3416–3425. [Google Scholar] [CrossRef]
  44. Mateo, E.; Zhu, L.; Li, G. Impact of XPM and FWM on the digital implementation of impairment compensation for WDM transmission using backward propagation. Opt. Express 2008, 16, 16124–16137. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Block diagram of cognitive dynamic system (CDS).
Figure 1. Block diagram of cognitive dynamic system (CDS).
Applsci 10 01150 g001
Figure 2. Decision making between S known situations using m focus levels.
Figure 2. Decision making between S known situations using m focus levels.
Applsci 10 01150 g002
Figure 3. The conventional orthogonal frequency division multiplexing (OFDM)-based fiber optic system. (ADC: Analog to Digital Converter, DAC: Digital to Analog Converter, FEC: Forward Error Correction, FFT: Fast Fourier Transform, IFFT: Inverse Fast Fourier Transform, P/S: Parallel to Serial, and S/P: Serial to Parallel).
Figure 3. The conventional orthogonal frequency division multiplexing (OFDM)-based fiber optic system. (ADC: Analog to Digital Converter, DAC: Digital to Analog Converter, FEC: Forward Error Correction, FFT: Fast Fourier Transform, IFFT: Inverse Fast Fourier Transform, P/S: Parallel to Serial, and S/P: Serial to Parallel).
Applsci 10 01150 g003
Figure 4. The OFDM-based fiber optic system enhanced with the CDS.
Figure 4. The OFDM-based fiber optic system enhanced with the CDS.
Applsci 10 01150 g004
Figure 5. Schematic of two popular machine learning approaches; (a) supervised learning (SL) and (b) reinforcement learning (RL).
Figure 5. Schematic of two popular machine learning approaches; (a) supervised learning (SL) and (b) reinforcement learning (RL).
Applsci 10 01150 g005
Figure 6. Conceptual implementation of proposed CDS.
Figure 6. Conceptual implementation of proposed CDS.
Applsci 10 01150 g006
Figure 7. Architecture of proposed CDS for cognitive decision-making (CDM) in non-Gaussian and nonlinear environments (NGNLE).
Figure 7. Architecture of proposed CDS for cognitive decision-making (CDM) in non-Gaussian and nonlinear environments (NGNLE).
Applsci 10 01150 g007
Figure 8. CDS architecture for long-haul NGNLE OFDM-based fiber optic link.
Figure 8. CDS architecture for long-haul NGNLE OFDM-based fiber optic link.
Applsci 10 01150 g008
Figure 9. Bit error rate (BER) for a conventional system with a linear equalizer.
Figure 9. Bit error rate (BER) for a conventional system with a linear equalizer.
Applsci 10 01150 g009
Figure 10. BER in perceptor after environmental actions (PF = 2, d = 4 Gb/s, focus level zero).
Figure 10. BER in perceptor after environmental actions (PF = 2, d = 4 Gb/s, focus level zero).
Applsci 10 01150 g010
Figure 11. BER using CDS in executive prediction for virtual actions (PF = 2, d = 4 Gb/s, focus level zero); (a) learning curve for virtual internal reward and (b) virtual BER versus executive internal cycle).
Figure 11. BER using CDS in executive prediction for virtual actions (PF = 2, d = 4 Gb/s, focus level zero); (a) learning curve for virtual internal reward and (b) virtual BER versus executive internal cycle).
Applsci 10 01150 g011
Figure 12. BER in perceptor after environmental actions (PF = 2, d = 4 Gb/s, focus level one). Dash lines: BER prediction in perceptor and continuous lines: BER using exact model.
Figure 12. BER in perceptor after environmental actions (PF = 2, d = 4 Gb/s, focus level one). Dash lines: BER prediction in perceptor and continuous lines: BER using exact model.
Applsci 10 01150 g012
Figure 13. BER using CDS in executive prediction for virtual actions (PF = 2, d = 4 Gb/s, focus level one); (a) learning curve for virtual internal reward and (b) virtual BER versus executive internal cycle).
Figure 13. BER using CDS in executive prediction for virtual actions (PF = 2, d = 4 Gb/s, focus level one); (a) learning curve for virtual internal reward and (b) virtual BER versus executive internal cycle).
Applsci 10 01150 g013
Figure 14. Results for the CDS versus perception-action cycle (PAC) numbers; (a) CDS learning curve and (b) BER versus PAC numbers.
Figure 14. Results for the CDS versus perception-action cycle (PAC) numbers; (a) CDS learning curve and (b) BER versus PAC numbers.
Applsci 10 01150 g014
Table 1. Comparison between proposed work and related work published in literature.
Table 1. Comparison between proposed work and related work published in literature.
ReferencesTechnique ImplementedQ-Factor ImprovementData Rate Enhancement
[20]Simple CDS2.7 dB13%
[21,23]CDS with action outcome prediction (Virtual actions)3.5 dB23%
This workNew CDS using different focus levels adaptively and Virtual actions7 dB43%
Table 2. The important notation list used in this paper.
Table 2. The important notation list used in this paper.
NotationsValue
KThe total number of environmental perception-action cycle (PACs)
kThe current PAC number (environmental)
MThe Maximum possible focus level
mThe current focus level
F m t o t a l , k The total number of discretized cells at kth PAC and mth focus level
CThe set of all possible actions in actions library
XThe set of all discrete situations of NGNLE
SMaximum number of the situations for CDM
LThe discrete time interval for AF calculation
X n k The current NGNLE situation, at time n
X ^ n k The training data from database using for modeling
X ¯ n k The estimated decision by the system
Y n k , m The observable at time n and kth PAC
Y ¯ n k , m The normalized observables at time n and kth PAC
Y ^ n k , m The discretized observables at time n and kth PAC
A F n k The assurance factor at time n and kth PAC
r w n k The internal reward at time n and kth PAC
q k ( X ^ n k , Y ^ n k , m ) The estimated posterior at time n and kth PAC
b k ( X ^ n k , Y ^ n k , m ) The estimated Bayesian model at time n and kth PAC
c k Current action applied on NGNLE
T h r e s h o l d ( m ) The predefined threshold at focus level m
T The number of virtual actions can predict by current available model at time n and kth PAC
P F m k The precision factor at kth PAC and focus level m
q k ( X ^ n ( k + 1 ) , t , Y ^ n ( k + 1 ) , t , m ) The tth prediction using posterior at kth PAC
A F n ( k + 1 ) , t The tth assurance factor prediction in executive
r w n ( k + 1 ) , t The tth internal reward prediction in executive
c k + 1 t The tth virtual actions at time n for prospective (k+1)th PAC
c k + 1 The final action for applying in (k+1)th PAC
Table 3. Numerical simulation parameters of the OFDM fiber optic communications system and CDS.
Table 3. Numerical simulation parameters of the OFDM fiber optic communications system and CDS.
Simulation ParametersValue
Fiber dispersion coefficient ( β 2 ) 22.1   ps 2 / km
Nonlinear coefficient ( γ ) 1.1   W 1 km 1
Fiber loss coefficient ( α ) 0.2   dB / km
Number of fiber spans (N) 20
Span length (Lspan)80 km (1600 km total length)
Noise figure ( N F ) 4.77   dB
Number of OFDM subcarriers per frame256
Subcarrier modulationQAM-16
Number of data frames128,000
Number of frames for modeling100,000
Number of frames for validation28,000
Discretization step d4 Gb/s
P F i = 0 k 2
P F i = 1 k 5
Complexity threshold 10 6 discretized cells
max ( F m = 0 t o t a l , k ) ,   k = 1 , 2 , 3 , 4 3600 discretized cells
max ( F m = 1 t o t a l , k ) ,   k = 5 , 6 , 7 , 8 , 9 518,400 discretized cells
F m = 2 t o t a l , k ,    k = 10 74,649,600 discretized cells
T h r e s h o l d m = 0   or   FEC    threshold 1.03 × 10 2
T h r e s h o l d m = 1   or   FEC    threshold 4.7 × 10 3
Fiber typeStandard single mode fiber (SMF)
Table 4. The CDS parameters at steady state.
Table 4. The CDS parameters at steady state.
Results or CDS ParametersValue
Data rate (k = 9)280 Gb/s
P F i = 0 k = 1 , 2 , , 9 2
P F i = 1 k = 5 , 6 , , 9 5
BER (k = 9) 4.4 × 10 3
Q-factor improvement (k = 9)6.98 dB

Share and Cite

MDPI and ACS Style

Naghshvarianjahromi, M.; Kumar, S.; Deen, M.J. Natural Brain-Inspired Intelligence for Non-Gaussian and Nonlinear Environments with Finite Memory. Appl. Sci. 2020, 10, 1150. https://doi.org/10.3390/app10031150

AMA Style

Naghshvarianjahromi M, Kumar S, Deen MJ. Natural Brain-Inspired Intelligence for Non-Gaussian and Nonlinear Environments with Finite Memory. Applied Sciences. 2020; 10(3):1150. https://doi.org/10.3390/app10031150

Chicago/Turabian Style

Naghshvarianjahromi, Mahdi, Shiva Kumar, and M. Jamal Deen. 2020. "Natural Brain-Inspired Intelligence for Non-Gaussian and Nonlinear Environments with Finite Memory" Applied Sciences 10, no. 3: 1150. https://doi.org/10.3390/app10031150

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop