Next Article in Journal
Prediction of Heat Transfer and Fluid Flow Effects on Entropy Generation in a Monolithic Catalytic Converter Using Large-Eddy Simulation
Previous Article in Journal
A Skew Logistic Distribution for Modelling COVID-19 Waves and Its Evaluation Using the Empirical Survival Jensen–Shannon Divergence
Previous Article in Special Issue
Naturalising Agent Causation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Metacognition as a Consequence of Competing Evolutionary Time Scales

1
Department of Biology, Allen Discovery Center at Tufts University, Medford, MA 02155, USA
2
23 Rue des Lavandières, 11160 Caunes Minervois, France
3
Wyss Institute for Biologically Inspired Engineering, Harvard University, Boston, MA 02138, USA
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(5), 601; https://doi.org/10.3390/e24050601
Submission received: 10 February 2022 / Revised: 15 April 2022 / Accepted: 19 April 2022 / Published: 26 April 2022
(This article belongs to the Special Issue Towards a Quantitative Understanding of Agency)

Abstract

:
Evolution is full of coevolving systems characterized by complex spatio-temporal interactions that lead to intertwined processes of adaptation. Yet, how adaptation across multiple levels of temporal scales and biological complexity is achieved remains unclear. Here, we formalize how evolutionary multi-scale processing underlying adaptation constitutes a form of metacognition flowing from definitions of metaprocessing in machine learning. We show (1) how the evolution of metacognitive systems can be expected when fitness landscapes vary on multiple time scales, and (2) how multiple time scales emerge during coevolutionary processes of sufficiently complex interactions. After defining a metaprocessor as a regulator with local memory, we prove that metacognition is more energetically efficient than purely object-level cognition when selection operates at multiple timescales in evolution. Furthermore, we show that existing modeling approaches to coadaptation and coevolution—here active inference networks, predator–prey interactions, coupled genetic algorithms, and generative adversarial networks—lead to multiple emergent timescales underlying forms of metacognition. Lastly, we show how coarse-grained structures emerge naturally in any resource-limited system, providing sufficient evidence for metacognitive systems to be a prevalent and vital component of (co-)evolution. Therefore, multi-scale processing is a necessary requirement for many evolutionary scenarios, leading to de facto metacognitive evolutionary outcomes.

1. Introduction

The idea of metacognition—“thinking about thinking”—is as old as introspective philosophy and is central to both cognitive neuroscience [1,2,3,4,5,6] and artificial intelligence [7,8,9,10,11]. Metacognition is sometimes defined to include exclusively human capabilities, e.g., theoretical self-knowledge [12]. In its simplest form, however, any regulator that is an integral part of some larger system can be viewed as a “metacognitive” model of the lower-level system components that it regulates [13]; systems that include such internal regulators display self-monitoring and self-regulation, the most basic attributes of metacognition noted in [12]. In humans, metacognition is broadly associated with executive control [14,15] and “deliberate” Process-1 problem solving [16,17,18], though whether there is a clean, architectural difference between this and “automated” Process-2 problem solving remains subject to considerable debate [19,20,21]. It is clear, in particular, that metacognitive processes are not always consciously introspectable. How such “high-level” metacognition relates to more prosaic forms of regulation, either in humans or across phylogeny, remains poorly understood and indeed relatively uninvestigated.
Here, we adopt a broad notion of metacognition, employing this term to indicate the function of regulating some “lower-level” computational process. An architectural component that implements metacognition will be referred to as a metaprocessor for the component being regulated. This purely functional definition of metacognition explicitly avoids the seemingly intractable philosophical problem of determining whether a given system is “cognitive”; our approach to this latter problem is detailed elsewhere [22]. It also explicitly depends on how the larger system of which the metaprocessor is a component is identified. In particular, this larger system must be identified as comprising both the metaprocessor and the system that it regulates. A thermostat, for example, can be considered a metaprocessor for the on/off switch of a furnace, provided that the thermostat, the switch, and the rest of furnace are considered to compose a single system, a “regulated furnace”; if the thermostat is considered independently of this larger system, it is simply a thermally regulated switch. The question of whether something is a metaprocessor is, therefore, not an ontological question, but rather an architectural question that depends on how the larger system that contains it is functionally decomposed.
With this understanding of metacognition as a function and metaprocessors as components, we address the question of why natural selection has produced systems capable of metacognition, i.e., systems regulated by metaprocessors. We suggest that metacognition can be expected whenever (evolutionary) fitness functions—as we will discuss in the context of active inference in Section 2.1 and in genetic algorithms in Section 3.2.3—vary on multiple timescales. As we will see, such fitness functions are crucially multi-scaled. Metacognition, in this view, provides a context-dependent switch between depth-first and breadth-first searches that allows the avoidance of local minima. As such, it can be expected to be both ancient and ubiquitous. We can, indeed, identify slowly acting regulators of faster control systems even in prokaryotes; e.g., acetylation that changes the sensitivity of the chemotaxis “switch” molecule CheY in E. coli [23]. This example reminds us that biological systems implement memories at multiple scales using various molecular, cellular, and bioelectric substrates [24].
In what follows, we first briefly review comparative studies that take human metacognition as a starting point in Section 2.1. We outline the computational resources required to implement metacognition in arbitrary systems, focusing on requirements for memory and an ability to measure duration, in Section 2.2. We review the conditions under which interactions between physical systems are mediated by Markov blankets (MBs) [25,26] in Section 2.3, and then provide a general overview of the framework of active inference to minimize variational free energy (VFE) as it applies to living systems from the cellular scale upwards [27,28,29,30,31] in Section 2.4. As this framework is provably general for systems that maintain states conditionally independent from those of their environments (or in quantum-theoretic terms: separable states) over macroscopic times [32,33], it provides a general formalism for discussing the evolution of computational architectures enabling metaprocessing. We turn in Section 3 to our main hypothesis: that metaprocessing, and hence metacognition as a function, can be expected to arise in any systems faced with selective pressures—effectively, sources of VFE—that vary on multiple timescales. We begin in Section 3.1 by proving, under generic assumptions, that metacognition is more energetically efficient, and hence confers higher fitness, than purely object-level cognition when selection operates at multiple timescales. We then turn in Section 3.2 to illustrations of this result, reviewing simulation studies of a variety of architectures, including multi-agent active inference networks, Lotka–Volterra-based systems of predator–prey interactions, coupled genetic algorithms, and coupled generative adversarial networks (GANs), showing how multiple timescales intrinsic to the relevant problem domains induce metacognition as a high-level function. We then discuss in Section 3.3 the emergence of coarse-graining representations of both time and space as a general consequence of metacognition. We conclude that metacognition is far from being human- or even mammal-specific and is not a special case; rather it can be expected at every level of biological organization and always addresses fundamentally the same problem, that of dealing effectively with uncertainties having different spatio-temporal scales.

2. Background

Before we investigate metacognition as an intrinsic function in generic terms, as well as in a variety of well-studied model paradigms, we first consider metacognition from an evolutionary perspective, outline the key computational resources necessary for metaprocessing, and briefly review the active inference framework as a provably general model of systems capable of metacognition.

2.1. Metacognition from an Evolutionary Perspective

The idea of metacognition as a function has two distinct theoretical antecedents: the philosophy and, later, psychology of deliberate, conscious control in humans, and the practical, engineering development of regulators and control systems. The definition of “conscious control” is of course problematic, as there are well-known and substantive disagreements about what consciousness is and how it is manifested (see for example [34,35] for recent reviews). Global workspace [36,37] and higher-order thought [38,39] approaches to consciousness naturally involve metacognition; others, for example Integrated Information Theory [40], do not. Even the notion of content of consciousness is controversial [41]. It is, moreover, as noted earlier, not clear what distinguishes “deliberate” from “automatic” processes [19,20,21]. These definitional controversies motivate our current purely functional approach to metacognition. This approach is broadly consonant with the systems engineering tradition, particularly as elaborated within artificial intelligence and robotics, where canonical metacognitive functions such as intrinsic motivation [42], curiosity-driven allocation of learning resources [43], and ethics-driven decision making [44] have become increasingly prominent components of autonomous or semi-autonomous systems.
Broadening the definition of metacognition away from human-specific capabilities and avoiding debates about consciousness allows explicit consideration of metacognition as an evolutionary development with potentially deep phylogenetic roots (but also see [22] for an approach to consciousness fully consistent with this evolutionary perspective). As expected, given the self-monitoring and self-regulating components of metacognition, canonical metacognitive tasks such as introspection, deliberate choice, and voluntary recall engage components of both executive [15] and default-mode [45] networks [46,47,48,49,50]. Such tasks can be translated to nonhumans by removing requirements for verbal communication and employing species-appropriate measures of choice, risk, confidence, and self-representation [51]. The involvement of homologous cortical structures in metacognition tasks in humans and other mammals provides additional support for human-like metacognition in other primates [52,53,54,55,56] and rodents [57]; claims of metacognitive ability based on behavioral tests and observation in the wild remain controversial [51,58,59]. Whether metacognitive abilities have been developed, presumably by convergent evolution, in invertebrates, e.g., cephalopods [60,61], also remains controversial.
Here, we offer an alternative approach to metacognition that focuses not on behavioral tests of function, but rather on the presence of architectural structures supporting metaprocessing. This architecture-driven approach avoids a priori assumptions derived from human uses of metacognition; hence, it leaves open the possibility of metacognitive functions in phylogenetically more diverse organisms that might not be noticed or recognized using human cognition as an implicit definition. This leads us to ask, from an abstract, architectural perspective, what computational resources are required to support metaprocessing.

2.2. Computational Resources for Metaprocessing

A generic metaprocessing architecture is shown in Figure 1: the metaprocessor samples both the input to and internal representations within an “object-level” processor and provides outputs that either regulate the object-level process or modify its output. Note that in this representation, meta- and object-level processors are regarded as components of a single containing system that receives inputs and produces outputs. Communication between object- and meta-level processors occurs via a defined channel that can, from a design perspective, be considered an application programming interface (API) and is restricted to the data structures supported by this channel. The Good Regulator Theorem requires that the metaprocessor be (i.e., encode or implement) a “model” of the object processor [13]. The regulatory capability of a metaprocessor is subject to the usual tradeoff between the accuracy, complexity, and computational efficiency of its model; overparameterized models or models that require too much time to execute are clearly nonoptimal. Barring these, regulatory capability increases as access to the internal state of the object processor increases, with the coding capacity of the communication channel as a hard limit. Increased access clearly requires increased computational and memory resources for the metaprocessor. From a design perspective, there is a tradeoff between the algorithmic and computational resources allocated to object- and meta-level processing. Even in a technological setting, the object level can become “frozen” for reasons outside a designer’s control, forcing the metaprocessor to take on tasks beyond simply monitoring and regulating the performance of the object processor. As evolution tends to proceed by “frozen accidents,” we can expect metaprocessors taking over what would, from a design perspective, be regarded as object-level tasks, to be ubiquitous in the architectures of living systems. Epigenetic regulation can, for example, be viewed as a meta-level regulator of gene expression that “takes over” the role ordinarily played by transcription factors when the latter cannot evolve either fast enough or with sufficient specificity to meet a selective challenge. In what follows, however, we will focus on the more traditional role of metaprocessors as regulators, not partial replacers, of object-level functions.
The architecture shown in Figure 1 is clearly hierarchical. The metaprocessor “knows” about the object-level processor only the information encoded in its inputs; the object-level processor, similarly, “knows” about the metaprocessor only the information encoded in the metaprocessor’s outputs. Any such architecture is amenable to virtualization: either processor can be replaced with any other processor having the same I/O behavior. The architecture in Figure 1 can, therefore, be considered a simple (two-level) virtual machine hierarchy [62]; it can be extended into a hierarchy of arbitrary depth. A thermostat, for example, is insensitive to whether the gas furnace it has controlled is swapped out for an electric furnace, or even for a heat pump. Genetic engineering is, similarly, effective to the extent that the expression and function of a gene are only minimally and controllably dependent on the cellular environment into which it is placed.
The local memory available to the metaprocessor determines both the maximum resolution of its model of the object-level process and the maximum window of history data that it can maintain. Hence, local memory is a critical resource for performance-based regulation. In simple systems such as thermostats or the CheY system in E. coli, limited memory can be compensated for by a long characteristic time for regulatory changes; thermostats take advantage of the thermal mass of rooms, while CheY acetylation is much slower than phosphorylation. More sophisticated systems that regulate toward long-term goals, however, require sufficient memory, implementing appropriate data structures, to maintain records of performance history. Metaprocessors that regulate learning systems to maximize learning efficiency provide an example. Any learning algorithm can be considered a function:
L : f , T f
where f   is a function and T is a set of training data. Hence, any learning algorithm can be considered a metaprocessor. Autonomous systems in complex environments must, however, select their own training data by implementing an exploration strategy that focuses on nontrivial but learnable features of the task environment [63]. Such selection systems may be fixed—effectively hardwired—or support learning. Improving training set selection through learning requires a higher-level, longer-time-window metaprocessor that regulates training-set selection by measuring learning progress. Humans accomplish this task heuristically, as expected for systems with limited computational and memory resources [64].
Metaprocessors serving long-term goals typically employ input from multiple sources and coordinate multiple types of actions; autonomous vehicles provide an example [65]. In such systems, the metaprocessor effectively serves multiple object-level processors, regulating their joint behavior toward both near-term (e.g., collision avoidance) and longer-term (e.g., timely arrival at a destination) objectives. As resources become constrained, shared memory and input queuing can be expected to replace true parallelism on both input and output sides. Global workspace (GW) models of human attention allocation and executive control invoke such resource-limited solutions for integrative metaprocessing [36,37]; the LIDA architecture replicates a GW model for robotic control [66].
From a design perspective, metaprocessor architectures are typically explicitly hierarchical and metaprocessors are typically explicitly centralized. Nonhierarchical distributed-system architectures, for example ACT-R [67], lack explicit metaprocessors. Hierarchical recurrent architectures, for example ART [68], similarly exhibit no explicit metaprocessing. We will show below, however, that metaprocessing emerges generically as an effective or apparent function when the interaction between a system and its environment is characterized by fitness functions with multiple characteristic timescales. To show this, we require a suitably generic way to describe both interaction and fitness, to which we now turn.

2.3. Interaction across a Markov Blanket

Biological systems are finite, and their interactions with their environments exchange only finite quantities of energy. The most general representation of such finite interactions is bipartite: some finite system S interacts with a finite environment E that is defined to be everything other than S . This definition renders the joint system S E both finite and closed. We can, therefore, represent the interaction as in Figure 2a: the systems S and E interact via a Hamiltonian (total energy) operator H S E   that is defined at the boundary separating S from E . Formally, the boundary is a decompositional boundary in the state space of the joint system S E ; it separates states of S from states of E . The Hamiltonian H S E   is, formally, a linear operator on this joint state space. The conservation of energy requires that the net energy flow between S and E is asymptotically zero, i.e., that the interaction is asymptotically adiabatic; we will assume for simplicity that it is adiabatic over whatever time scale is of interest.
We now make an explicit assumption that the states of the systems S and E are independently well-defined. This is always the case in classical physics; in quantum theory, it is the assumption that S and E are not entangled (i.e., the joint quantum state S E is separable and factors as S E = S E ) over the time period of interest. In this case, the interaction H S E can be represented, without loss of generality, as a sequence of exchanges of finite information between S and E , i.e., S writes a finite bit string on , which E then reads, and the cycle reverses [22,33,69]—see [70,71] for technical details. The information exchanged between S and E —hence the information encoded on in each interaction cycle—can be specified exactly: at each instant, encodes the current eigenvalue of the operator H S E .
Under these conditions—finite interaction in the absence of quantum entanglement—the decompositional boundary functions as an MB between S and E [72]. An MB surrounding a system S is, by definition, a set of states m = s , a such that internal states i of the blanketed system S (if any) depend on external or environmental states e only via their effects on m [25,26]. In practice, an MB exists whenever the state space of S E includes states not on the boundary . Friston [27,28] introduced the decomposition of the MB into sensory states s and active states a , with the further constraint that sensory states are not influenced by internal states and active states are not influenced by external states; this distinction is illustrated in Figure 2b and, in two biological contexts, in Figure 3. Any MB is, clearly, completely symmetrical; the “sensory” states of E are the “active” states of S and vice versa. An MB generalizes the function of an API, in effect defining, and therefore limiting, the communication channel between any system S and its environment E .
Note that nothing in the above places any constraints on the internal interactions H S   and H E of S and E , respectively. In particular, neither H S nor H E can be inferred from the interaction H S E defined at the boundary that implements the MB separating S from E . The existence of the MB thus places significant restrictions on what S can “know” about E and vice-versa. Because S has no direct access to the state e of E , S cannot measure the dimension of e , and hence cannot determine the number of degrees of freedom of E . The MB similarly prevents any direct access to the internal interaction H E of E . We can, therefore, construct any finite decomposition E = F G with internal interaction H E = H F + H G + H F G without affecting what S can detect, i.e., without affecting the information encoded on the MB by H S E in any way. Hence, S cannot “know” about decompositions of E . Any “subsystems” of E represented by S are, effectively, sectors of the MB that are defined by computational processes implemented by S ; see [22,33,69,73] for formal details and further discussion. As the MB is completely symmetrical, these considerations apply equally to E .
As emphasized by Friston [32], any system that occupies a nonequilibrium steady state (NESS), or more generally, has a well-defined NESS density, must be separated from its environment by an MB. This condition is, however, not necessary; indeed, no system with an MB could perform inferences or enact nontrivial behaviors if it were. Even with a fixed H S E , changes in the eigenvalue encoded on the MB, and hence in the energy transferred across it, will generally correlate with changes in the internal states of both S and E . Time dependence in H S E will also induce time-varying states in both S and E (see [33] for further discussion). The Free-Energy Principle (FEP) [27,28,29,32] is the statement that closed joint systems S E will asymptotically approach a constant H S E and, hence, that both S and E have NESS densities as attractors; it is shown in [33] that this principle is, for quantum systems, asymptotically equivalent to the Principle of Unitarity, i.e., of the conservation of information within the joint system.

2.4. Active Inference Framework

The Bayesian inference framework enables quantitative models linking sensory mechanisms (i.e., inputs) with functional behaviors (i.e., outputs) in arbitrary classical [32] and quantum [33] systems. This framework has been applied extensively to biological systems [28,29,30,31,74,75,76,77,78]; here, we consider it more generally. The variational free energy (VFE) that is being minimized in Bayesian inference follows out of classical analytical and statistical physics considerations as a unique form of a least action principle. With the partition defined by the MB in hand, we can interpret internal states as parametrizing some arbitrary probability density q e (its “beliefs”) over external states. This replaces the Lagrangian being used to compute the gradient descent in classical least action (or variational) principles with a VFE functional F of form (cf. Equation (8.4) in [32]):
F m , i = ln   p m , i + D K L ( q e | | p ( e | m , i ) , Surprisal Bound
where
D K L p | q = p x   ln p x q x   d x
is the Kullback–Leibler Divergence defined as the expectation of the logarithmic difference between the probabilities p and q . This makes Equation (2) a logarithmic difference between the (variational) density or Bayesian “beliefs” about external states q e and actual probability densities p ( e | m , i ) of external states given the MB state m and the internal state i . Note that the VFE is a measure of uncertainty about how the current behavior of the environment E (i.e., the external state e ) will affect the future state m of the MB; the VFE is thus completely distinct from the thermodynamic free energy required to drive computation, as discussed in Section 3.1 below.
The first term in Equation (2) is known as (Bayesian negative log) model evidence, or marginal likelihood, describing the likelihood that the sensory inputs (i.e., the MB state m ) were generated by the system’s generative model, implicit in the internal state i . Because negative evidence is equivalent to unexpected sensory inputs, this term is also referred to as surprisal. The second term is referred to as relative entropy, which minimizes the divergence between the variational and posterior densities q e   and p e | m , i ,   respectively. As a result, maximizing model evidence—by either changing i to encode new beliefs or acting on E so as to alter m —is equivalent to minimizing the VFE F m , i   of the system [28,29,30,32]. Because the divergence of the second term can never be less than zero, free energy is an upper bound on the negative log evidence. For this reason, the Kullback–Leibler term is also referred to as a bound on surprisal.
With this definition of VFE, the FEP can be stated as the principle that any system S with an MB will behave so as to minimize VFE, i.e., so as to maximize its ability to predict the future actions of its environment E on its MB. This is active inference [27,28,29,32]. Maximal predictive ability clearly implies minimal time variation in both H S E and H S , i.e., S occupying or being very close to its asymptotic NESS. The converse is also true:
Theorem 1.
Time variation in H S E increases VFE for S E
Proof. 
Time variation in H S E is time variation in the spectrum of H S E , i.e., in the eigenvalues of H S E encoded on the MB. VFE measures the unpredictability of these eigenvalues; hence, variation in the eigenvalues increases VFE. □
Theorem 1 imposes an additional limit on what S can “know” both about E and about itself. Consider a time evolution U = S E U = S E that “moves” the decompositional boundary within a fixed joint system U . This evolution corresponds to an exchange of degrees of freedom between S and E ; S engulfing part of E or vice versa would be an example. Any such evolution changes the Hamiltonian, H S E H S E . If by theoretical fiat we choose to regard S and S as “the same system”—e.g., we treat an organism as “the same thing” after it ingests a meal—we can conclude via Theorem 1 that the VFE experienced by that system has increased. Its VFE may, of course, thereafter decrease if the new interaction H S E proves more predictable than the previous H S E .
Theorem 1 has a significant consequence for the FEP: no system can reliably predict, from the information available on its MB, whether it has reached its NESS. Indeed, to do so would require reliably predicting that its environment has reached its respective NESS, a feat explicitly ruled out by the MB condition.
For the far-from equilibrium systems that predominate in biology, the active inference formalism transforms an intractable integration problem of a global thermodynamic potential into a tractable integration over a probabilistic model of how a system “thinks” it should behave, i.e., a model of the behavior of the states of its MB. To illustrate this, recall Equation (1) and consider how a system learns about, and hence enhances its ability to predict the behavior of, its environment. At each instant, the state m of the MB can be considered the “next” element of the training set T . The internal state i of S can be considered to implement some function f ; indeed this f is, in the language of the FEP, S ’s generative model at that instant. If S is an artificial neural network, for example, f is implemented by the node transfer function (which can be taken to be identical for all nodes) and the weight values. In the early stages of learning, i.e., far from the NESS, significant variation in m will induce significant variation in f and hence in i . An active system, e.g., an organism or an autonomous robot, will modify m by behaving, i.e., acting on E to provoke some response. As the predictive capability of f improves, i.e., as i approaches in NESS, changes in m will invoke smaller changes in f , with no further changes in f (i.e., no further learning) as the limit in which “perfect prediction” and the NESS are achieved.
This principle of active inference has recently been applied directly to biological self-organization, beyond its initial applications in neuroscience, from lower levels of biological complexity such as a primordial soup [28], to higher levels of cells and tissues during morphogenesis [29,31]. Indeed, it has been shown to map quite well to the evolution of larger-scale, higher-complexity systems. If each of a collection of small-scale systems (e.g., cells or individual organisms) can minimize its own, individual VFE (and hence maximize its predictive success) by closely associating with other small-scale systems, a larger-scale system (e.g., a multi-cellular organism or an organismal community) will emerge; such a larger-scale system will remain stable as a macroscopic, collective entity, and hence have a well-defined MB at its larger scale, but only if its dynamics successfully minimize VFE at its larger scale [79]. In this context, natural selection at each scale can be viewed as Bayesian model selection (also referred to as structure learning, because the structure of the generative model itself is evolved and learned [80]) at that scale [30]. This view of natural selection as Bayesian belief updating can be traced back to Fisher’s fundamental theorem of natural selection [81], which equates the rate of increase in fitness to the genetic variance in fitness [81,82]. Fisher’s theorem itself has been cast in information theoretic terms by S.A. Frank in relating the rate of fitness increase to the amount of information that a population acquires about its environment based on the price equation [83], which itself can readily be shown to be an instance of Bayes’ theorem [81]. For an evolving organism, each instantaneous interaction with the environment serves as a “training example” for the generative model as discussed above [84]. How selection pressure generates new neuronal structures underlying new generative models as its own variational free energy minimizing process has been shown in [75]. The fundamental operations of expanding and adapting generative models in response to selection pressure have been outlined in [85] and form a gradual emergence of complexity, balancing model complexity and accuracy demands from increased environmental dynamic complexity driving evolution.
The limit in which selection (i.e., the modification of the generative model) stops and the NESS is achieved is the limit of “perfect adaptation” to the local environment. This process can be formulated in the traditional language of fitness as follows. As “fitness” can be considered a probability of future existence—i.e., of survival and/or reproduction—it can be expressed as a (instantaneous) function g :   e , m , i     0 , 1 , where, as above, e , m , and i , are (instantaneous) external, blanket, and internal states, respectively (hence m , i is the total organism state). Perfect predictions minimize VFE and maximize survival probability; indeed, the fundamental prediction of any system is its own future existence [32]. Hence, with appropriate normalization, we can write g e , m , i = 1 δ , where δ is the total prediction error of the generative model implemented by the internal state i when exposed to the external state e as represented on the blanket state m (see also [86], where this is made time-dependent using a path-integral formalism). Systems with large values of δ are unlikely to survive/reproduce; those with small δ are more likely to survive/reproduce. It is important, moreover, to note that nothing in the theory guarantees the stability of emergent large-scale structures; indeed, many organisms are only facultatively multi-cellular, and as human history amply demonstrates, multi-organism communities often collapse. It is also important to note that while MBs are often associated intuitively with cell membranes or other biophysically implemented spatial boundaries, the MB itself is a decompositional boundary in an abstract state space that may or may not assign spatial degrees of freedom to every state.
It is not our intention here to develop a general theory of evolution as Bayesian satisficing; for some initial steps in this direction, see [84]. Our interest is in a particular aspect of this process, the development of metaprocessor architectures as a way of obtaining more predictive efficacy at a smaller free-energy cost. We also explicitly set aside the question of how one agent identifies another and tracks the other’s identity through time. Identification and tracking require that some criterial component of the identified system maintains an NESS; detailed models of the identification and tracking process are developed in [33,69].

3. Results

In the following sections, we will introduce a formal description of metacognition in an evolutionary setting and discuss how different time scales in the interaction dynamics of simple, well-known and -studied systems induce metacognition in these systems. In Section 3.1, we will formally show how metacognitive abilities confer a fitness advantage in an evolutionary setting based on information theoretic energy considerations. In Section 3.2, we start by reviewing some different computational examples of such simple interaction networks and discuss how complex system dynamics in each case necessitate some form of metacognition. We will first look at multi-agent active inference networks, which are simultaneously the most general models of interaction, and exhibit the most explicit form of metacognition we discuss here. Then, we move on to the more specific, established computational frameworks of multi-agent/network models, including Lotka–Volterra-based systems of predator–prey interactions, coupled genetic algorithms, and GANs, highlighting and exploring their various roles in modeling coadaptation and coevolution. While we will mostly focus on two agent/network interactions, we note that an arbitrary number n of interactions can, while often technically challenging, be decomposed, without the loss of generality, into n binary interactions, each between one of the n agents and the remaining n 1 agents considered as a single system. This follows directly from the Markov blanket condition; at a deeper level, it follows from the associativity of vector-space decompositions [73]. Finally, in Section 3.3, we will discuss how spatio-temporally coarse-grained structures emerge naturally in any resource-limited system with sufficiently complex interaction dynamics.

3.1. Formal Investigation of Metacognition in Evolution

Our primary claim in this paper is that selection on multiple timescales induces metacognition. We can make this claim precise by employing the general formalism for a finite system S interacting with a finite environment E via a Markov blanket MB developed in Section 2.3. Interesting interactions involve nontrivial information processing between reading the MB (i.e., sensation) and writing on (i.e., action) on the part of both S and E . Landauer’s principle [87,88] imposes a free-energy cost of at least ε = k B T   l n 2 per erased bit (and hence per rewritten bit) on (classical) information processing, where k B is Boltzmann’s constant and T > 0 is the temperature of the system performing the computing. Note that this is a lower limit; a realistic system will have a per-bit processing cost of ε = β k B T , with β > l n 2 a measure of thermodynamic efficiency. As this free energy must be sourced, and the waste heat dissipated, through asymptotically, we assume for simplicity that it is sourced and dissipated through on every cycle. We can, therefore, represent as in Figure 2b. The bits allocated to free-energy acquisition and waste heat dissipation are uninformative to S , as they are burned as fuel for computation. The bits allocated to action by S are similarly uninformative to S , although they are informative to E . The bits allocated to sensation by S are informative to S and serve as the inputs to S ’s computations. This distinction between informative and uninformative bits is often not made explicitly (such as in Figure 3), where the flow of free energy as incoming fuel and outgoing waste is ignored; however, it is required whenever the joint system S E is thermodynamically closed [33]. To emphasize this distinction, we will use s for informative sensory states and a for states encoding “interesting” actions, i.e., actions other than waste heat dissipation.
With this formalism, we can place lower limits on the free energy cost of the computational processes that determine S ’s actions on E , given various assumptions about E ’s actions on S and S ’s computational architecture. As noted above, E ’s actions on S implement natural selection. The characteristic times of E ’s actions are, therefore, the characteristic times of the selective pressures that S faces.
In the simplest case, E acts on S via some stochastic process that varies randomly with a characteristic timescale τ. As S has, in principle, no access to the internal dynamics of E except through , from S ’s perspective, τ is the response time of E to actions by S . We assume for simplicity that S ’s response time is also τ, i.e., that S is capable of computing its next action in time τ from detecting E ’s current action.
We also assume for simplicity that S devotes the same number n of bits on to sensation and action, i.e., S computes a function f :   0 , 1 n     0 , 1 n , and that this function depends, in general, on the values of all n input bits (i.e., S does not ignore informative bits). If computing this function requires m steps, including all relevant classical memory write operations, then S has a free-energy cost of n m ε per sensation–action cycle.
Now consider a situation in which E acts on S via two independent stochastic processes with two different characteristic timescales, τ 1 and τ 2 , and assume that S detects and processes the inputs from these two environmental processes independently. Assume also that S ’s thermodynamic efficiency is the same for detecting and processing the inputs from these two environmental processes. In this case we can factor the state s   into two sectors, s 1 and s 2 , encoding n 1 and n 2 bits, respectively, the states of which vary at τ 1   and τ 2 , respectively. Suppose now that τ 1 = τ and τ 2 > τ 1 , i.e., suppose S experiences independent selective pressures with two different characteristic timescales. In this case, S ’s free-energy cost to process only the fast-varying inputs from s 1   is n 1 m ε per sensation–action cycle, while S ’s free-energy cost to process only the slowly varying inputs from s 2   is τ / τ 2   n 2 m ε per sensation–action cycle. Hence, S ’s total free-energy cost for processing the two inputs separately is:
Ξ = n 1 + τ / τ 2 n 2 m ε + Δ ,
where Δ is the overhead cost, which must also be sourced as free energy from E , of detecting the change in s 2 and devoting resources to processing it. Whether Ξ < n m ε , i.e., whether processing the inputs separately is more energy efficient than processing all n input bits together on every cycle, clearly depends on the value of Δ. Hence, we can state:
Theorem 2.
If informative inputs vary at different timescales τ 1 and τ 2 , τ 2 > τ 1 , processing the inputs separately is energetically advantageous provided its cost Δ < 1 τ / τ 2 n 2 m ε , where n 2 is the number of bits of the slowly varying input, m is the number of computational steps for each input bit, and ε is the per-bit thermodynamic cost.
Proof. 
The cost of processing the inputs together on each cycle is n m ε ; hence, 1 τ / τ 2 n 2 m ε is the energy saved by processing the inputs separately if the Δ = 0 . Provided Δ remains below this number, separate processing is advantageous. □
Separate processing of inputs is, clearly, a simple form of metaprocessing; the metaprocessor is the “switch” that detects a change in s 2 and devotes processing resources to it; hence, Δ is the overhead cost of metaprocessing. For example, the sector s 1 may encode inputs from an “object” that S identifies in E , while s 2 encodes the “background” surrounding the object, i.e., the entire rest of E . Distinguishing an object from its environmental background requires distinct rates of state change; typically, objects change their states faster than the background does, e.g., by moving [89]. We can, therefore, consider segmenting objects from their environmental backgrounds a simple and evolutionarily ancient metacognitive function.
Other things being equal, higher energetic efficiency confers a fitness advantage. Hence, we can state:
Corollary 1.
Organisms implementing metacognition will generically dominate organisms that do not.
Proof. 
Other things being equal, the fitness advantage of any energy cost savings due to metaprocessing is exponentially amplified. Hence, in N generations, the more efficient organisms will have a fitness advantage of 1 + n m ε Ξ N . □
Selective pressures with multiple timescales, therefore, generically favor metacognition. The fact that even simple organisms such as E. coli employ metaprocessing architectures is, in this case, not surprising. Organisms favor metaprocessing architectures for the same reason software engineers do: because they are more efficient.
To highlight the biological relevance of optimizing bit-wise information processing energy costs, it is noteworthy that individual cells and single-cell organisms can, and must, in order to survive, detect and compute very small differences in information [90]. For example, in bacterial chemotaxis, it has been shown that fast-moving E. coli can respond to concentration changes as low as 3.2 nM [91] in as little as one second, the time available to evaluate ligand concentrations between “tumbles” [92]. As this concentration value corresponds to only about three ligand molecules in the volume of the cell, assumed to be one femtoliter, this actually suggest single-molecule detection [90]. As Aquino and Endres have shown, biochemical signaling precision against a background of large thermal noise and in far-from-equilibrium systems is enhanced by active cellular processes such as receptor internalization, allowing quick removal and subsequent internalization-rate-dependent time averaging of sensed ligands [90]. As such processes of endocytosis almost always require some form of energy consumption such as through ATP [93], we should expect single cells and single-cell organisms to evolve metaprocessing architectures based on the considerations above. Indeed, for this exact reason, Landauer’s principle has recently been applied to, and confirmed in, cellular biochemical networks for steady-state computations of ligand concentrations [94,95].
In the next section, we examine several well-studied architectures, showing in each case how they use metaprocessing to improve efficiency.

3.2. General Models of Two-System Interaction with Selection across Different Time Scales

3.2.1. Multi-Agent Active Inference Networks

When coupling active-inference systems together to study multi-agent behavior, the first thing to notice is that actions from one agent are natively coupled into the sensory inputs of the others and vice versa without need for further extension, as opposed to the other modeling frameworks below. This is because of the way the MB partitions the systems into active, sensory, internal, and external states from the start: active states of one subsystem are directly coupled with the sensory states of others, while the external (also called hidden) states are made up of the internal states of others, as internal states between agents are hidden from each other by the intervening MB. What is left to state explicitly is exactly how the updating of beliefs and subsequent actions of one agent affect those of others in the same context of coadaptation and coevolution as above.
Variational free energy, the term to be minimized at the center of active inference, places an upper bound on (logarithmic) model evidence (i.e., the likelihood of sensory data with respect to an internal model), and thereby on the dispersion of a particle’s internal states and their Markov blanket. Because in a multi-agent system the reduction of the dispersion of sensory, and therefore internal states, necessitates the minimization of the dispersion of active states of other agents, we can equate this goal as each agent working to limit the behavioral flexibility of the others. When different agents have the same type of generative model, the system will quickly, after the random initialization of prior beliefs, converge on a stable, quasi-static configuration. This was the case in our previous work on morphogenesis as Bayesian inference, where the generative model encoded a target morphology as the goal structure [29,31]. In the coevolutionary systems we are discussing in this paper, however, this is no longer the case. In such a system with different generative models, the task becomes for each agent to learn the generative model of its environment. Fortunately, we do not need to reinvent the mechanism of learning in a purely active inference framework, but can instead integrate deep neuronal networks into an active inference model, which is what has recently been achieved under the deep active inference framework [86]. This framework deals specifically with the problem of an agent having to contend with environmental dynamics to which it not only does not have direct access to, but also whose underlying functional form does not have to coincide with the functional form of the agent’s generative model of the world. To solve this problem, deep active inference employs deep and recurrent neural networks [96,97] as flexible functional approximators for the internal dynamics of the agent, allowing the agent’s generative model to learn a good approximation of true world dynamics [86]. On top of that, this framework employs evolution strategies, an extension to genetic algorithms that focus on unbiasedness and adaptive control of parameters [98]. These evolution strategies estimate gradients on the variational free energy bound on the average agent’s surprise over a population of agents [86], surprise here meaning how unexpected sensory inputs vary compared to an agent’s prior beliefs, which is what is being minimized in active inference.
A concrete application of deep active inference to adaptive behavior in a human setting has recently been implemented in the form of studying emotional valence [99]. This paper shows that the type of deep hierarchical structure employed in deep active inference can allow an agent to hold and utilize information in working memory as a basis for future valenced actions. This is relevant to this discussion for two reasons:
(1) We have recently shown that the concept of valence goes far beyond neuronal systems, is at the core of the fundamental drive to self-preservation underlying organisms as living agents, and is very compatible with the basic computational approaches to learning [100,101].
(2) By maintaining internal valence representations in a deep active inference model, the ensuing affective agent is capable of optimizing confidence in action selection preemptively. These valence representations can then in turn be optimized by leveraging the (Bayes-optimal) updating term for subjective fitness, tracking changes in fitness estimates and lending a sign to otherwise unsigned divergences between predictions and outcomes [99]. This formulation of affective inference then corresponds to implicit metacognition [99] and is exactly the necessary capability that will emerge in our considerations of coevolution below.

3.2.2. Predator–Prey Models

Complex temporal ecological interaction has been identified as a driving force behind evolution. Despite the extensive evolutionary records supporting this notion, the actual prediction of coevolutionary outcomes is extremely difficult because of a lack of access to the experimental records of the actual dynamic interactions of different species in concert with their environment. One of the simplest ways to computationally model such a coevolutionary interaction is by building on predator–prey models, which predict the population dynamics of two different species in a predator–prey relationship over time. Predator–prey models are typically based on Lotka–Volterra equations of the form:
d x d t = α x β x y d y d t = δ x y γ y ,
initially proposed in the context of chemical reactions [102], where x and y are the number of prey and predator organisms, and d x d t and d y d t the instantaneous growth rates of the two populations, respectively, with α, β, γ, δ as positive real parameters describing the individual traits determining the interaction between the two species.
This system of linear partial differential equations results, in nontrivial parameter settings, in oscillating dynamics where the growth peaks of the predator population cause a collapse of the prey population, which in turn results in a collapse of the predator population, followed by a rebound of the prey population and so on (cf. Figure 4). This simplified cyclic dependence, however, is built on the assumption of species life history traits being fixed over ecologically relevant time scales [103]. Moreover, once this population-level framework is extended to (co-)evolution, implicit inbuilt timescales of the system such as life cycle and environmental shifts quickly necessitate the explicit formulation and analysis of these time scales for the system to be solvable [104,105].
A first step towards introducing the ability of either population to adapt as part of an evolutionary process was provided by the Rosenzweig–MacArthur extension [107], which includes density-dependent population growth and a functional response. Because this extension allows for self-limitation in prey and handling limitation in the predator populations, prey adaptation over evolutionary time scales can then be included in stabilizing the system in more complex cases such as more than one coexisting predator species [108].
Coevolution can then be incorporated following [103] by allowing the predator and prey traits determined through the parameters α and β (and with them implicitly δ and γ ) from Equation (5) to vary over time:
d α d t = V a α α i 1 x   d x d t α i = α   d β d t = V b β β i 1 y   d y d t β i = β   .
Here, V a α and V b β denote the genetic trait variances of predator and prey, while the remaining terms in these trait equations determine the individual fitness gradients for the traits, whose derivatives are taken with respect to an individual’s phenotype ( α i and β i ).
Much work has been carried out on finding stabilities, as well as critical transitions in the population dynamics that are the basis of evolutionary shifts, using, among others, bifurcation analysis to show that different time scales in prey and predator populations reduce the model dimensionality enough to allow for different behavioral strategies to influence system stability and persistence necessary for evolution to act upon [104,105,109,110].
The temporal patterns emerging from such treatments of the predator–prey model crucially help explain experimental evidence in the ecology of temporal (as well as spatial) separation in terms of behaviors [111,112] and life cycles [113,114], in which different species have seemingly learned the temporal patterns of their cohabitants. Indeed, when analyzing predator–prey models in the context of different ecological and evolutionary timescales, alternating, periodic patterns of different prey and predator pairs have been found, and it indeed has been shown that such evolutionary timescales are essential for the maintenance of different species in a system [115].
It is in this experimental context that computational models of reinforcement and deep reinforcement learning have been applied to the within-lifetime behavior, as well as the evolution of new behavior and physical properties on longer time scales [106,116,117,118,119]. Specifically, it has been shown that the reinforcement learning of predators is beneficial to the stability of the ecosystem, while the learning of prey makes the ecosystem oscillate and leads to a higher risk of extinction for predators [118]. Crucially, the same research shows that the co-reinforcement learning of predators and prey leads to predators finding a more suitable way to survive with their prey, meaning the number of predators is more stable and greater than when only predators or only prey learn [118], further supporting the notion that each species perceives and adapts to the temporal patterns of other species.
In other words, by highlighting the clear adaptive and learning components in coevolving interactions and modeling approaches thereof, we identify components of a metaprocessing architecture that provide us with a higher-level, predictive view of coevolution. To further elaborate on the idea of coevolving species, we next review work on coupled genetic algorithms, which allows us to better study temporal adaptation over long time periods.

3.2.3. Coupled Genetic Algorithms

While Darwin is credited with originating the core concept of evolution through natural selection, it was not until the 20th century that scientists discovered the various molecular processes that created the genetic diversity upon which natural selection could occur [120]. Once the fundamental processes of mutation and crossover were sufficiently illuminated, evolution could be reduced to a surprisingly simple yet efficient algorithm, which Holland first formulated as a tool not just to simulate evolution, but also to solve other computational problems by operating as a general-purpose classifier system [121]. The basic concept of a standard evolutionary algorithm is to represent a population of individual organisms as part of a species to be evolved (or solution to be found) as sequences, or strings, of individual units called bits (which can represent genes or even individual nucleotides when modeling actual evolution) that are in initially randomized states of a predefined set (cf. Figure 5). Furthermore, one must define a scorable attribute to the string called the fitness function that can be computed at each time step. In the case of evolution, this fitness function describes how fit, i.e., how probable to appear in the next generation, any given phenotype is given the organism’s environment. Then, at each simulation step, a certain amount of mutation (random changing of states of a limited number of bits), as well as crossover (switching of states between different regions or different strings) is performed, and the fitness function for the different outcomes is calculated. Outcomes that decrease fitness are eliminated, while outcomes that increase it will be passed to the next iteration, corresponding to the next generation in an evolutionary setting. Thereby, outcomes that increase fitness appear more frequent in subsequent iterations, allowing the solution space to be sampled very quickly yet broadly [122,123].
The two biggest limitations of genetic algorithms are that (1) they can readily arrest in local minima of the solution space and (2) they are very sensitive in their efficiency to parameters governing the mutation and crossover probabilities, as well as the definition of the fitness function itself, leading to a lot of problem-specific finetuning of the algorithm [122,124,125]. An important step towards remedying this problem and creating more efficient and robust genetic algorithms has been taken in the form of introducing adaptation to the genetic algorithm, in which either the probabilities of mutation and crossover are made dependent on the fitness values of previous generations [126], or the fitness function itself is dynamically varied according to environmental constraints [127]. This straightforward computational finding has wide-ranging implications for biological evolution, specifically explaining the source of the rapid adaptation of species in novel environments, such as is observed in invasive species [126].
More importantly for this work, it is this inclusion of an adaptive property to a genetic algorithm that allows us to study coevolution and the emergence of the temporal patterns in which we are interested. The most fundamental aspect of coevolution is that the fitness functions of each species will invariably depend on the evolution of other organisms, creating the need to include time-varying and tracking functions into the genetic algorithm as above, otherwise the convergence cannot be achieved on realistic time scales [128,129].
We have already implicitly shown how a need for coevolution emerges in the interaction between different species through the example of predator–prey modeling in the previous section, and now we turn to how this is achieved using coupled genetic algorithms. Coupled genetic algorithms are referred to as such because they couple the fitness function of different coevolving individuals and systems [130,131,132,133], thereby each computing the selection function for the other, exactly as shown in predator–prey interactions [134,135]. We note that if behavior during the lifetime of individuals significantly affects fitness and therefore evolution, this coupling can replace the “all at once” sequential fitness calculation of a standard genetic algorithm with a more partial, but continuous, fitness evaluation, which occurs during the entire lifetime of an individual, called lifetime fitness evaluation [136].
Interestingly, it has been shown that there are two ways to maintain stability and successful long-term adaptation in a coevolutionary algorithm of a predator–prey type system [134]: First, one can include a sufficiently long memory in the subspecies of how fit and therefore well-adapted it has been compared to the other species in previous lifetimes. Second, one can balance the coevolution by making the reproduction rate dependent on the performance of a population, leading to different reproduction rates in both populations, just as we have seen in the previous section [113,114]. Both solutions implicitly cause each subspecies to learn the temporal patterns of the coevolving species.
Furthermore, research on coupled fitness landscapes in coevolving systems suggests that evolutionary dynamics through natural selection would favor individuals that achieve higher coevolutionary adaptive success, if the sustained average fitness of coevolving partners depends upon the ruggedness and coupling of their landscapes [137]. The emergence of characteristic time scales in this framework, consistent with our theoretical and experimental observations above, is then described as a necessary condition (acting via selective pressure) towards achieving evolutionary adaptability. This is achieved by forcing a system to a “liquid” phase state “poised at the edge of chaos”, where cascades (or avalanches) of evolutionary change propagate across all spatial scales in a power law distribution of characteristic cascade size and frequency [137].
In order to make this implicit learning of other coevolving strategies explicit, we now turn to a third type of model, called coupled GANs, which will allow us to directly express and study agents that learn each other’s evolutionary strategies.

3.2.4. Coupled Generative Adversarial Networks

A basic generative adversarial network consists of a generator and a discriminator.
The objective of the generative model is to synthesize data, while the objective of the discriminative model is to distinguish real data from fake data. The key novelty in this approach as compared to many other machine learning approaches, both supervised and unsupervised, is that the learning of the generator is indirect through the discriminator, as it is not actually trying to maximize how close its generated outputs are to real outputs, but instead aims to counteract the ability of the discriminator to tell the difference between fake and real inputs. Crucially, because the discriminator itself is also updated dynamically, this becomes an adaptive, competitive process that necessitates a lot fewer restrictions on the type of generative model we begin with. However, it still allows the generator to focus its generation on a smaller manifold close to the real data, which is why they can generate more precise outputs than other models [138]. One way to apply a generative adversarial network in a dynamic adaptive setting similar to the simulations of coadapting and coevolving species above has recently been provided by Talas et al., where the generative network corresponded to prey developing better camouflage, while the predator was represented by the discriminating network trying to discern the prey from its environment [139]. As would be expected in a real, yet constrained to one physical dimension, predator–prey evolutionary arms race, the simulation ends in either an equilibrium or the “extinction” of one or both components, corresponding to the discriminator finding 100% of the generated targets on every iteration.
In order to apply this method to more complex, i.e., multi-agent (more species modeled by more generative and discriminative networks) and/or multi-directional (one species being modeled by both a generative and a discriminating network, making adaptation truly bidirectional), one only needs to couple multiple GANs together (cf. Figure 6). Indeed, coupled GANs are at the forefront of novel research on finding computational solutions to problems of learning joint distributions, where tuples (pairs) of corresponding data sets are not provided, resulting in a more biologically realistic case of unsupervised domain learning [140,141]. Indeed, in such frameworks, the learning of spatio-temporal separation and necessary memory conditions that we have seen earlier can be modeled explicitly by employing Spatio-Temporally Coupled Generative Adversarial Networks (STC-GANs). These networks learn powerful temporal representations by extracting rich dynamic features and high-level spatial contexts via a spatio-temporal encoder in a future frame generation model [142]. Here, STC-GANs are used for predictive scene parsing, which captures contextual appearance information and dynamic temporal representation from prior frames to generate future scene parsing results. In other words, this transforms the implicit learning of temporal patterns in coevolutionary contexts into an explicit problem of predicting future adaptive patterns from coevolving species in an environment.
These examples all illustrate the core result of Section 3.1: that metacognition emerges when selection acts on multiple timescales. We now consider the emergence of coarse-grained structures in metacognition, showing when and how they emerge in natural systems.

3.3. Spatio-Temporally Coarse-Grained Structures Emerge Naturally in Any Resource-Limited System with Sufficiently Complex Interaction Dynamics

We have seen multiple computational reasons to include an explicit representation of time and spatial scales in interacting model systems based on tractability arguments and have seen how such representations emerge in model systems with built-in time scales. We now consider how these representations emerge and function in general.
We start by pointing out that the originally biologically motivated and now heavily utilized computational algorithm we discussed above (Section 3.2.2) is already implicitly scaled (or chunked): cross-over in genetic algorithms. Besides the obvious biological motivation of genes as chunks, it is a lot more computationally efficient to enhance mutation-only searches with cross-over, as it allows the algorithm to search deep in parts of the space defined by fixed subsequences. The main advantage of the inbuilt coarse-graininess which the cross-over provides is that it allows the search algorithm to maintain search diversity by constantly combining novel mutations across the size of the genome, reducing the risk of landing on local minima. In other words, by coarse-graining the different search operators across different time scales, we drastically increase the efficiency of the algorithm in limited computational times by interleaving depth-first (i.e., mutation) and breadth-first (i.e., cross-over) searches. From this perspective, it makes sense that the biological genome would evolve to incorporate multiple scales of organizations, because otherwise adaptive evolution within time constraints relevant to selection pressure would be unlikely.
Another example of scaled spatio-temporal structures is presented by morphogenesis in multi-cellular organisms. Morphogenesis, by definition, concerns the establishment of the shape of an entire organism, meaning that successful morphogenetic processes are, when viewed as a search path in a morphogenetic free-energy landscape, not primarily about individual cell behavior. What this means is that at different stages of a morphogenetic process, individual cells migrate, rearrange, or differentiate not necessarily based on individual cellular decisions, but based on environmental feedback integrated from large numbers of surrounding cells. Therefore, morphogenesis is about morphogenetic fields, signaling concentrations that are chunked into corresponding cell types [143,144]. In this view, the spatial scale introduced by multi-cellularity is an optimal solution to the problem of obtaining a large, functionally complex shape. Moreover, this scaled structure is also represented in the temporal scales of morphogenesis, where fast time scales continually integrate into emerging slower time scales as multi-cellularity increases with morphogenesis.
Finally, we turn back to the question of evolution by highlighting some key properties of multi-level selection theory. We have already seen in our discussions about co-evolution that learning and anticipation are practically required in order for coadapting and coevolving systems to stabilize in sufficiently complex temporal interactions. It seems therefore unlikely that higher levels of multi-evolved systems would emerge in a pure depth-first search without any coarse-graining of spatio-temporal scales. In fact, recent work in formally linking evolutionary processes to principles of learning seems very promising [145] in explaining how modularity evolves in dynamic environments [146], and consequentially how those structures emerge from learned past environments by generalizing to new environments [147]. Indeed, evolutionary and developmental processes can be viewed as executing the same dynamics, which is well-described by the active inference framework, just at different spatial and temporal scales [84,148].
Lastly, while highly intuitive, we highlight the fact that despite the above considerations on the emergence of time- and space-scaled structures in a closed, controlled environment, the actual background of all living things is of course highly structured in time and space by purely physical constraints. This includes the seasonal [149], lunar [150], and day–night cycles [151,152], as well as various weather and geological patterns that build a scaled foundation upon which evolution is built.

4. Conclusions

The central aim of this work was to highlight how the emergence of multi-scale processing is a necessary requirement for many evolutionary scenarios and constitutes a form of metacognition based on an architectural definition of metaprocessing.
We showed that metacognitive systems must evolve naturally when fitness landscapes vary on multiple time scales. We then illustrated this by reviewing coevolutionary modeling frameworks based on predator–prey models and genetic algorithms. Here, quasi-stable coevolution was only possible when implicit or explicit considerations on time scales and methods of learning were included.
We described how new approaches in generative adversarial networks have the potential to link the adaptive learning components of a coevolving system more natively and tightly to a predictive framework in the case of spatio-temporal coupled generative adversarial networks, while keeping a very intuitive form of competitive selection.
We showed that the new emerging application of active inference to biological systems of various complexity comes with an inbuilt formalized structure of active and sensory states that form an interface (the Markov blanket) between the internal (belief-encoding and hence predictive) and external (environmental) states. This interface is, from a design perspective, an application programming interface connecting two virtual machines, with the coding capacity of the application programming interface determining the extent to which the two machines/agents can communicate. Thus, from remarkably simple assumptions, we produce a direct formulation of metacognition in terms of learning the generative model of other agents in a multi-agent active inference framework. Furthermore, incorporating deep learning directly into active inference in recent work [86,99] has the potential to supply a complete model of adaptive, multi-scaled coevolution that would allow one to link complex learning outcomes to remarkably simple, valence-based goal structures that fit well into more established homeostatic considerations on biological drive [100].
Finally, we discussed how time-scaled and coarse-grained structures must emerge naturally in any resource-limited system with sufficiently complex interaction dynamics. This is because, as we proved, under generic assumptions, metacognition is more energetically efficient, and hence confers higher fitness, than purely object-level cognition when selection operates at multiple timescales. We underlined this point by showing that existing computational approaches to coevolution all incorporate, implicitly or explicitly, forms of metacognition without which coadaptation, and hence (co-)evolution, cannot be efficiently achieved. This realization is furthermore strengthened by the observation that the physical environment by which evolution is constrained is already highly structured in time and space.
We thereby conclude that, given these physical constraints on time and resources, evolution could only have converged on solutions that contain metacognitive aspects. This is because the architectural concept of metaprocessing underlying our use of the term metacognition is essential for evolution to have resulted in such diverse and highly structured ecosystems as we observe today.
Crucially, from this follows that we should expect to find various distinct time scales in the structure and behavior of evolved organisms, giving us an important insight into their dynamic interactions with other species, as well providing us with a vital tool with which to manipulate coadaptive outcomes.
The exciting challenge of the coming years will be to elucidate exactly how specific temporal scales manifest in various levels of a coevolving biological system, so that one day we can predict evolutionary outcomes with enough accuracy to act, when necessary, before the collapse of an ecosystem, including our own.

Author Contributions

Conceptualization, C.F., F.K. and M.L.; writing—original draft preparation, C.F. and F.K.; writing—review and editing, C.F., F.K. and M.L.; funding acquisition, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This publication was made possible through the support of Grant 62212 from the John Templeton Foundation. The opinions expressed in this publication are those of the author(s) and do not necessarily reflect the views of the John Templeton Foundation.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Burgess, P.W.; Wu, H. Rostral prefrontal cortex (Brodmann area 10). In Principles of Frontal Lobe Function; Oxford University Press: New York, NY, USA, 2013; pp. 524–544. [Google Scholar]
  2. Flavell, J.H. Metacognition and cognitive monitoring: A new area of cognitive–developmental inquiry. Am. Psychol. 1979, 34, 906. [Google Scholar] [CrossRef]
  3. Fleming, S.M.; Daw, N.D. Self-evaluation of decision-making: A general Bayesian framework for metacognitive computation. Psychol. Rev. 2017, 124, 91. [Google Scholar] [CrossRef]
  4. Koriat, A.; Levy-Sadot, R. Processes underlying metacognitive judgments: Information-based and experience-based monitoring of one’s own knowledge. In Dual-Process Theories in Social Psychology; The Guilford Press: New York, NY, USA, 1999. [Google Scholar]
  5. Shea, N.; Frith, C.D. The Global Workspace Needs Metacognition. Trends Cogn. Sci. 2019, 23, 560–571. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Winkielman, P.; Schooler, J.W. Consciousness, metacognition, and the unconscious. In The Sage Handbook of Social Cognition; SAGE Publications: London, UK, 2012; pp. 54–74. [Google Scholar]
  7. Azevedo, R. Reflections on the field of metacognition: Issues, challenges, and opportunities. Metacognition Learn. 2020, 15, 91–98. [Google Scholar] [CrossRef]
  8. Cox, M.T. Metacognition in computation: A selected research review. Artif. Intell. 2005, 169, 104–141. [Google Scholar] [CrossRef] [Green Version]
  9. Goertzel, B. Artificial general intelligence: Concept, state of the art, and future prospects. J. Artif. Gen. Intell. 2014, 5, 1. [Google Scholar] [CrossRef] [Green Version]
  10. Kotseruba, I.; Tsotsos, J.K. 40 years of cognitive architectures: Core cognitive abilities and practical applications. Artif. Intell. Rev. 2020, 53, 17–94. [Google Scholar] [CrossRef] [Green Version]
  11. Sloman, A. Varieties of Metacognition in Natural and Artificial Systems; AAAI Press: Cambridge, MA, USA, 2011. [Google Scholar]
  12. Drigas, A.; Mitsea, E. 8 Pillars X 8 Layers Model of Metacognition: Educational Strategies, Exercises &Trainings. Int. J. Online Biomed. Eng. IJOE 2021, 17, 115. [Google Scholar] [CrossRef]
  13. Conant, R.C.; Ross Ashby, W. Every good regulator of a system must be a model of that system. Int. J. Syst. Sci. 1970, 1, 89–97. [Google Scholar] [CrossRef]
  14. Friedman, N.P.; Robbins, T.W. The role of prefrontal cortex in cognitive control and executive function. Neuropsychopharmacology 2022, 47, 72–89. [Google Scholar] [CrossRef]
  15. Menon, V.; D’Esposito, M. The role of PFC networks in cognitive control and executive function. Neuropsychopharmacology 2022, 47, 90–103. [Google Scholar] [CrossRef] [PubMed]
  16. Evans, J.S.B.T. Dual-Processing Accounts of Reasoning, Judgment, and Social Cognition. Annu. Rev. Psychol. 2008, 59, 255–278. [Google Scholar] [CrossRef] [Green Version]
  17. Evans, J.S.B.T.; Stanovich, K.E. Dual-Process Theories of Higher Cognition. Perspect. Psychol. Sci. 2013, 8, 223–241. [Google Scholar] [CrossRef] [PubMed]
  18. Kahneman, D. Thinking, Fast and Slow; Farrar, Straus and Giroux: New York, NY, USA, 2011. [Google Scholar]
  19. Chater, N. Mind Is Flat: The Remarkable Shallowness of the Improvising Brain; Yale University Press: London, UK, 2018. [Google Scholar]
  20. Fields, C.; Glazebrook, J.F. Do Process-1 simulations generate the epistemic feelings that drive Process-2 decision making? Cogn. Processing 2020, 21, 533–553. [Google Scholar] [CrossRef] [PubMed]
  21. Melnikoff, D.E.; Bargh, J.A. The Mythical Number Two. Trends Cogn. Sci. 2018, 22, 280–293. [Google Scholar] [CrossRef]
  22. Fields, C.; Glazebrook, J.F.; Levin, M. Minimal physicalism as a scale-free subtrate for cognition and consciousness. Neurosci. Conscious. 2021, 2021, niab013. [Google Scholar] [CrossRef] [PubMed]
  23. Baron, S.; Eisenbach, M. CheY acetylation is required for ordinary adaptation time inEscherichia colichemotaxis. FEBS Lett. 2017, 591, 1958–1965. [Google Scholar] [CrossRef] [Green Version]
  24. Fields, C.; Levin, M. Multiscale memory and bioelectric error correction in the cytoplasm-cytoskeleton-membrane system. Wiley Interdiscip. Rev. Syst. Biol. Med. 2018, 10, e1410. [Google Scholar] [CrossRef]
  25. Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference; Morgan Kaufmann: San Francisco, CA, USA, 1988. [Google Scholar]
  26. Clark, A. How to knit your own Markov blanket. In Philosophy and Predictive Processing; The Free Press: New York, NY, USA, 2017. [Google Scholar]
  27. Friston, K. The free-energy principle: A unified brain theory? Nat. Rev. Neurosci. 2010, 11, 127–138. [Google Scholar] [CrossRef] [PubMed]
  28. Friston, K. Life as we know it. J. R. Soc. Interface 2013, 10, 20130475. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Friston, K.; Levin, M.; Sengupta, B.; Pezzulo, G. Knowing one’s place: A free-energy approach to pattern regulation. J. R. Soc. Interface 2015, 12, 20141383. [Google Scholar] [CrossRef] [Green Version]
  30. Kirchhoff, M.; Parr, T.; Palacios, E.; Friston, K.; Kiverstein, J. The Markov blankets of life: Autonomy, active inference and the free energy principle. J. R. Soc. Interface 2018, 15, 20170792. [Google Scholar] [CrossRef]
  31. Kuchling, F.; Friston, K.; Georgiev, G.; Levin, M. Morphogenesis as Bayesian inference: A variational approach to pattern formation and control in complex biological systems. Phys. Life Rev. 2019, 33, 88–108. [Google Scholar] [CrossRef]
  32. Friston, K. A free energy principle for a particular physics. arXiv 2019, arXiv:1906.10184. [Google Scholar]
  33. Fields, C.; Friston, K.; Glazebrook, J.F.; Levin, M. A free energy principle for generic quantum systems. arXiv 2021, arXiv:2112.15242. [Google Scholar]
  34. Sattin, D.; Magnani, F.G.; Bartesaghi, L.; Caputo, M.; Fittipaldo, A.V.; Cacciatore, M.; Picozzi, M.; Leonardi, M. Theoretical Models of Consciousness: A Scoping Review. Brain Sci. 2021, 11, 535. [Google Scholar] [CrossRef] [PubMed]
  35. Signorelli, C.M.; Szczotka, J.; Prentner, R. Explanatory profiles of models of consciousness-towards a systematic classification. Neurosci. Conscious. 2021, 2021, niab021. [Google Scholar] [CrossRef]
  36. Baars, B.J.; Franklin, S. How conscious experience and working memory interact. Trends Cogn. Sci. 2003, 7, 166–172. [Google Scholar] [CrossRef] [Green Version]
  37. Dehaene, S.; Changeux, J.-P. Neural mechanisms for access to consciousness. Cogn. Neurosci. 2004, 3, 1145–1158. [Google Scholar]
  38. Gennaro, R.J. Higher-order theories of consciousness. In The Bloomsbury Companion to the Philosophy of Consciousness; Bloomsbury Publishing: London, UK, 2018; p. 142. [Google Scholar]
  39. Lau, H. Consciousness, Metacognition, & Perceptual Reality Monitoring. PsyArXiv 2019. [Google Scholar] [CrossRef]
  40. Oizumi, M.; Albantakis, L.; Tononi, G. From the Phenomenology to the Mechanisms of Consciousness: Integrated Information Theory 3.0. PLoS Comput. Biol. 2014, 10, e1003588. [Google Scholar] [CrossRef] [Green Version]
  41. Paoletti, P.; Ben-Soussan, T.D. Reflections on Inner and Outer Silence and Consciousness Without Contents According to the Sphere Model of Consciousness. Front. Psychol. 2020, 11, 1807. [Google Scholar] [CrossRef]
  42. Oudeyer, P.-Y.; Kaplan, F. What is intrinsic motivation? A typology of computational approaches. Front. Neurorobotics 2009, 1, 6. [Google Scholar] [CrossRef] [Green Version]
  43. Gottlieb, J.; Lopes, M.; Oudeyer, P.-Y. Motivated Cognition: Neural and Computational Mechanisms of Curiosity, Attention, and Intrinsic Motivation; Emerald Group Publishing Limited: Bingley, UK, 2016; pp. 149–172. [Google Scholar] [CrossRef] [Green Version]
  44. Millar, J. An Ethics Evaluation Tool for Automating Ethical Decision-Making in Robots and Self-Driving Cars. Appl. Artif. Intell. 2016, 30, 787–809. [Google Scholar] [CrossRef]
  45. Smallwood, J.; Bernhardt, B.C.; Leech, R.; Bzdok, D.; Jefferies, E.; Margulies, D.S. The default mode network in cognition: A topographical perspective. Nat. Rev. Neurosci. 2021, 22, 503–513. [Google Scholar] [CrossRef]
  46. Fleming, S.M.; Huijgen, J.; Dolan, R.J. Prefrontal Contributions to Metacognition in Perceptual Decision Making. J. Neurosci. 2012, 32, 6117–6125. [Google Scholar] [CrossRef] [Green Version]
  47. Fleming, S.M.; Ryu, J.; Golfinos, J.G.; Blackmon, K.E. Domain-specific impairment in metacognitive accuracy following anterior prefrontal lesions. Brain 2014, 137, 2811–2822. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Fleming, S.M.; Weil, R.S.; Nagy, Z.; Dolan, R.J.; Rees, G. Relating introspective accuracy to individual differences in brain structure. Science 2010, 329, 1541–1543. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. McCurdy, L.Y.; Maniscalco, B.; Metcalfe, J.; Liu, K.Y.; De Lange, F.P.; Lau, H. Anatomical Coupling between Distinct Metacognitive Systems for Memory and Visual Perception. J. Neurosci. 2013, 33, 1897–1906. [Google Scholar] [CrossRef] [Green Version]
  50. Ye, Q.; Zou, F.; Dayan, M.; Lau, H.; Hu, Y.; Kwok, S.C. Individual susceptibility to TMS affirms the precuneal role in meta-memory upon recollection. Brain Struct. Funct. 2019, 224, 2407–2419. [Google Scholar] [CrossRef]
  51. Hampton, R.R. Multiple demonstrations of metacognition in nonhumans: Converging evidence or multiple mechanisms? Comp. Cogn. Behav. Rev. 2009, 4, 17–28. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Middlebrooks, P.G.; Sommer, M.A. Metacognition in monkeys during an oculomotor task. J. Exp. Psychol. Learn. Mem. Cogn. 2011, 37, 325–337. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Miyamoto, K.; Osada, T.; Setsuie, R.; Takeda, M.; Tamura, K.; Adachi, Y.; Miyashita, Y. Causal neural network of metamemory for retrospection in primates. Science 2017, 355, 188–193. [Google Scholar] [CrossRef]
  54. Miyamoto, K.; Setsuie, R.; Osada, T.; Miyashita, Y. Reversible Silencing of the Frontopolar Cortex Selectively Impairs Metacognitive Judgment on Non-experience in Primates. Neuron 2018, 97, 980–989.e6. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Cai, Y.; Jin, Z.; Zhai, C.; Wang, H.; Wang, J.; Tang, Y.; Kwok, S.C. Time-Sensitive Prefrontal Involvement in Associating Confidence with Task Performance Illustrates Metacognitive Introspection in Monkeys; Cold Spring Harbor Laboratory: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  56. Kwok, S.C.; Cai, Y.; Buckley, M.J. Mnemonic Introspection in Macaques Is Dependent on Superior Dorsolateral Prefrontal Cortex But Not Orbitofrontal Cortex. J. Neurosci. 2019, 39, 5922–5934. [Google Scholar] [CrossRef] [PubMed]
  57. Masset, P.; Ott, T.; Lak, A.; Hirokawa, J.; Kepecs, A. Behavior- and Modality-General Representation of Confidence in Orbitofrontal Cortex. Cell 2020, 182, 112–126.e18. [Google Scholar] [CrossRef]
  58. Bayne, T.; Brainard, D.; Byrne, R.W.; Chittka, L.; Clayton, N.; Heyes, C.; Mather, J.; Ölveczky, B.; Shadlen, M.; Suddendorf, T.; et al. What is cognition? Curr. Biol. 2019, 29, R608–R615. [Google Scholar] [CrossRef] [PubMed]
  59. Gallup, G.G., Jr.; Anderson, J.R. Self-recognition in animals: Where do we stand 50 years later? Lessons from cleaner wrasse and other species. Psychol. Conscious. Theory Res. Pract. 2020, 7, 46. [Google Scholar]
  60. Mather, J. What is in an octopus’s mind? Anim. Sentience 2019, 4, 1. [Google Scholar] [CrossRef]
  61. Schnell, A.K.; Amodio, P.; Boeckle, M.; Clayton, N.S. How intelligent is a cephalopod? Lessons from comparative cognition. Biol. Rev. 2021, 96, 162–178. [Google Scholar] [CrossRef] [PubMed]
  62. Smith, J.E.; Ravi, N. The architecture of virtual machines. Computer 2005, 38, 32–38. [Google Scholar] [CrossRef] [Green Version]
  63. Gottlieb, J.; Oudeyer, P.-Y.; Lopes, M.; Baranes, A. Information-seeking, curiosity, and attention: Computational and neural mechanisms. Trends Cogn. Sci. 2013, 17, 585–593. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Ten, A.; Kaushik, P.; Oudeyer, P.-Y.; Gottlieb, J. Humans monitor learning progress in curiosity-driven exploration. Nat. Commun. 2021, 12, 5972. [Google Scholar] [CrossRef] [PubMed]
  65. Yeong, D.J.; Velasco-Hernandez, G.; Barry, J.; Walsh, J. Sensor and Sensor Fusion Technology in Autonomous Vehicles: A Review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef] [PubMed]
  66. Franklin, S.; Madl, T.; D’Mello, S.; Snaider, J. LIDA: A Systems-level Architecture for Cognition, Emotion, and Learning. IEEE Trans. Auton. Ment. Dev. 2014, 6, 19–41. [Google Scholar] [CrossRef]
  67. Anderson, J.R. Human Symbol Manipulation Within an Integrated Cognitive Architecture. Cogn. Sci. 2005, 29, 313–341. [Google Scholar] [CrossRef]
  68. Grossberg, S. Adaptive Resonance Theory: How a brain learns to consciously attend, learn, and recognize a changing world. Neural Netw. 2013, 37, 1–47. [Google Scholar] [CrossRef]
  69. Fields, C.; Glazebrook, J.F.; Marcianò, A. Reference Frame Induced Symmetry Breaking on Holographic Screens. Symmetry 2021, 13, 408. [Google Scholar] [CrossRef]
  70. Addazi, A.; Chen, P.; Fabrocini, F.; Fields, C.; Greco, E.; Lulli, M.; Marciano, A.; Pasechnik, R. Generalized holographic principle, gauge invariance and the emergence of gravity à la Wilczek. Front. Astron. Space Sci. 2021, 8, 1. [Google Scholar] [CrossRef]
  71. Fields, C.; Marcianò, A. Holographic Screens Are Classical Information Channels. Quantum Rep. 2020, 2, 326–336. [Google Scholar] [CrossRef]
  72. Fields, C.; Marcianò, A. Markov blankets are general physical interaction surfaces. Phys. Life Rev. 2019, 33, 109–111. [Google Scholar] [CrossRef] [PubMed]
  73. Fields, C. Some Consequences of the Thermodynamic Cost of System Identification. Entropy 2018, 20, 797. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  74. Sajid, N.; Convertino, L.; Friston, K. Cancer Niches and Their Kikuchi Free Energy. Entropy 2021, 23, 609. [Google Scholar] [CrossRef] [PubMed]
  75. Hesp, C.; Ramstead, M.; Constant, A.; Badcock, P.; Kirchhoff, M.; Friston, K. A multi-scale view of the emergent complexity of life: A free-energy proposal. In Evolution, Development and Complexity; Springer: Berlin/Heidelberg, Germany, 2019; pp. 195–227. [Google Scholar]
  76. Constant, A.; Ramstead, M.J.D.; Veissière, S.P.L.; Campbell, J.O.; Friston, K.J. A variational approach to niche construction. J. R. Soc. Interface 2018, 15, 20170685. [Google Scholar] [CrossRef] [PubMed]
  77. Bruineberg, J.; Rietveld, E.; Parr, T.; Van Maanen, L.; Friston, K.J. Free-energy minimization in joint agent-environment systems: A niche construction perspective. J. Theor. Biol. 2018, 455, 161–178. [Google Scholar] [CrossRef]
  78. Friston, K. A Free Energy Principle for Biological Systems. Entropy 2012, 14, 2100–2121. [Google Scholar] [CrossRef]
  79. Fields, C.; Levin, M. Somatic multicellularity as a satisficing solution to the prediction-error minimization problem. Commun. Integr. Biol. 2019, 12, 119–132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  80. Friston, K.; Parr, T.; Zeidman, P. Bayesian model reduction. arXiv 2018, arXiv:1805.07092. [Google Scholar]
  81. Campbell, J.O. Universal Darwinism As a Process of Bayesian Inference. Front. Syst. Neurosci. 2016, 10, 49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  82. Fisher, R.A. The Genetical Theory of Natural Selection; Oxford: London, UK, 1930. [Google Scholar]
  83. Frank, S.A. Natural selection. V. How to read the fundamental equations of evolutionary change in terms of information theory. J. Evol. Biol. 2012, 25, 2377–2396. [Google Scholar] [CrossRef]
  84. Fields, C.; Levin, M. Does Evolution Have a Target Morphology? Organisms. J. Biol. Sci. 2020, 4, 57–76. [Google Scholar]
  85. Pezzulo, G.; Parr, T.; Friston, K. The evolution of brain architectures for predictive coding and active inference. Philos. Trans. R. Soc. B 2022, 377, 20200531. [Google Scholar] [CrossRef] [PubMed]
  86. Ueltzhöffer, K. Deep active inference. Biol. Cybern. 2018, 112, 547–573. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  87. Landauer, R. Information is a physical entity. Phys. A Stat. Mech. Its Appl. 1999, 263, 63–67. [Google Scholar] [CrossRef] [Green Version]
  88. Landauer, R. Irreversibility and Heat Generation in the Computing Process. IBM J. Res. Dev. 1961, 5, 183–191. [Google Scholar] [CrossRef]
  89. Fields, C. Trajectory Recognition as the Basis for Object Individuation: A Functional Model of Object File Instantiation and Object-Token Encoding. Front. Psychol. 2011, 2, 49. [Google Scholar] [CrossRef] [Green Version]
  90. Aquino, G.; Endres, R.G. Increased accuracy of ligand sensing by receptor internalization. Phys. Rev. E 2010, 81, 021909. [Google Scholar] [CrossRef] [Green Version]
  91. Mao, H.; Cremer, P.S.; Manson, M.D. A sensitive, versatile microfluidic assay for bacterial chemotaxis. Proc. Natl. Acad. Sci. USA 2003, 100, 5449–5454. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  92. Berg, H. Random Walks in Biology; Princeton University Press: Princeton, NJ, USA, 1993. [Google Scholar]
  93. Mukherjee, S.; Ghosh, R.N.; Maxfield, F.R. Endocytosis. Physiol. Rev. 1997, 77, 759–803. [Google Scholar] [CrossRef]
  94. Mehta, P.; Lang, A.H.; Schwab, D.J. Landauer in the Age of Synthetic Biology: Energy Consumption and Information Processing in Biochemical Networks. J. Stat. Phys. 2016, 162, 1153–1166. [Google Scholar] [CrossRef] [Green Version]
  95. Mehta, P.; Schwab, D.J. Energetic costs of cellular computation. Proc. Natl. Acad. Sci. USA 2012, 109, 17978–17982. [Google Scholar] [CrossRef] [Green Version]
  96. Karpathy, A.; Johnson, J.; Fei-Fei, L. Visualizing and Understanding Recurrent Networks. arXiv 2015, arXiv:1506.02078. [Google Scholar]
  97. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  98. Hansen, N.; Arnold, D.V.; Auger, A. Evolution Strategies; Springer: Berlin/Heidelberg, Germany, 2015; pp. 871–898. [Google Scholar] [CrossRef]
  99. Hesp, C.; Smith, R.; Parr, T.; Allen, M.; Friston, K.J.; Ramstead, M.J.D. Deeply Felt Affect: The Emergence of Valence in Deep Active Inference. Neural Comput. 2021, 33, 398–446. [Google Scholar] [CrossRef]
  100. Lyon, P.; Kuchling, F. Valuing what happens: A biogenic approach to valence and (potentially) affect. Philos. Trans. R. Soc. Lond. B Biol. Sci. 2021, 376, 20190752. [Google Scholar] [CrossRef]
  101. Levin, M. Life, death, and self: Fundamental questions of primitive cognition viewed through the lens of body plasticity and synthetic organisms. Biochem. Biophys. Res. Commun. 2021, 564, 114–133. [Google Scholar] [CrossRef]
  102. Lotka, A.J. Contribution to the Theory of Periodic Reactions. J. Phys. Chem. 1910, 14, 271–274. [Google Scholar] [CrossRef] [Green Version]
  103. Cortez, M.H.; Weitz, J.S. Coevolution can reverse predator-prey cycles. Proc. Natl. Acad. Sci. USA 2014, 111, 7486–7491. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  104. Poggiale, J.-C.; Aldebert, C.; Girardot, B.; Kooi, B.W. Analysis of a predator–prey model with specific time scales: A geometrical approach proving the occurrence of canard solutions. J. Math. Biol. 2020, 80, 39–60. [Google Scholar] [CrossRef] [Green Version]
  105. Vanselow, A.; Wieczorek, S.; Feudel, U. When very slow is too fast-collapse of a predator-prey system. J. Theor. Biol. 2019, 479, 64–72. [Google Scholar] [CrossRef] [Green Version]
  106. Park, J.; Lee, J.; Kim, T.; Ahn, I.; Park, J. Co-Evolution of Predator-Prey Ecosystems by Reinforcement Learning Agents. Entropy 2021, 23, 461. [Google Scholar] [CrossRef] [PubMed]
  107. Rosenzweig, M.L.; MacArthur, R.H. Graphical representation and stability conditions of predator-prey interactions. Am. Nat. 1963, 97, 209–223. [Google Scholar] [CrossRef]
  108. Velzen, E. Predator coexistence through emergent fitness equalization. Ecology 2020, 101, e02995. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  109. Auger, P.; Kooi, B.W.; Bravo De La Parra, R.; Poggiale, J.-C. Bifurcation analysis of a predator–prey model with predators using hawk and dove tactics. J. Theor. Biol. 2006, 238, 597–607. [Google Scholar] [CrossRef] [PubMed]
  110. Cortez, M.H.E.; Stephen, P. Understanding Rapid Evolution in Predator-Prey Interactions Using the Theory of Fast-Slow Dynamical Systems. Am. Nat. 2010, 176, E109–E127. [Google Scholar] [CrossRef]
  111. Romero-Muñoz, A.; Maffei, L.; Cuéllar, E.; Noss, A.J. Temporal separation between jaguar and puma in the dry forests of southern Bolivia. J. Trop. Ecol. 2010, 26, 303–311. [Google Scholar] [CrossRef]
  112. Karanth, K.U.; Sunquist, M.E. Behavioural correlates of predation by tiger (Panthera tigris), leopard (Panthera pardus) and dhole (Cuon alpinus) in Nagarahole, India. J. Zool. 2000, 250, 255–265. [Google Scholar] [CrossRef]
  113. Martin, A.; Simon, C. Temporal Variation in Insect Life Cycles. BioScience 1990, 40, 359–367. [Google Scholar] [CrossRef]
  114. Kingsolver, J.G.; Arthur Woods, H.; Buckley, L.B.; Potter, K.A.; Maclean, H.J.; Higgins, J.K. Complex Life Cycles and the Responses of Insects to Climate Change. Integr. Comp. Biol. 2011, 51, 719–732. [Google Scholar] [CrossRef] [Green Version]
  115. Laan, J.D.V.D.; Hogeweg, P. Predator—prey coevolution: Interactions across different timescales. Proc. R. Soc. London. Ser. B Biol. Sci. 1995, 259, 35–42. [Google Scholar] [CrossRef]
  116. Bosiger, Y.J.; Lonnstedt, O.M.; McCormick, M.I.; Ferrari, M.C.O. Learning Temporal Patterns of Risk in a Predator-Diverse Environment. PLoS ONE 2012, 7, e34535. [Google Scholar] [CrossRef] [PubMed]
  117. Ishii, Y.; Shimada, M. The effect of learning and search images on predator–prey interactions. Popul. Ecol. 2010, 52, 27–35. [Google Scholar] [CrossRef]
  118. Wang, X.; Cheng, J.; Wang, L. A reinforcement learning-based predator-prey model. Ecol. Complex. 2020, 42, 100815. [Google Scholar] [CrossRef]
  119. Yamada, J.; Shawe-Taylor, J.; Fountas, Z. Evolution of a Complex Predator-Prey Ecosystem on Large-scale Multi-Agent Deep Reinforcement Learning. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020. [Google Scholar]
  120. Wilson, E.; Morgan, T. Chiasmatype and crossing over. Am. Nat. 1920, 54, 193–219. [Google Scholar] [CrossRef] [Green Version]
  121. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  122. Forrest, S. Genetic algorithms: Principles of natural selection applied to computation. Science 1993, 261, 872–878. [Google Scholar] [CrossRef] [Green Version]
  123. Forrest, S. Genetic algorithms. ACM Comput. Surv. CSUR 1996, 28, 77–80. [Google Scholar] [CrossRef]
  124. Forrest, S.; Mitchell, M. What makes a problem hard for a genetic algorithm? Some anomalous results and their explanation. Mach. Learn. 1993, 13, 285–319. [Google Scholar]
  125. Shrestha, A.; Mahmood, A. Improving Genetic Algorithm with Fine-Tuned Crossover and Scaled Architecture. J. Math. 2016, 2016, 4015845. [Google Scholar] [CrossRef] [Green Version]
  126. Prentis, P.J.; Wilson, J.R.U.; Dormontt, E.E.; Richardson, D.M.; Lowe, A.J. Adaptive evolution in invasive species. Trends Plant Sci. 2008, 13, 288–294. [Google Scholar] [CrossRef] [PubMed]
  127. Kazarlis, S.; Petridis, V. Varying Fitness Functions in Genetic Algorithms: Studying the Rate of Increase of the Dynamic Penalty Terms; Springer: Berlin/Heidelberg, Germany, 1998; pp. 211–220. [Google Scholar] [CrossRef]
  128. Jin, Y.; Branke, J. Evolutionary Optimization in Uncertain Environments—A Survey. IEEE Trans. Evol. Comput. 2005, 9, 303–317. [Google Scholar] [CrossRef] [Green Version]
  129. Handa, H. Fitness Function for Finding out Robust Solutions on Time-Varying Functions; ACM Press: New York, NY, USA, 2006. [Google Scholar]
  130. Bull, L. On coevolutionary genetic algorithms. Soft Comput. 2001, 5, 201–207. [Google Scholar] [CrossRef]
  131. Bull, L. Coevolutionary Species Adaptation Genetic Algorithms: A Continuing SAGA on Coupled Fitness Landscapes; Springer: Berlin/Heidelberg, Germany, 2005; pp. 322–331. [Google Scholar] [CrossRef]
  132. Bull, L. Coevolutionary species adaptation genetic algorithms: Growth and mutation on coupled fitness landscapes. In Proceedings of the 2005 IEEE Congress on Evolutionary Computation, Edinburgh, UK, 2–5 September 2005; pp. 559–564. [Google Scholar]
  133. Eiben, A.E.; Smith, J.E. Coevolutionary Systems; Springer: Berlin/Heidelberg, Germany, 2015; pp. 223–229. [Google Scholar] [CrossRef]
  134. Paredis, J. Coevolution, Memory and Balance. IJCAI 1999, 10, 1212–1217. [Google Scholar]
  135. Mitchell, W.A. Multi-behavioral strategies in a predator-prey game: An evolutionary algorithm analysis. Oikos 2009, 118, 1073–1083. [Google Scholar] [CrossRef]
  136. Paredis, J. Coevolutionary computation. Artif. Life 1995, 2, 355–375. [Google Scholar] [CrossRef]
  137. Kauffman, S.A.; Johnsen, S. Coevolution to the edge of chaos: Coupled fitness landscapes, poised states, and coevolutionary avalanches. J. Theor. Biol. 1991, 149, 467–505. [Google Scholar] [CrossRef]
  138. Gonog, L.; Zhou, Y. A Review: Generative Adversarial Networks. In Proceedings of the 2019 14th IEEE Conference on Industrial Electronics and Applications (ICIEA), Xi’an, China, 19–21 June 2019. [Google Scholar]
  139. Talas, L.; Fennell, J.G.; Kjernsmo, K.; Cuthill, I.C.; Scott-Samuel, N.E.; Baddeley, R.J. CamoGAN: Evolving optimum camouflage with Generative Adversarial Networks. Methods Ecol. Evol. 2020, 11, 240–247. [Google Scholar] [CrossRef]
  140. Liu, M.-Y.; Tuzel, O. Coupled generative adversarial networks. Adv. Neural Inf. Processing Syst. 2016, 29, 469–477. [Google Scholar]
  141. Wang, J.; Jiang, J. Conditional Coupled Generative Adversarial Networks for Zero-Shot Domain Adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  142. Qi, M.; Wang, Y.; Li, A.; Luo, J. STC-GAN: Spatio-Temporally Coupled Generative Adversarial Networks for Predictive Scene Parsing. IEEE Trans. Image Processing 2020, 29, 5420–5430. [Google Scholar] [CrossRef]
  143. Levin, M. The wisdom of the body: Future techniques and approaches to morphogenetic fields in regenerative medicine, developmental biology and cancer. Regen. Med. 2011, 6, 667–673. [Google Scholar] [CrossRef] [Green Version]
  144. Levin, M. Morphogenetic fields in embryogenesis, regeneration, and cancer: Non-local control of complex patterning. Biosystems 2012, 109, 243–261. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  145. Watson, R.A.; Szathmáry, E. How Can Evolution Learn? Trends Ecol. Evol. 2016, 31, 147–157. [Google Scholar] [CrossRef] [Green Version]
  146. Kashtan, N.; Alon, U. Spontaneous evolution of modularity and network motifs. Proc. Natl. Acad. Sci. USA 2005, 102, 13773–13778. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  147. Parter, M.; Kashtan, N.; Alon, U. Facilitated Variation: How Evolution Learns from Past Environments To Generalize to New Environments. PLoS Comput. Biol. 2008, 4, e1000206. [Google Scholar] [CrossRef]
  148. Fields, C.; Levin, M. Scale-Free Biology: Integrating Evolutionary and Developmental Thinking. BioEssays 2020, 42, 1900228. [Google Scholar] [CrossRef]
  149. Tauber, C.A.; Tauber, M.J. Insect seasonal cycles: Genetics and evolution. Annu. Rev. Ecol. Syst. 1981, 12, 281–308. [Google Scholar] [CrossRef]
  150. Lathe, R. Fast tidal cycling and the origin of life. Icarus 2004, 168, 18–22. [Google Scholar] [CrossRef]
  151. Gordon, R.; Mikhailovsky, G. There were plenty of day/night cycles that could have accelerated an origin of life on Earth, without requiring panspermia. In Planet Formation and Panspermia: New Prospects for the Movement of Life through Space; Wiley: Hoboken, NJ, USA, 2021; pp. 195–206. [Google Scholar]
  152. Gehring, W.; Rosbash, M. The coevolution of blue-light photoreception and circadian rhythms. J. Mol. Evol. 2003, 57, S286–S289. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Generic architecture of a metaprocessor. A metaprocessor (red) regulates an object processor (blue) both internally and externally. The metaprocessor requires its own local memory of the object processor’s behavior.
Figure 1. Generic architecture of a metaprocessor. A metaprocessor (red) regulates an object processor (blue) both internally and externally. The metaprocessor requires its own local memory of the object processor’s behavior.
Entropy 24 00601 g001
Figure 2. General form of bipartite interaction dynamics in the absence of quantum entanglement. (a) Finite systems S and E interact via a boundary that serves as an MB. (b) From S’s perspective, comprises sensory bits (in state s, red shading) that are inputs to S’s computations and action bits (in state a, blue shading) that are S’s outputs. When the free-energy costs of computation are taken into account, as in Section 3.1 below, some input bits must be allocated to fuel for computation and some output bits must be allocated to waste heat dissipation (hatched area); these bits are uninformative to S.
Figure 2. General form of bipartite interaction dynamics in the absence of quantum entanglement. (a) Finite systems S and E interact via a boundary that serves as an MB. (b) From S’s perspective, comprises sensory bits (in state s, red shading) that are inputs to S’s computations and action bits (in state a, blue shading) that are S’s outputs. When the free-energy costs of computation are taken into account, as in Section 3.1 below, some input bits must be allocated to fuel for computation and some output bits must be allocated to waste heat dissipation (hatched area); these bits are uninformative to S.
Entropy 24 00601 g002
Figure 3. Markov blanket state separation in active inference. Middle: The internal (pink) and external states (white) of each agent are separated by a Markov blanket, comprised of sensory (grey) and active (green) states. This architecture can be applied to various forms of information processing, two of which are shown above and below. By associating the gradient flows of the Markov blanket partition with Bayesian belief updating, self-organization of internal states—in response to sensory fluctuations—can be thought of as perception, while active states couple internal states back to hidden external states vicariously, to provide a mathematical formulation of action and behavior. Top: Visual processing. Internal states are made up of the brain, which directs movement of the eye through abductor muscles as active states. Sensory states as the photoreceptors in the eye perceive the external states in the visual field of view. Bottom: Transcriptional machinery. Internal states are given by the gene expression levels plus epigenetic modifications. Intracellular components such as ribosomes, smooth and rough endoplasmic reticulum, and Golgi apparatus implement protein translation and delivery as the active states. Sensory states correspond to the surface states of the cell membrane, such as ligand receptors, ion channel states, and gap junctions between cells. External states are associated with extracellular concentrations and the states of other cells. Image credits: The cell schematic has been adapted from an image by Judith Stoffer supplied at the National Institute of General Medical Sciences, NIH (CC BY-NC 2.0). The brain schematic has been adapted from an image of a head cross-section on openclipart.org by Kevin David Pointon under a public domain license, and an image of an eye with abductor muscles from Wikimedia by Patrick J. Lynch (CC BY-NC 2.5).
Figure 3. Markov blanket state separation in active inference. Middle: The internal (pink) and external states (white) of each agent are separated by a Markov blanket, comprised of sensory (grey) and active (green) states. This architecture can be applied to various forms of information processing, two of which are shown above and below. By associating the gradient flows of the Markov blanket partition with Bayesian belief updating, self-organization of internal states—in response to sensory fluctuations—can be thought of as perception, while active states couple internal states back to hidden external states vicariously, to provide a mathematical formulation of action and behavior. Top: Visual processing. Internal states are made up of the brain, which directs movement of the eye through abductor muscles as active states. Sensory states as the photoreceptors in the eye perceive the external states in the visual field of view. Bottom: Transcriptional machinery. Internal states are given by the gene expression levels plus epigenetic modifications. Intracellular components such as ribosomes, smooth and rough endoplasmic reticulum, and Golgi apparatus implement protein translation and delivery as the active states. Sensory states correspond to the surface states of the cell membrane, such as ligand receptors, ion channel states, and gap junctions between cells. External states are associated with extracellular concentrations and the states of other cells. Image credits: The cell schematic has been adapted from an image by Judith Stoffer supplied at the National Institute of General Medical Sciences, NIH (CC BY-NC 2.0). The brain schematic has been adapted from an image of a head cross-section on openclipart.org by Kevin David Pointon under a public domain license, and an image of an eye with abductor muscles from Wikimedia by Patrick J. Lynch (CC BY-NC 2.5).
Entropy 24 00601 g003
Figure 4. Schematic of adaptive learning in predator–prey models. Left: A typical predator–prey model simulation based on Equation (5), characterized by persistent population size oscillations of both predator and prey, where peaks in prey precede peaks in predator populations. Right: Predator–prey model simulations with learned adaptive coevolution of both predator and prey based on Equation (6) and the work done by Park et al. [106], characterized by initially dampened oscillations followed by only marginally oscillating, stable population sizes of both predator and prey.
Figure 4. Schematic of adaptive learning in predator–prey models. Left: A typical predator–prey model simulation based on Equation (5), characterized by persistent population size oscillations of both predator and prey, where peaks in prey precede peaks in predator populations. Right: Predator–prey model simulations with learned adaptive coevolution of both predator and prey based on Equation (6) and the work done by Park et al. [106], characterized by initially dampened oscillations followed by only marginally oscillating, stable population sizes of both predator and prey.
Entropy 24 00601 g004
Figure 5. Coupled Genetic Algorithm Schematic. Top: Schematic of a standard genetic algorithm. An initial population (purple) of individuals or chromosomes (orange), consisting of individual bits or genes (black), reproduce with crossover of random sets of genes and mutations. Fitness according to a fixed fitness function is calculated for each chromosome, after which individuals with highest fitness are selected and cloned to form a new population. After multiple iterations of this process, the algorithm stops when no higher fitness can be generated in any new individual. Bottom: Schematic of a coupled genetic algorithm. Coevolving populations undergo reproduction as usual but are then coupled in the computation of each other’s fitness function, which in turn are variable and can incorporate integration of past fitness functions as a form of memory, at which point learning methods can be implemented.
Figure 5. Coupled Genetic Algorithm Schematic. Top: Schematic of a standard genetic algorithm. An initial population (purple) of individuals or chromosomes (orange), consisting of individual bits or genes (black), reproduce with crossover of random sets of genes and mutations. Fitness according to a fixed fitness function is calculated for each chromosome, after which individuals with highest fitness are selected and cloned to form a new population. After multiple iterations of this process, the algorithm stops when no higher fitness can be generated in any new individual. Bottom: Schematic of a coupled genetic algorithm. Coevolving populations undergo reproduction as usual but are then coupled in the computation of each other’s fitness function, which in turn are variable and can incorporate integration of past fitness functions as a form of memory, at which point learning methods can be implemented.
Entropy 24 00601 g005
Figure 6. Schematic of Coupled Generative Adversarial Networks. Top: Schematic of a standard generative adversarial network applied to the problem of camouflage. A generator (blue) neural network generates data xg with the goal of emulating the potential distributions of real data xr as best as possible. In this example, inspired by Talas et al. [139], the generator generates stripe patterns that, when emerged in the habitat of a Siberian tiger, are aimed at fooling the discriminator in not being able to discern generated tiger patterns from tree patterns in the background. A second neural network called discriminator receives unlabeled data from both real and generated distributions and aims to label them correctly by origin. Bottom: Schematic of a coupled generative adversarial network. When fitness landscapes of two populations are coupled, each generative adversarial network type agent tries to internally generate and discern trait patterns that correspond to the observed traits of the other population, thereby learning and predicting the traits of the other population. Because in a coupled generative adversarial network weights between the first few generator layers, and last few discriminator layers are shared, this configuration allows the coupled generative adversarial network to learn a joint fitness distribution underlying traits without correspondence supervision [140]. Tiger and birch tree images are adapted from images taken from flickr.com (last accessed 02/09/2022) under the CC BY-NC-SA 2.0 license.
Figure 6. Schematic of Coupled Generative Adversarial Networks. Top: Schematic of a standard generative adversarial network applied to the problem of camouflage. A generator (blue) neural network generates data xg with the goal of emulating the potential distributions of real data xr as best as possible. In this example, inspired by Talas et al. [139], the generator generates stripe patterns that, when emerged in the habitat of a Siberian tiger, are aimed at fooling the discriminator in not being able to discern generated tiger patterns from tree patterns in the background. A second neural network called discriminator receives unlabeled data from both real and generated distributions and aims to label them correctly by origin. Bottom: Schematic of a coupled generative adversarial network. When fitness landscapes of two populations are coupled, each generative adversarial network type agent tries to internally generate and discern trait patterns that correspond to the observed traits of the other population, thereby learning and predicting the traits of the other population. Because in a coupled generative adversarial network weights between the first few generator layers, and last few discriminator layers are shared, this configuration allows the coupled generative adversarial network to learn a joint fitness distribution underlying traits without correspondence supervision [140]. Tiger and birch tree images are adapted from images taken from flickr.com (last accessed 02/09/2022) under the CC BY-NC-SA 2.0 license.
Entropy 24 00601 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kuchling, F.; Fields, C.; Levin, M. Metacognition as a Consequence of Competing Evolutionary Time Scales. Entropy 2022, 24, 601. https://doi.org/10.3390/e24050601

AMA Style

Kuchling F, Fields C, Levin M. Metacognition as a Consequence of Competing Evolutionary Time Scales. Entropy. 2022; 24(5):601. https://doi.org/10.3390/e24050601

Chicago/Turabian Style

Kuchling, Franz, Chris Fields, and Michael Levin. 2022. "Metacognition as a Consequence of Competing Evolutionary Time Scales" Entropy 24, no. 5: 601. https://doi.org/10.3390/e24050601

APA Style

Kuchling, F., Fields, C., & Levin, M. (2022). Metacognition as a Consequence of Competing Evolutionary Time Scales. Entropy, 24(5), 601. https://doi.org/10.3390/e24050601

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop