Next Article in Journal
Finite-Time Stochastic Stability Analysis of Permanent Magnet Synchronous Motors with Noise Perturbation
Next Article in Special Issue
Detection of Static and Mobile Targets by an Autonomous Agent with Deep Q-Learning Abilities
Previous Article in Journal
Alpha-Beta Hybrid Quantum Associative Memory Using Hamming Distance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Subjective Trusts for the Control of Mobile Robots under Uncertainty

1
Department Industrial Engineering, Ariel University, Ariel 4076414, Israel
2
Laboratory for AI, Machine Learning, Business and Data Analytics, Tel Aviv University, Tel Aviv 6997801, Israel
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(6), 790; https://doi.org/10.3390/e24060790
Submission received: 4 May 2022 / Revised: 2 June 2022 / Accepted: 4 June 2022 / Published: 5 June 2022

Abstract

:
The paper deals with the methods of handling imperfect information and the control of mobile robots acting under uncertainty. It is assumed that the robots act in autonomous regime and are controlled internally without referring the externally defined probabilities of the states and actions. To control the activity of the robots, we suggest the novel multi-valued logic techniques based on the recently developed measures, known as the subjective trusts. In addition to specification of the robots’ movements, such a technique allows for direct definition of the robots’ swarming using the methods of artificial neural networks with mobile neurons. The suggested methods are verified by numerical simulations and running examples. The resulting framework forms a basis for processing non-probabilistic uncertainties and making individual decisions with imperfect information.

1. Introduction

Functionality of autonomous systems includes three main processes: observing and measuring the environmental and system’s states, making decisions and prescribing the future actions, and conducting the chosen action aimed to change the state of the environment or of the system. In the case of complete certainty in observations and errorless decision making and acting, the models of control implement the methods originated by Maxwell in 1868 [1]. Following this approach, behavior of the system is defined by a certain dynamical system that prescribes the further actions with respect to the current observations. In the case of uncertainty, the control process is more complicated and requires handling imperfect or incomplete information about the states of the environment and of the system, processing inaccurate and ambiguous decisions and correcting erroneous actions [2,3].
Starting in the later 1950s, operating with uncertainties in the autonomous systems control is based on probabilistic methods [4,5,6]. It is assumed that the uncertainties in the observation results and the errors in the decisions and actions can be defined in the terms of random variables and their probabilities.
However, the use of probabilistic methods involves presumptions, which follows from the nature of probability, but do not hold in the framework of autonomous systems control. The difficulties in application of the probabilistic methods in the control of the swarm are due to two reasons.
The first is inconsistence of the probabilities to the considered situation. For example, consider a driver in the traffic flow who observes a pedestrian at the roadside. The probability that the pedestrian will cross the road on the red light is close to zero, but since it is possible, each experienced driver is ready to brake near the crosswalk. For the other example, consider observation of unknown or incompletely defined objects in a fog. Since the objects are unknown, the distances to the objects in the fog are also undefined. Hence, the probability that the certain object is observed is undefinable. Most such difficulties are considered in detail in the possibility theory [7,8].
The second reason is more complicated and follows from the nature of decision-making in the swarms, which does not imply direct minimization of the probabilities of outcomes or maximization of the probabilities of rewards (see the classical results by Kahneman and Tversky [9] and their recent validation [10]). Even in the simple case of motion toward the single target, an optimal behavior of the swarm involves a certain level of altruism of the swarm members such that in some steps the agents deny their movements to allow the movements of their neighbors [11]. In each specific task, such steps can be specified or at least approximated by stochastic optimization, however, in the general case, control of the swarm and the reactive activity of its members that leads to near optimal teleological behavior is still unclear.
In the paper we address a novel method for handling uncertainty based on the recently developed subjective trust measures [12,13]. Together with the parameterized version of the uninorm [14] and absorbing norm [15] aggregators, these measures form a formal algebra [16,17], which, in a certain sense, connects probabilistic and possibilistic approaches. Recently these measures were applied for formal description of prospects and demonstrated good correspondence with the experimental results [12]. In this paper, we continue this direction and define control of the swarm that allows to count the altruistic behavior of the agents without considering specific models and optimization of the swarm activity.
Starting from the observed relation between the subjective trusts, possibilities and necessities, we define the extended version of the uninorm and absorbing norm aggregators and the extended version of the algebra with these aggregators. The obtained algebra is considered as the algebra of control variables, which are governed directly by the decisions made by the means of the same algebra. Such an approach allows resolving uncertainties in the decision making and in the control at the same stage of the system’s evolution using the same tools. In addition, since the suggested algebra acts on the bounded interval, the processes are naturally limited, which prevents divergence of the controls even in the cases of erroneous decisions.
The suggested method is illustrated by the example of control of mobile robots. In the example, we start from the known model of a neural network with mobile neurons [18] described using subjective trusts [13] and extend this model to the use of control variables in the defined algebra. Then, swarming of mobile robots [19,20] is defined by direct implementation of the methods widely used in the descriptions of the ensembles of the neurons [21]. However, in contrast to the known methods, mobility of the neurons in the network and, consequently, of the robots in the swarm, is based both on the internal states of the neurons and the robots and on the states of the synapses, which are the values of the interconnections. In addition, we define the process of separation and unification of the synapses and specify the corresponding motion of the neurons and robots.
It should be noticed that the first attempt to use the concept of trusts for control of the mobile robots most likely appeared in the works [22,23]. In this performance centric method, the robot’s behavior is defined by the human-to-robot trust and by the human self-confidence. In other words, the higher the trust to the robot, the higher the quality of its performance and, similarly, the trust to the human is higher the better the human controls the robot. The suggested method, in contrast, can be considered as decision centric, in which the trust is governed by a priori defined algebraic rules, and the performance of the robot is a result of the operations with the trusts.
In addition, in the suggested control method the robot is defined as a dipole consisting of two neurons. However, in contrast to the original dipole activity [24], in the suggested model the dipoles are also governed by operators of the suggested algebra.
Numerical simulations verify the methods and demonstrate their convenience for definition of internal control under uncertainty with imperfect information about the states of the system and its actions. Computational complexity of the algorithms of robots’ control depends on the implemented methods of calculation of the subjective trusts. In the simulations, the trusts were calculated using hyperbolic tangent function, which does not add complexity to the algorithms.

2. Algebra of Control Variables

Control of the mobile robot is defined in the extended version of recently developed algebra that allows consideration in the same framework for both logical and arithmetical operations. We start with general definition of this algebra of multivalued logic and then present its extension on the control variables.

2.1. Algebraic Structure for Multivalued Logic

Let θ : [ 0 ,   1 ] × [ 0 ,   1 ] [ 0 ,   1 ] be a uninorm (or uninorm aggregator) [14] with neutral or identity element θ [ 0 ,   1 ] , and ϑ : [ 0 ,   1 ] × [ 0 ,   1 ] [ 0 ,   1 ] be an absorbing norm (or absorbing norm aggregator also known as null norm) [15] with absorbing element ϑ [ 0 ,   1 ] . The first aggregator, with respect to the value of θ , extends Boolean a n d and o r operators and the second aggregator extends Boolean n o t   x o r operator.
Similar to the Boolean binary operators, the uninorm θ and absorbing norm ϑ are symmetric and meet the commutative and associative properties. In addition, the uninorm θ is transitive. The neutral θ and absorbing ϑ elements play the role of unit and zero for their operators, respectively, such that for any x [ 0 ,   1 ] it holds that θ θ x = x and ϑ ϑ x = ϑ . It was proven [25] that for any x , y [ 0 ,   1 ] there exist the functions u : ( 0 ,   1 ) ( , ) and v : ( 0 ,   1 ) ( , ) called generator functions such that
x θ y = u 1 ( u ( x ) + u ( y ) ) ,
x ϑ y = v 1 ( v ( x ) × v ( y ) ) .
Generator functions u and v are continuous, strictly monotonously increasing with the zeroes u ( θ ) = v ( ϑ ) = 0 and with the limits lim x 0 u ( x ) = lim x 0 v ( x ) = and lim x 1 u ( x ) = lim x 1 v ( x ) = + . The last property allows definition of the uninorm and absorbing norm on the boundary values 0 and 1 of the interval [ 0 ,   1 ] .
It is easy to show [17] that the inverse functions u 1 : ( , ) ( 0 ,   1 ) and v 1 : ( , ) ( 0 ,   1 ) have the same properties as probability distributions and, in general, can be defined by any positive sigmoid functions with the values in the interval [ 0 ,   1 ] . Then, generator functions u and v are the inverses of the corresponding sigmoid functions and can be considered as quantile functions.
The values from the interval [ 0 ,   1 ] together with the operations defined by the uninorm and the absorbing norm form an algebra A = ( [ 0 , 1 ] , θ , ϑ ) , in which uninorm θ acts as a summation operator with zero θ and absorbing norm ϑ acts as a multiplication operator with the zero ϑ [16,17]. For completeness, algebra A also defined the operations
x θ y = u 1 ( u ( x ) u ( y ) ) ,
x ϑ y = v 1 ( v ( x ) / v ( y ) ) ,
where v ( y ) 0 .
Algebra A extends the Boolean algebra B =   { 0 ,   1 } ,   ,   , which with usual binary conjunction and disjunction operators, and its multivalued version =   [ 0 ,   1 ] ,   ,   defined using t -norm and t -conorm [26]. In addition, algebra A can be considered as an arithmetic system on the interval [ 0 ,   1 ] in which uninorm θ and absorbing norm ϑ are associated respectively with the weighted arithmetical summation “ + ” and multiplication “ × ”. Consequently, operations θ and ϑ are associated, respectively, with the weighted arithmetical subtraction “ ” and division “ / ”.
In general, algebra A is non-distributive, that is
( x θ y ) ϑ z ( x ϑ z ) θ ( y ϑ z ) ,
while u v or θ ϑ ; however, if u = v or θ = ϑ then the distributivity property holds [16,17].
Assume that generator functions u and v are equivalent with equivalent neutral and absorbing elements and denote w = u = v and η = θ = ϑ . For an event A the trusts τ ( A ) about A are interpreted as follows:
  • τ ( A ) = 1 means that A is necessary and τ ( A ) = 0 means that A is impossible;
  • τ ( A ) = w 1 ( 1 ) means that A is probable and τ ( A ) = w 1 ( 1 ) means that A is improbable;
  • τ ( A ) = w 1 ( w 1 ( 1 ) ) means that A is possible and τ ( A ) = w 1 ( w 1 ( 1 ) ) means that A is unnecessary.
Such interpretation means that if, for example, the trust in the event A is τ ( A ) = w 1 ( 1 ) , then the probability of the event A is p ( A ) = 1 and vice versa, and if the trust in the event A is τ ( A ) = w 1 ( 1 ) , then the probability of the event A is p ( A ) = 0 and vice versa. For the possibility and the necessity, the interpretation is similar.
For example, assume that generator function w is
w ( x ) = ln ( x b / ( 1 x b ) ) ,       x ( 0 ,   1 ) ,
where b = 1 / log 2 ( η ) . Then the inverse w 1 of the generator function is a sigmoid function
w 1 ( ξ ) = ( exp ( ξ ) / ( 1 + exp ( ξ ) ) ) 1 / b ,       ξ ( , ) .
These functions are shown in Figure 1.
The trusts defined by these functions for the pairs of necessity-impossibility, probability-improbability, and possibility-unnecessity are shown in Figure 2.
It is seen that on one side the trust τ ( A ) = 0.73 in probable events is higher than the trust τ ( A ) = 0.67 in possible events, but is lower than the trust τ ( A ) = 1.00 in necessary events. On the other side, the trust τ ( A ) = 0.27 in improbable events is higher than the trust τ ( A ) = 0.00 in impossible events, but is lower than the trust τ ( A ) = 0.33 in unnecessary events.
If A is interpreted as a proposition, then from a logical point of view the indicated truth values are interpreted as follows:
  • τ ( A ) = 1 means that A is objectively true and τ ( A ) = 0 means that A is objectively false;
  • τ ( A ) = w 1 ( 1 ) means that A is subjectively true (true from the observer’s point of view) and τ ( A ) = w 1 ( 1 ) means that A is subjectively false (false from the observer’s point of view);
  • τ ( A ) = w 1 ( w 1 ( 1 ) ) means that A seems to be true and τ ( A ) = w 1 ( w 1 ( 1 ) ) means that A is seems to be false.
The truth values of the subjective truth and false are denoted by I w and O w , respectively, and these values hold the following true [16,18]:
0 < O w < η < I w < 1 ,
where 0 and 1 represent Boolean false and true values that are limiting values for subjective false and subjective true, respectively. The truth value η indicates that it is undecidable whether the proposition A is true or false.
The suggested algebra with uninorm and absorbing norm operations allows direct definition of the control of mobile robots and their swarming. In the next section, we define such a system using the extension of the considered algebra based on the model of a neural network with mobile neurons.

2.2. Algebra of Control Variables

Let us adopt the presented above algebra A to direct the handling of the control variables. For convenience, assume that the control variables obtain their values from the interval [ 1 ,   1 ] . Such an assumption is certainly not necessary, but it meets most cases of the control systems.
Formally it is required to extend algebra A to the interval [ 1 ,   1 ] . The first option is to apply direct and inverse linear transformations
c = 2 x 1 ,       x [ 0 ,   1 ] ,
x = ( c + 1 ) / 2 ,       c [ 1 ,   1 ] .
Then, the values c of control variables are converted to the trusts x and then after appropriate handling the resulting trusts x are converted to the controlled values c .
The second option is to define the extended algebra A * , which inherits the properties of the algebra A , but acts on the interval [ 1 ,   1 ] . Let u * : ( 1 ,   1 ) ( , ) and v * : ( 1 ,   1 ) ( , ) be continuous, strictly monotonously increasing functions with the zeroes u * ( θ * ) = v * ( ϑ * ) = 0 and with the limits lim x 1 u * ( x ) = lim x 1 v * ( x ) = and lim x 1 u * ( x ) = lim x 1 v * ( x ) = + .
Using these functions, the extended uninorm θ * : [ 1 ,   1 ] × [ 1 ,   1 ] [ 1 ,   1 ] and absorbing norm ϑ * : [ 1 ,   1 ] × [ 1 ,   1 ] [ 1 ,   1 ] are defined similarly to the Equations (1) and (2):
x θ * y = u * 1 ( u * ( x ) + u * ( y ) ) ,
x ϑ * y = v * 1 ( v * 1 ( x ) × v * 1 ( y ) ) .
The inverse operations then are
x θ * y = u * 1 ( u * ( x ) u * ( y ) ) ,
x ϑ * y = v * 1 ( v * ( x ) / v * ( y ) ) ,
where v * ( y ) 0 .
Consequently, an extension of the algebra A to the interval [ 1 ,   1 ] is a triple A * = ( [ 1 , 1 ] , θ * , ϑ * ) , in which extended uninorm θ * and absorbing norm ϑ * specify algebraic operations with zeroes θ * and ϑ * , respectively.
Algebra A * inherits the properties of the algebra A , but since it is defined on the interval [ 1 , 1 ] the trusts x [ 1 , 1 ] can be directly considered as control variables c and vice versa. This algebra can be considered as a continuous version of both the balanced ternary numerical system and the ternary logic.
The defined algebra A * can be used for direct definition of the control and learning system. Below we consider the extension of previously defined neural network with mobile neurons. This network is a basis for specification of navigation and swarming of mobile robots.

3. Neural Network with Mobile Neurons in Algebra A *

Control of the mobile robots is defined using the extended algebra presented above and uninorm and absorbing norm aggregators. Such an approach allows direct application of formal operations for specification of the robots’ actions. The swarming process is then defined using the model of neural network with mobile Tsetlin neurons.

3.1. States of the Neurons and of the Synapses

In general, in a neural network with mobile neurons it is assumed that the connectivity between the neurons depends on the distance between the neurons, and then the training process leads to restructuring the network and forming ensembles of the neurons. On the other hand, the distance between the neurons is defined with respect to the connectivity between the neurons; thus, the learning process is governed by the neurons motion [18].
Let us consider a neural network defined in the algebra A * . Formally this network follows the previously defined Tsetlin network acting in algebra A [13], but the use of the interval [ 1 ,   1 ] instead of the interval [ 0 ,   1 ] results in more convenient specification of the neurons’ motion.
Assume that the i th neuron obtains the inputs via l synapses and that at each time t = 0 , 1 , 2 , the input value x k ( t ) appears at the k th synapse, k = 1 , 2 , , l , which is characterized by the weight ω k ( t ) . The weight can be considered as a transition possibility, which is the ability of the synapse to transfer the input value or as a transition trust that is the trust that the synapse transmits the input value. In the considered framework, transition possibility and transition trust have the same meaning and specify the level of connectivity between the neurons. At the same time, as a logical value, the weight ω k ( t ) is an operand of multi-valued logical operations.
In the Tsetlin network defined in A * , each neuron acts as follows. The input value x k ( t ) is aggregated with the weight ω k ( t ) by the uninorm θ * or by the absorbing norm ϑ * and the resulting value is transmitted to the neuron. The neuron compares the obtained value with its internal state s i ( t ) using the absorbing norm ϑ * . Finally, using the uninorm θ * the neuron aggregates the results of such a comparison for all l inputs and specifies the obtained result as an output z i ( t ) , which can also be (or can be not) used as the next internal state s i ( t + 1 ) .
Formally, such activity is defined as follows. Let ( x 1 ( t ) , x 2 ( t ) , , x l ( t ) ) be the values appearing at the synapses of the neuron at time t , s i ( t ) be the neuron’s internal state and ( ω 1 ( t ) , ω 2 ( t ) , , ω l ( t ) ) be the weights of the neuron’s synapses. The output values of synapses k = 1 , 2 , , l are defined using the uninorm θ * (u-synapses)
y k ( t ) = x k ( t ) θ * ω k ( t )
or using the absorbing norm ϑ * (a-synapses)
y k ( t ) = x k ( t ) ϑ * ω k ( t ) .
Then each of these values is compared with the current state of the neuron using the absorbing norm, that is
c k ( t ) = y k ( t ) ϑ * s i ( t ) ,
and, finally, the results of comparisons are aggregated using the uninorm, and the output of the neuron is specified by
z i ( t ) = θ * k = 1 l c k ( t ) .
The next state s i ( t + 1 ) of the neuron is specified by a certain state function, which in the simplest case is
s i ( t + 1 ) = z i ( t ) .
Let us consider the weights ω ( t ) [ 1 , 1 ] of the synapses, which specify the strength of the connection between the neurons. The changes of these values are associated with the learning process. The weights can be defined either externally by a certain training process or internally using the states of the neurons. In the first case, the network converges to definite configuration, which leads to the desired result on the output neurons, while in the second case the network demonstrates certain self-organization [27] and can change its configuration with respect to the states of the neurons and of the environment. In the suggested model, we follow the second approach and define the transition possibilities with respect to the states of the interconnecting neurons.
Let N i and N j be two neurons connected by the synapse S i j . The states of the neurons at time t are denoted by s i ( t ) and s j ( t ) and the weight of the synapse is denoted by ω i j ( t ) . Notice that in the notation of the weight ω k the upper index stands for k th input of the neuron and in the notation ω i j the bottom pair of indices represents the neurons interconnected by the considered synapse. Using the absorbing norm ϑ * , the value ω i j ( t ) is defined as
ω i j ( t ) = s i ( t ) ϑ * s j ( t ) ,
which, in essence, is a result of the comparison between the neurons’ states s i ( t ) and s j ( t ) .
In the network with mobile neurons, the weights ω i j ( t ) are applied for two goals. The first is a conventional use for specification of the output y j ( t ) of the neuron N j (see Equations (15) and (16)), and the second is a definition of the value of the potential function u in the location of the synapse. The second use is possible since in the considered approach the network is implemented in the form of physical devices with definite locations on a plane. Namely, if the neurons N i and N j are associated with the mobile robots located in the points c i and c j , then it is assumed that the synapse is located at the geometrical center c ¯ i j of the line connecting these points. Figure 3 shows an example of locations of three neurons N 1 , N 2 , and N 3 and three synapses S 1 , S 2 , and S 3 connecting these neurons.
In the definition of the potential function, the neurons and, consequently, the robots are considered as obstacles and the potential function in their locations obtains the highest value 1 . In the other points of the plane the values of the potential function are defined by an appropriate smoothing. The potential function for the neurons and the synapses shown in Figure 3 with the neurons’ states s 1 = 1 , s 2 = 1 and s 3 = 1 and the weights ω 13 = ω 12 = 1 and ω 23 = 1 is shown in Figure 4.
It is seen that the locations of the neurons and the location of the third synapse that connects the neurons N 2 and N 3 potential are positive and the neurons (which are the mobile robots) are repulsed from these locations. However, the locations of the synapses that connect the neuron N 1 with the neurons N 2 and N 3 potential are negative, and the neurons (or robots) are attracted to these locations.

3.2. Reactive Learning and Motion of the Neurons

Learning in the considered network is based on the changes of the connectivity between the neurons and on the movements of the neurons with respect to the defined potential function. In addition, it is assumed that the neuron can substitute the synapse, to which location the neuron arrived, which results in creating two new synapses, connecting the arrived neuron with the neurons, which were connected by the substituted synapse. The process of synapse substitution and creation of new synapses is shown in Figure 5.
In the considered scenario, synapse S 12 before substitution had a negative value and attracted the neuron. After division, the synapse S 1 k starts with the values equivalent to the value of synapse S 12 and then changes it to the value calculated by Equation (20) with respect to the states of the neurons N 1 and N k (for synapse S 1 k ) and N k and N 2 (for synapse S k 2 ).
In a similar manner, two neighboring neurons can repulse the neuron with the same sign of the state value. The process, in which the neurons N 1 and N 2 repulse neuron N k , is shown in Figure 6.
Similar to the Coulomb law for the electrically charged particles, in the considered model it is assumed that each neuron attracts the neurons with different signs of the states and repulses the neurons with the same sign of the state. Using the value
a r p ( N i , N j ) = u * 1 ( θ * ( s i ϑ * s j ) ) = u * 1 ( θ * ω i j ) ,
of attraction/repulsion between the neurons N i and N j , where s i and s j are the states of the neurons N i and N j , respectively, the attraction/repulsion force for these neurons is defined by the formula
F ( N i , N j ) = λ   a r p ( N i , N j ) / d i s t ( N i , N j ) ,
where d i s t ( N i , N j ) is a geometrical distance between the neurons and λ is an attraction/repulsion coefficient. In the Equation (21), the value s i ϑ * s j represents similarity between the states, and in the Equation (22) the distance between the neurons is defined by the metric of the space, where the network acts; in the considered case it is Euclidian distance.
Certainly, attraction/repulsion can be defined differently with respect to the needs and requirements of the network. For example, attraction and repulsion can be defined using the known aggregation function [19]
J a g r ( N i , N j ) = d i s t ( N i , N j ) ( a b exp ( d i s t 2 ( N i , N j ) c ) ) ,
where the states of the synapses and of the neurons are expressed by the parameters a , b , and c , or by using different similarity measures.

3.3. Simulation of the Network Activity

To clarify activity of the defined network with mobile neurons let us consider the following example. Assume that the network is in the gridded square domain of the size N × N = 100 × 100 cells and that the number of the neurons in the network is n = 25 .
Each neuron N i , i = 1 , 2 , , n , is connected with the neighboring neurons N j , j = 1 , 2 , n i n 1 , j i , which are located at the distances d i s t ( N i , N j ) bounded by certain predefined threshold d m a x . For the illustrations below we used Euclidian distance and the threshold d m a x = 30 .
If the distance d i s t ( N i , N j ) is less than the threshold d m a x , then the state of the synapse S i j connecting the neurons N i and N j , which is the weight ω i j , is calculated according to Equation (20) with respect to the current states s i ( t ) and s j ( t ) of the neurons N i and N j , correspondingly. Otherwise, if the distance d i s t ( N i , N j ) is greater than the threshold d m a x , then it is assumed that the neurons are not connected and the weight ω i j is set to zero.
The attraction/repulsion force F ( N i , N j ) for the neurons N i and N j is calculated according to Equation (22), where attraction/repulsion coefficient λ is a tenths part of the domain diagonal, that is, λ = 0.1 2 N 2 = 10 2 = 14.14 .
Following the attraction/repulsion forces, the neurons move in the resultant directions by the steps proportional to the values of the resultant attraction/repulsion forces. In the illustrations it is assumed that the lengths δ of the steps are equivalent to the values of the attraction/repulsion forces.
The states s i ( t ) of the neurons are initialized by random such that s i ( 0 ) are drawn from the interval [ 1 ,   1 ] with respect to uniform distribution, and then these values as well as the weights of the connections are updated using the Equations (15)–(19). Examples of the evolution of the network structures are shown in Figure 7 and Figure 8 (the videos of the evolution of the network structures can be found in Supplementary Materials, Videos S1 and S2, respectively). In Figure 7 the starting configuration is a regular lattice and in Figure 8 the starting configuration is random.
It is seen that initially the regular configuration of the network (Figure 7a) is disturbed by the first movements of the neurons (Figure 7b), and this distortion is enough for serious change of the network configuration at the next step (Figure 7c). Figure 7d shows the neuron groups observed after the 100th step of each neuron.
Evolution of the network configuration starting from random configuration is shown in Figure 8.
It is seen that, similar to the previous example, the first motion of the neurons slightly changes the initial configuration of the network (cf. Figure 8a,b). However, already at the third step the configuration of the network changes seriously (Figure 8c). Figure 8d shows the configuration of the network after the 100 th step of each neuron, which demonstrates that the initially connected network is divided to non-connected neurons, which preserve their locations.
In order to clarify the evolution of the network structure, let us consider the following example. Assume that the network includes n = 4 interconnected neurons as it is shown in Figure 9.
Assume that the initial states s i ( 0 ) of the neurons N i , i = 1 , ,   4 , are
s 1 ( 0 ) = 0.75 ,   s 2 ( 0 ) = 0.25 ,   s 3 ( 0 ) = 0.25   and   s 4 ( 0 ) = 0.75 .
Then the weights ω i j ( 0 ) , i , j = 1 , ,   4 , of the connections defined by Equation (20) are defined by the matrix:
ω ( 0 ) = ( 0 0.2435 0.2435 0 0.2435 0.7383 0.0651 0.2435 0.2435 0.0652 0.7383 0.2435 0 0.2435 0.2435 0 ) .
Direct application of the Equations (15) and (17)–(19) results in the following states s i ( 1 ) , i = 1 , ,   4 , at the time t = 1 :
s 1 ( 1 ) = 0.0256 ,   s 2 ( 1 ) = 0.0485 ,   s 3 ( 1 ) = 0.0817   and   s 4 ( 1 ) = 0.9534 .
Then, the weights ω i j ( 1 ) , i , j = 1 , ,   4 , are:
ω ( 1 ) = ( 0 0.0012 0.0012 0 0.0021 0.0478 0.0040 0.0905 0.0021 0.0040 0.0478 0.0905 0 0.1518 0.1518 0 ) .
By the same manner, the states s i ( 1 ) , i = 1 , ,   4 , at the time t = 2 are
s 1 ( 2 ) = 0.0498 ,   s 2 ( 2 ) = 0.0910 ,   s 3 ( 2 ) = 0.1450   and   s 4 ( 2 ) = 0.2477 .
The weights ω i j ( 2 ) , i , j = 1 , ,   4 , at this time are
ω ( 2 ) = ( 0 0.0046 0.0046 0 0.0073 0.0126 0.0133 0.0231 0.0073 0.0133 0.0126 0.0231 0 0.0369 0.0369 0 ) .
Finally, already at the time t = 3 the states become closer to zero
s 1 ( 3 ) = 0.0008 ,   s 2 ( 3 ) = 0.0057 ,   s 3 ( 3 ) = 0.0187   and   s 4 ( 3 ) = 0.0542 ,
and the weights obtain the zero values ω i j ( 3 ) = 0 , i , j = 1 , ,   4 ,
ω ( 3 ) = ( 0 ) .
Thus, the neurons are separate and at the next step, t = 4 , the states also obtain zero values
s 1 ( 4 ) = 0 ,   s 2 ( 4 ) = 0 ,   s 3 ( 4 ) = 0   and   s 4 ( 4 ) = 0 ,
which are preserved infinitely. Since the movements of the neurons are defined by the similarity between the neurons’ states (see Equation (21)), then the motions are zeroed, and the neurons stay at their locations.
The dynamics of the neural network with mobile neurons allows both its usage in traditional machine learning applications, and as a model of swarm activity. In the next section we implement the second possibility and apply the constructed network for specification of the swarm activity.

4. Robot States and Movements

In the suggested approach, activity of the group of mobile robots is modeled using a neural network with mobile neurons. The states of the robots are associated with the states of the neurons, and the neurons’ assembling is associated with the robots’ swarming and creating subgroups in the swarm.
Motion of the neurons can be mapped into the motion of the robots in two ways: either by direct transformation of the neuron’s movements into the movements of the robot or by the use of separate networks for specification of the movements in different dimension. In the first case, the activity if the swarm mimics the activity of the network presented in Section 3.3, while in the second case the swarm dynamics demonstrate additional properties; below we consider the second definition of the swarm activity.

4.1. Robot States under Uncertainty

Consider a mobile robot acting on a plane and assume that the state of the robot is defined by its location and heading. The robot can move one step forward, turn left or right and then move one step forward in a new direction and stay in its current location with current heading.
If the state of the robot is certain, then the next state of the robot is defined by the chosen movement: if the movement is certain, then the next state is also certain and if the movement is uncertain, the next state is uncertain with the same grade of uncertainty. If the state (either the robot’s location or heading or both) is uncertain, then the next state of the robot is also uncertain and for a uncertain movement it includes both uncertainty of the state and of the movement.
In the probabilistic terms, uncertainties of the states and movements are defined by the states’ and transitions’ probabilities. Then the dynamics of the are is governed by the Markov process, which specifies the probabilities of the next robot’s locations and heading. In the terms of subjective trusts, the robot’s dynamics is defined by similar manner, but instead of the Markov process the subjective Markov process [13] is used.
Denote by ξ ( t ) = ( ξ 1 ( t ) , ξ 2 ( t ) ) the vector of the decision-maker’s trusts regarding the robot’s heading at time t . It is assumed that
-
The trust vector ξ ( t ) = ( 1 ,   1 ) means that the heading of the robot is necessary ;
-
The trust vector ξ ( t ) = ( 1 , 1 ) means that the heading of the robot is necessary ;
-
The trust vector ξ ( t ) = ( 1 ,   1 ) means that the heading of the robot is necessary ;
-
The trust vector ξ ( t ) = ( 1 , 1 ) means that the heading of the robot is necessary .
The intermediate values ξ 1 ( t ) , ξ 2 ( t ) ( 1 ,   1 ) specify the levels of the trust that the robot has certain direction.
The trusts of the decision-maker about the turns of the robot are represented by the matrix
T ( t ) = τ 11 ( t ) τ 12 ( t ) τ 21 ( t ) τ 22 ( t ) ,
of the trusts such that
-
The trust matrix T ( t ) = 1 0 0 1 means preserving current direction of the robot with necessity;
-
The trust matrix T ( t ) = 1 0 0 1 means turn left with necessity;
-
The trust matrix T ( t ) = 1 0 0 1 means turn right with necessity.
The intermediate values τ i j ( t ) ( 1 ,   1 ) , i , j = 1 , 2 , specify the levels of the trust that the robot turns in certain direction.
Notice that the suggested notation coincides (up to inambiguous mapping) with the usual definition of headings by the vectors ξ = ( 0 ,   1 ) , ξ = ( 0 , 1 ) , ξ = ( 1 ,   0 ) and ξ = ( 1 ,   0 ) , and of the turns by the rotation matrix
R = cos ( φ ) sin ( φ ) sin ( φ ) cos ( φ ) ,
where φ is a rotation angle such that R = 1 0 0 1 means “preserve the current heading”, R = 0 1 1 0 means “turn left” and R = 0 1 1 0 means “turn right”. In addition, the trust vector ξ ( t ) and the matrix T ( t ) can be considered in the terms of quantum controlled mobile robots and the observations of their states [28] that provides an additional framework for the suggested model.
The turns of the robot in the terms of the trusts are defined by the same manner as its turns using rotation matrix. Namely, if ξ ( t ) = ( ξ 1 ( t ) , ξ 2 ( t ) ) is the trust vector of the robot’s state at time t , then the trust vector ξ ( t + 1 ) = ( ξ 1 ( t + 1 ) , ξ 2 ( t + 1 ) ) of the state at time t + 1 is defined as follows
ξ ( t + 1 ) = ξ ( t ) ϑ * T ( t ) ,
where, similar to the usual multiplication rule of the vector and the matrix,
ξ 1 ( t + 1 ) = ( ξ 1 ( t ) ϑ * τ 11 ( t ) ) θ * ( ξ 2 ( t ) ϑ * τ 21 ( t ) ) , ξ 2 ( t + 1 ) = ( ξ 1 ( t ) ϑ * τ 12 ( t ) ) θ * ( ξ 2 ( t ) ϑ * τ 22 ( t ) ) .
Evolution of the trusts defined by Equation (26) is known as the subjective Markov process [13].
In order to define transition trusts τ i j ( t ) , i , j = 1 , 2 , let us associate the values ξ 1 ( t ) and ξ 2 ( t ) as the states s 1 ( t ) and s 2 ( t ) of the Tsetlin neurons N 1 and N 2 (see Section 3.1), which are connected with themselves and with each other. Then, following the definitions of the neuron’s functions (15)–(19), the transition trusts are defined as
τ 11 ( t ) = θ * k = 1 l ( x k ( t ) θ * ω k 1 ( t ) ) ,   τ 12 ( t ) = x 1 ( t ) θ * ω 12 ( t ) , τ 21 ( t ) = x 2 ( t ) θ * ω 21 ( t ) ,   τ 22 ( t ) = θ * k = 1 l ( x k ( t ) θ * ω k 2 ( t ) ) ,
or as
τ 11 ( t ) = θ * k = 1 l ( x k ( t ) ϑ * ω k 1 ( t ) ) ,   τ 12 ( t ) = x 1 ( t ) ϑ * ω 12 ( t ) , τ 21 ( t ) = x 2 ( t ) ϑ * ω 21 ( t ) ,   τ 22 ( t ) = θ * k = 1 l ( x k ( t ) ϑ * ω k 2 ( t ) ) ,
with respect to the type of synapses—u-synapses (Equation (15)) or a-synapses (Equation (16))—used in the connections between the neurons.
Connections of the neurons are shown in Figure 10.
In the figure, the transition trusts τ i j ( t ) , i , j = 1 , 2 , represent internal controls in the robot and the trusts τ i j ( t ) , i , j = 3 , , k , l , represent connections of the robot with the other robots in the swarm.
Finally, let us consider the mobility of the neurons and its use for definition of the robots’ swarming. Let 1 and 2 be two mobile robots moving on a plane by conducting the indicated above motions.
Assume that the state of each robot is defined by the corresponding trust vector: vector ξ 1 ( t ) = ( ξ 1 1 ( t ) , ξ 2 1 ( t ) ) for the robot 1 and vector ξ 2 ( t ) = ( ξ 1 2 ( t ) , ξ 2 2 ( t ) ) for the robot 2 . As above, let us associate the elements of the trust vectors ξ 1 ( t ) and ξ 2 ( t ) with the states of the neurons. In other words, we assume that the state of each robot is defined by the pair of neurons, that is 1 = ( N 1 1 , N 2 1 ) and 2 = ( N 1 2 , N 2 2 ) , and the states s i j ( t ) of the neurons are associated with the trusts ξ i j ( t ) , i , j = 1 , 2 .
Then, the attraction and repulsion between the robots are the combination of the attraction and repulsion between each pair of the neurons associated with the robots. Using Equation (22), the attraction/repulsion force F ( 1 , 2 ) between the robots consists of four forces
F ( 1 , 2 ) = { F ( N i 1 , N j 2 )   : i , j = 1 ,   2 } .
The scheme of attraction and repulsion of the robots is shown in Figure 11.
It is seen that each robot together with the applied forces can be considered as a dipole. Consequently, the group of the robots forms a dipole dynamical system [24] governed by the rules based on extended uninorm and absorbing norm. Below, we present numerical simulations of the behavior of such system.

4.2. Simulation of the Robots’ Motion and Swarming

Activity of the group of the robots was simulated in a similar setting as the above considered activity of the neural network with mobile neurons. The group includes n = 25 robots i = ( N 1 i , N 2 i ) , i = 1 , 2 , , n , acting in the gridded square domain of the size N × N = 100 × 100 .
Each neuron N 1 i and N 2 i of each robot i is connected with the neurons N 1 j and N 2 j of the neighboring robot j , j = 1 , 2 , , n i n 1 , j i , which are located at the distances less or equal to the threshold d m a x . As above, in the simulations the distances are Euclidian, and the threshold is d m a x = 30 . The value of the attraction/repulsion coefficient λ is also λ = 0.1 2 N 2 = 14.14 .
In the first simulations, the states s 1 i ( t ) and s 2 i ( t ) of the neurons are initialized by random, such that both s 1 i ( t ) and s 2 i ( t ) are drawn from the interval [ 1   1 ] with respect to uniform distribution and then these values as well as the weights of the connections are updated using Equations (15)–(19). Figure 12 shows the locations of the robots starting from the ordered configuration (the video can be found in Supplementary Materials, Video S3). In this figure and the figures below the neuron N 1 i is depicted by a white circle and the neuron N 2 i is depicted by a black circle.
It is seen that similar to the behavior of the neural network with mobile robots (see Figure 7), the initially ordered configuration of the robots’ locations (Figure 12a) is disturbed by the first movements of the robots (Figure 12b), and this distortion increases with the next movement (Figure 12c). Figure 12d shows configuration of the robots’ locations after the   100 th movement of each robot.
Locations of the robots starting from random configuration are shown in Figure 13 (the video can be found in Supplementary Materials, Video S4).
It is seen that the initially dense group of the robots (Figure 13a) diffuses by the first and the second movements of the robots (Figure 13b,c). Figure 13d shows scattered configuration of the robots’ locations after the 100 th movement of each robot.
Notice that in both scenarios the states of both N 1 i and N 2 i in each i th robot, i = 1 , 2 , , 25 , were initialized equivalently by random values from the interval [ 1 ,   1 ] . In other words, the robots were not directed, and this fact is seen in the figures, where the neurons N 1 i and N 2 i attracted and repulsed with no concern to their index.
In the final simulations, in contrast, the states s 1 i ( t ) and s 2 i ( t ) of the neurons are initialized by random, such that s 1 i ( 0 ) are drawn from the interval [ 0 ,   1 ] and s 2 i ( t ) are drawn from the interval [ 1 ,   0 ] with respect to uniform distribution. Thus, in these scenarios the robots are directed in such a manner that the neurons with the states of opposite signs attract and the neurons with the states of the same signs repulse. The states and the weights of the connections are updated using the Equations (15)–(19).
As above, in the simulations the robots started from the locations in ordered configuration and from the locations in random configuration. Figure 14 shows the locations of the robots starting from the ordered configuration (the video can be found in Supplementary Materials, Video S5).
It is seen that the movement of the robots is similar to the movement in the scenario illustrated by Figure 12, with the expected difference in attraction and repulsion of the robots’ neurons with equal and different signs depicted by black and white circles, respectively.
Finally, locations of the robots starting from random configuration are shown in Figure 15 (the video can be found in Supplementary Materials, Video S6).
As it was expected, in this scenario, the motion of the robots is similar to the motion illustrated by Figure 13, with the already-mentioned difference in the repulsion because of the difference between the signs of the neurons’ states.
The simulations demonstrate that starting from arbitrary locations the robots demonstrate variative motion in the group. The states of the neurons in the robots differ from the boundary values; as a result, the connections between the robots are preserved. Together with that, the states and connections of some of the robots converge to the steady state values such that the robots keep their locations.
The observed simulated behavior of the robots follows the behavior of the suggested neural network with mobile neurons, which governs each of the robots’ neurons, which allows consideration of decision making and the actions in the same framework.

5. Discussion

The suggested approach of the control of mobile robots considers decision making and operating as a unified process based on the novel method of handling imperfect information and acting under uncertainty. The method is based on the extension of recently developed algebra with parameterized unary and binary aggregators, which implement multi-valued logical operations.
Formally, the suggested method follows thee general scheme of fuzzy control (see, e.g., [29,30]), but, in contrast to the usual techniques, it is focused on the algebraic model and on the relation between possibilities, probabilities, and necessities. The measures of uncertainty are considered as subjective trusts, which specify both decisions and actions of the system.
Control of the robots’ motion in the group is based on the neural network with the mobile neurons. The idea of such a network with mobile neurons was suggested by Apolloni et al. [18]. In addition to the usual activity of neural network, in this network the mobility of the neurons is applied, and it both defines and is defined by the connectivity between the neurons. In the suggested model the network consists of the Tsetlin neurons, and connections between the neurons and their mobility are governed by the states of the neurons. The activity of the network is defined in the developed algebra of the subjective trusts.
From a control point of view, each robot is specified as a dipole [24], which originally was used by Tchieu et al. in the model of behavior of fish flocks. In the suggested model, the robot consists of two mobile neurons acting in the neural network with obvious limitation led by the size of the robot. Such a model allows to consider the control of the robot by the same techniques as the control of quantum controlled mobile robots [6,18], but with stressing the difference in the topology of the actions’ space.
Since the suggested model applies subjective measures of uncertainty, control is considered from the internal point of view and follows the topology of a non-oriented surface. Such topology most likely differs from the topology of the controls used in the performance centered model by Saeidi [22,23] and of the controls in the model of quantum control [28,31]; these issues will be addressed in future studies.
Finally, the paper does not address teleologic issues or specific applications and can be considered as a proof of concept. Numerical simulations have demonstrated the feasibility of the suggested method. At the next steps, the suggested method is planned to be implemented in the search and foraging algorithms. In the simple case of following toward a single target, the exact values of the trusts can be defined using the recently calculated probabilities of altruistic and egoistic steps of the agents [11], while in more complicated scenarios, additional efforts and comparisons of the suggested method with the known techniques will be required.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/e24060790/s1, Videos S1–S6.

Author Contributions

Conceptualization, E.K. and A.R.; methodology, E.K. and A.R.; software, E.K.; validation, E.K. and A.R.; writing—original draft preparation, E.K.; writing—review and editing, A.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maxwell, J.C. On governors. Proc. R. Soc. Lond. 1868, 16, 270–283. [Google Scholar]
  2. Aliev, R.; Huseynov, O. Decision Theory with Imperfect Information; World Scientific: Singapore, 2014. [Google Scholar]
  3. Feldbaum, A.A. Theory of Optimal Automated Systems; PhysMatGiz: Moscow, Russia, 1963. [Google Scholar]
  4. Aoki, M. Optimization of Stochastic Systems; Academic Press: New York, NY, USA; London, UK, 1967. [Google Scholar]
  5. Åström, K.J. Introduction to Stochastic Control Theory; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  6. Bertsekas, D.P.; Shreve, S.E. Stochastic Optimal Control: The Discrete Time Case; Academic Press: New York, NY, USA, 1978. [Google Scholar]
  7. Dubois, D.; Prade, H. Possibility Theory; Plenum: New York, NY, USA, 1988. [Google Scholar]
  8. Zadeh, L.A. Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets Syst. 1978, 1, 3–28. [Google Scholar] [CrossRef]
  9. Kahneman, D.; Tversky, A. Prospect theory: An analysis of decision under risk. Econometrica 1979, 47, 263–292. [Google Scholar] [CrossRef] [Green Version]
  10. Ruggeri, K.; Ali, S.; Berge, M.L.; Bertoldo, G.; Bjørndal, L.D.; Cortijos-Bernabeu, A.; Davison, C.; Demić, E.; Esteban-Serna, C.; Friedemann, M.; et al. Replicating patterns of prospect theory for decision under risk. Nat. Hum. Behav. 2020, 4, 622–633. [Google Scholar] [CrossRef] [PubMed]
  11. Hassoun, M.; Kagan, E. On the right combination of altruism and randomness in the motion of homogeneous distributed autonomous agents. Nat. Comput. 2021. [Google Scholar] [CrossRef]
  12. Kagan, E.; Rybalov, A. Subjective trusts and prospects: Some practical remarks on decision making with imperfect information. Oper. Res. Forum 2022, 3, 19. [Google Scholar] [CrossRef]
  13. Kagan, E.; Rybalov, A.; Yager, R. Sum of certainties with the product of reasons: Neural network with fuzzy aggregators. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2022, 30, 1–18. [Google Scholar] [CrossRef]
  14. Yager, R.; Rybalov, A. Uninorm aggregation operators. Fuzzy Sets Syst. 1996, 80, 111–120. [Google Scholar] [CrossRef]
  15. Batyrshin, I.; Kaynak, O.; Rudas, I. Fuzzy modeling based on generalized conjunction operations. IEEE Trans. Fuzzy Syst. 2002, 10, 678–683. [Google Scholar] [CrossRef]
  16. Fodor, J.; Rudas, I.; Bede, B. Uninorms and absorbing norms with applications to image processing. In Proceedings of the Information Conference SISY, 4th Serbian-Hungarian Joint Symposium on Intelligent Systems, Subotica, Serbia, 29–30 September 2006; pp. 59–72. [Google Scholar]
  17. Kagan, E.; Rybalov, A.; Siegelmann, H.; Yager, R. Probability-generated aggregators. Int. J. Intell. Syst. 2013, 28, 709–727. [Google Scholar] [CrossRef]
  18. Apolloni, B.; Bassis, S.; Valerio, L. Training a network of mobile neurons. In Proceedings of the International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 1683–1691. [Google Scholar]
  19. Gazi, V.; Passino, K.M. Swarm Stability and Optimization; Springer: Berlin, Germany, 2011. [Google Scholar]
  20. Kagan, E.; Shvalb, N.; Ben-Gal, I. (Eds.) Autonomous Mobile Robots and Multi-Robot Systems: Motion-Planning, Communication, and Swarming; John Wiley & Sons: Chichester, UK, 2019. [Google Scholar]
  21. Fregnac, Y. Hebbian cell ensembles. In Encyclopedia of Cognitive Science; Nature Publishing Group: Berlin, Germany, 2003; pp. 320–329. [Google Scholar]
  22. Saeidi, H. Trust-Based Control of (Semi) Autonomous Mobile Robotic Systems. Ph.D. Thesis, Clemson University, Clemson, SC, USA, 2016. [Google Scholar]
  23. Saeidi, H.; Wang, Y. Incorporating trust and self-confidence analysis in the guidance and control of (semi) autonomous mobile robotic systems. IEEE Robot. Autom. Lett. 2018, 4, 239–246. [Google Scholar] [CrossRef]
  24. Tchieu, A.A.; Kanso, E.; Newton, P.K. The finite-dipole dynamical system. Proc. R. Soc. A 2012, 468, 3006–3026. [Google Scholar] [CrossRef]
  25. Fodor, J.; Yager, R.; Rybalov, A. Structure of uninorms. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 1997, 5, 411–427. [Google Scholar] [CrossRef]
  26. Klement, E.; Mesiar, R.; Pap, E. Triangular Norms; Kluwer Academic Publishers: Dordrecht, The Netherlands, 2000. [Google Scholar]
  27. Hopfield, J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Kagan, E.; Ben-Gal, I. Navigation of quantum-controlled mobile robots. In Recent Advances in Mobile Robotics; InTech: Rijeka, Czech Republic, 2011; pp. 311–326. [Google Scholar]
  29. Jantzen, J. Foundations of Fuzzy Control; John Wiley & Sons: Chichester, UK, 2013. [Google Scholar]
  30. Passino, K.M.; Yurkovich, S. Fuzzy Control; Addison-Wesley: Menlo Park, CA, USA, 1998. [Google Scholar]
  31. Benioff, P. Quantum Robots and Environments. Phys. Rev. A 1998, 58, 893–904. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Generator function (a) and inverse generator function (b).
Figure 1. Generator function (a) and inverse generator function (b).
Entropy 24 00790 g001
Figure 2. Interpretation of the trusts’ values.
Figure 2. Interpretation of the trusts’ values.
Entropy 24 00790 g002
Figure 3. Locations of three neurons N 1 , N 2 , and N 3 and three synapses S 1 , S 2 , and S 3 connecting the neurons.
Figure 3. Locations of three neurons N 1 , N 2 , and N 3 and three synapses S 1 , S 2 , and S 3 connecting the neurons.
Entropy 24 00790 g003
Figure 4. Potential function for the neurons and the synapses shown in Figure 3. The states of the neurons are s 1 = 1 , s 2 = 1 and s 3 = 1 and the weights of the connections (the states of the synapses) are ω 13 = ω 12 = 1 and ω 23 = 1 .
Figure 4. Potential function for the neurons and the synapses shown in Figure 3. The states of the neurons are s 1 = 1 , s 2 = 1 and s 3 = 1 and the weights of the connections (the states of the synapses) are ω 13 = ω 12 = 1 and ω 23 = 1 .
Entropy 24 00790 g004
Figure 5. The process of synapse substitution and creation of new synapses: (a) neuron N k , which is connected with the neurons N i and N j via the synapses S i k and S k j , moves toward the synapse S 12 , which connects the neurons N 1 and N 2 ; (b) the synapse S 12 is “divided” into two synapses; (c) neuron N k substitutes synapse S 12 in its location and connects with the neurons N 1 and N 2 via new synapses S 1 k and S k 2 .
Figure 5. The process of synapse substitution and creation of new synapses: (a) neuron N k , which is connected with the neurons N i and N j via the synapses S i k and S k j , moves toward the synapse S 12 , which connects the neurons N 1 and N 2 ; (b) the synapse S 12 is “divided” into two synapses; (c) neuron N k substitutes synapse S 12 in its location and connects with the neurons N 1 and N 2 via new synapses S 1 k and S k 2 .
Entropy 24 00790 g005
Figure 6. The process of repulsion of the neuron and union of the synapse: (a) neurons N 1 and N 2 repulse neuron N k ; (b) synapses S 1 k and S k 2 move toward each other; (c) synapses S 1 k and S k 2 are united into the synapse S 12 .
Figure 6. The process of repulsion of the neuron and union of the synapse: (a) neurons N 1 and N 2 repulse neuron N k ; (b) synapses S 1 k and S k 2 move toward each other; (c) synapses S 1 k and S k 2 are united into the synapse S 12 .
Entropy 24 00790 g006aEntropy 24 00790 g006b
Figure 7. Evolution of the network structure: (a) initial regular configuration of the network, t = 0 ; (b) configuration after the first movement of each neuron, t = 1 ; (c) configuration after two movements of each neuron, t = 2 ; and (d) configuration after hundred movements of each neuron, t = 100 .
Figure 7. Evolution of the network structure: (a) initial regular configuration of the network, t = 0 ; (b) configuration after the first movement of each neuron, t = 1 ; (c) configuration after two movements of each neuron, t = 2 ; and (d) configuration after hundred movements of each neuron, t = 100 .
Entropy 24 00790 g007
Figure 8. Evolution of the network structure: (a) initial random configuration of the network, t = 0 ; (b) configuration after the first movement of each neuron, t = 1 ; (c) configuration after two movements of each neuron, t = 2 ; and (d) configuration after hundred movements of each neuron, t = 100 .
Figure 8. Evolution of the network structure: (a) initial random configuration of the network, t = 0 ; (b) configuration after the first movement of each neuron, t = 1 ; (c) configuration after two movements of each neuron, t = 2 ; and (d) configuration after hundred movements of each neuron, t = 100 .
Entropy 24 00790 g008
Figure 9. The network with four interconnected neurons.
Figure 9. The network with four interconnected neurons.
Entropy 24 00790 g009
Figure 10. Connection of the neurons that represent the state of the robot.
Figure 10. Connection of the neurons that represent the state of the robot.
Entropy 24 00790 g010
Figure 11. Attraction/repulsion forces between the robots. The bold arrows indicate the heading of the robots.
Figure 11. Attraction/repulsion forces between the robots. The bold arrows indicate the heading of the robots.
Entropy 24 00790 g011
Figure 12. Locations of the robots starting from the position in ordered configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Figure 12. Locations of the robots starting from the position in ordered configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Entropy 24 00790 g012
Figure 13. Locations of the robots starting from the position in random configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Figure 13. Locations of the robots starting from the position in random configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Entropy 24 00790 g013
Figure 14. Locations of the directed robots starting from the position in ordered configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Figure 14. Locations of the directed robots starting from the position in ordered configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Entropy 24 00790 g014
Figure 15. Locations of the directed robots starting from the position in random configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Figure 15. Locations of the directed robots starting from the position in random configuration: (a) initial locations, t = 0 ; (b) locations after the first movement of each robot, t = 1 ; (c) locations after the second movement of each robot, t = 2 ; and (d) locations after the 100th movement of each robot, t = 100 .
Entropy 24 00790 g015aEntropy 24 00790 g015b
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kagan, E.; Rybalov, A. Subjective Trusts for the Control of Mobile Robots under Uncertainty. Entropy 2022, 24, 790. https://doi.org/10.3390/e24060790

AMA Style

Kagan E, Rybalov A. Subjective Trusts for the Control of Mobile Robots under Uncertainty. Entropy. 2022; 24(6):790. https://doi.org/10.3390/e24060790

Chicago/Turabian Style

Kagan, Eugene, and Alexander Rybalov. 2022. "Subjective Trusts for the Control of Mobile Robots under Uncertainty" Entropy 24, no. 6: 790. https://doi.org/10.3390/e24060790

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop