Next Article in Journal
Multibody Model with Foot-Deformation Approach for Estimating Ground Reaction Forces and Moments and Joint Torques during Level Walking through Optical Motion Capture without Optimization Techniques
Previous Article in Journal
A Real-Time License Plate Detection and Recognition Model in Unconstrained Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cognitive Control Architecture for the Practical Realization of UAV Collision Avoidance

1
Aviation Engineering School, Air Force Engineering University, Xi’an 710038, China
2
Unit 93535 of PLA, Rikaze 857060, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2790; https://doi.org/10.3390/s24092790
Submission received: 26 March 2024 / Revised: 20 April 2024 / Accepted: 24 April 2024 / Published: 27 April 2024

Abstract

:
A highly intelligent system often draws lessons from the unique abilities of humans. Current humanlike models, however, mainly focus on biological behavior, and the brain functions of humans are often overlooked. By drawing inspiration from brain science, this article shows how aspects of brain processing such as sensing, preprocessing, cognition, obstacle learning, behavior, strategy learning, pre-action, and action can be melded together in a coherent manner with cognitive control architecture. This work is based on the notion that the anti-collision response is activated in sequence, which starts from obstacle sensing to action. In the process of collision avoidance, cognition and learning modules continuously control the UAV’s repertoire. Furthermore, simulated and experimental results show that the proposed architecture is effective and feasible.

1. Introduction

One of the key hallmark requirements of unmanned aerial vehicles (UAVs) is the ability to avoid obstacles autonomously [1,2,3,4]. Many scholars have studied automatic collision avoidance, with the aim of pushing forward the development of UAVs. Beard [5] constructed Voronoi diagrams to avoid known obstacles. But Voronoi’s method overlooked the obstacles’ effective range, as S. L. Wang [6] pointed out. Wang further developed Voronoi diagrams into Laguerre diagrams to overcome this drawback [7]. However, diagram-based methods cannot deal with pop-up obstacles [8], since those methods require all the obstacles’ information, which is almost impossible to apply in a dynamic environment. Also, many experts are engaged in making UAVs more autonomous. They have proposed many approaches in the literature, such as A-Star and D-Star [9,10,11], potential fields [12], model predictive control [13], and evolutionary algorithmic techniques [14,15].
Although many efforts have been made in collision avoidance, UAVs’ anti-collision capability remains a difficult problem for two reasons: (1) The traditional methods assume that the obstacles are few and far between each other. (2) The time consumed is limitless. So, the traditional methods may not be practical for real-time collision avoidance, but this may be solved by referring to advances in brain-inspired cognitive systems [16].
The ability to avoid obstacles quickly is easy for humans but difficult for UAVs [17]. Douglas Hofstadter envisioned this when he said that “for any program to handle letterforms with the flexibility that human beings do, it would have to possess full-scale artificial intelligence” [18]. Many researchers have conjectured that collision avoidance could be achieved by incorporating insights from brain and cognitive science research.
One may avoid crowded obstacles without any thinking. We are not born with avoidance skills, but through explicit and repetitive training, we can develop what is called an acquired conditioned reflex (CR) [19,20,21]. People can learn the skill through special training.
The foundation of artificial intelligence (AI) lies in its capacity for self-learning, a concept. This self-learning capability is not just a theoretical construct but a fundamental requirement for the development of autonomous systems. J. Weng [22], in 2001, delineated an autonomous learning mechanism that is pivotal in the construction of intelligent robots capable of functioning with a degree of autonomy [23,24]. This mechanism is designed to enable robots to learn from their interactions with the environment, thereby improving their performance over time. M. S. Lim [25] contributed to the field by proposing a rapid and efficient method for synthesizing humanoid robot hands that can adeptly grasp planar polygonal objects. This innovation is a step toward more dexterous and versatile robotic manipulation. M. Asada [26] developed a synthetic approach for cognitive robots, which was a significant step in endowing machines with cognitive abilities that were once the exclusive domain of biological systems. The approaches mentioned in [27] further build upon these ideas, expanding the scope of AI in robotics. Behavior-based control architectures, as suggested by Simmons [28], Rosenblatt, and Thorpe [29], draw inspiration from Brooks’ [30] behavior-based subsumption architecture. These architectures are designed to facilitate complex behaviors in autonomous mobile robots through a decentralized control system that allows for the integration of various behavioral modules. In 2007, Bermudez and Verschure [21] proposed a bio-inspired UAV system, marking a significant attempt to integrate biological intelligence into the automatic control of UAVs. This system aims to leverage principles from biology to enhance the adaptability, efficiency, and autonomy of UAVs, particularly in dynamic and challenging environments. The integration of these cognitive and learning principles into UAV systems is crucial for creating machines that can operate with a higher level of autonomy and intelligence. It is through this lens of cognitive learning and adaptive behavior that the future of UAV technology is being shaped, and this has the potential to revolutionize the way we approach robotics and AI.
In this paper, we introduce a cognitive control architecture for UAVs that is inspired by the human brain’s conditioned-reflex mechanisms. Our approach significantly enhances the real-time collision avoidance capabilities of UAVs by reducing their computational complexity and response time. The main contributions of this paper include the development of an innovative learning-based strategy that allows UAVs to adapt to dynamic environments and avoid obstacles more efficiently. We have conducted extensive simulations and experiments to validate the effectiveness of our proposed system, which demonstrates a substantial improvement over traditional methods in terms of both safety and operational efficiency.

2. A Cognitive Control Architecture for the Conditioned Reflex Cycle in the Anti-Collision Behavior of UAVs

2.1. The Brain Pathway of the Conditioned Reflex Cycle in Human Anti-Collision Behavior

Humans have a fundamental function, which is to find a safe path in crowded environments based on the CR. The classical CR refers to a learning process in which a potent stimulus (e.g., obstacles) is paired with a previously neutral stimulus (e.g., anti-collision movements); its neural substrates are now beginning to be understood [31,32]. The more familiar the crowded environment is, the faster and safer will be the path obtained. So, the CR-based method usually takes less time than normal ways.
For complex problems, humans can find optimal solutions through cognition and learning [33,34]. Infants may not know how to avoid obstacles in the complex environment, but they can cognize the obstacles and learn how to deal with obstacles as they grow up. This process is regarded as a growth of the brain, especially the parts concerning anti-collision behavior. There are three important parts when humans are performing collision avoidance, which are the cortex, the thalamus and the basal ganglia. In the mammalian brain, feedback connections between brain areas play important roles in collision avoidance [35]. Rivara [36] came up with the study that suggested that the cerebral cortex has many functions, including movement planning, behavior control, and motion execution.
The thalamus has multiple functions, and it is generally believed to act as a relay station or hub, relaying information between different subcortical areas and the cerebral cortex [37]. The thalamus receives signals from cortex and produces awareness of the obstacles [38].
Popular theories implicate the basal ganglia primarily in action selection—in helping to decide which of several possible behaviors to execute [39]. In more specific terms, the basal ganglia’s primary function is likely to control and regulate activities of the motor and premotor cortical areas so that voluntary movements can be performed smoothly [40]. In conclusion, the basal ganglia is strongly interconnected with the cortex and thalamus, and its function is to control movements [41]. An anti-collision mechanism in humans can be derived based on the studies of brain science, as shown in Figure 1.
In Figure 1, the anti-collision of normal method and CR method is presented with black flow and red flow respectively.
To normal anti-collision reflex (see in Figure 1, the black flow), there are 8 procedures in fulfilling complete anti-collision. (1). Sense. Human must have a way, see or touch, to sense the environment. (2). Preprocess. The prefrontal cortex receives and synthesizes information from the outside world [42]. The perceived information is then sent to frontal cortex, functioning on transforming original visual information to classified and reduced signals [43]. This preprocessing towards information makes subsequent steps more effective. (3). Cognize. In mammalian brain anatomy, the thalamus is the large mass of gray matter in the dorsal part of the diencephalon of the brain with several functions such as relaying of sensory signals, including motor signals, to the cerebral cortex, and the regulation of consciousness and alertness [38]. The function of thalamus, regarding anti-collision, is to classify obstacles and to have a consciousness of familiar environment, that is to cognize obstacles. (4). Obstacles learning. The cortex, in medicine, is the cerebral cortex which covers the brain, purposing on learning and reserving knowledge, and mapping the knowledge between each other [44], in other word, to develop brain through learning and to think a way out in crowded obstacles. (5). Behave. The basal ganglia is in the brains of vertebrates including humans, which is situated at the base of the forebrain. The basal ganglia is associated with a variety of functions including: control of voluntary motor movements, procedural learning, routine behaviors and emotion [41]. Its function is to design anti-collision strategy and behave safely in the environment. (6). Strategies learning. The function of frontal cortex is to memorize anti-collision strategies, and match them with obstacles. (7). Pre-act. The anti-collision strategies cannot be decoded and acted by limbs, so they should be clarified so as to transform the strategies to body language. (8). Act. The limbs follow the commands from brain to avoid crowded obstacles.
To CR-based anti-collision (see in Figure 1, the red flow): Conditioned reflex occurs when a conditioned stimulus is paired with an unconditioned stimulus. After the pairing is repeated, the organism exhibits a conditioned response to the conditioned stimulus when the conditioned stimulus is presented along, and it must be acquired through experience. According to Bouton [45], conditioned reflex does not involve the acquisition of any new behavior, but rather the tendency to respond in old ways to new stimuli. In conditioned reflex anti-collision, there are only two processes, which are sense (9) and act (10). Humans can avoid obstacles by body language through training and experience. In a crowded environment, humans do not have any time to go through the normal anti-collision process. To make themselves safe, they can only move without thinking. But the conditioned reflex is based on normal anti-collision behavior.

2.2. Cognitive Control Architecture for UAVs

In collision avoidance, traditional UAVs only have two levels (a guidance level and a motion level) to plan a safe path. However, the guiding algorithms always stay the same. Even in the most familiar environment, the UAV must calculate every path from scratch.
In a cognitive UAV, there are three levels (see Figure 2; cognition level, guidance level, and motion level). The flight control strategy is executed by the guidance level, which generates control variables of the UAVs. The motion level responds to the guidance level, which transforms the control variables into throttle commands, which regulate the rotational speed of each motor, so as to control the altitude and attitude. What differs from traditional UAVs is the cognition level, which makes a UAV learn what obstacles are and how to avoid obstacles without much thinking in familiar environments.
As shown in Figure 2, information about obstacles is transferred to the cognition controller so as to produce avoidance strategies. The strategies are sent to the guidance controller to calculate and generate flight instructions. Finally, the motion controller decodes flight instructions into flying motions that directly control a UAV. In this paper, we assume that a cognitive UAV can decode avoidance strategies and obtain flying motions by itself.

2.3. The Brain-Like Conditioned Reflex Cycle of Cognitive UAVs

Building models with reference to the mammalian brain is an important area of study for UAVs [46,47]. Collision avoidance in a complex environment has met with tremendous difficulties for it requires the following cognitive characters:
  • Universality: Cognitive algorithms design a universal schema rather than preprogramming and reprogramming for special missions, so they can enrich themselves incrementally.
  • Interactivity: Humans act as teachers rather than programmers. We can only influence the process and the context but are definitely not the decision-makers.
  • Learnability: Cognition is a gradual and cumulative process. Advanced knowledge relies on basic ability.
We introduce a hierarchical model called the recursive brain network (RBN) that incorporates brain-science studies to fulfill the cognitive demands. In RBN, the anti-collision process can be described as shown in Figure 3.
In Figure 3, the intricate flow processes that govern anti-collision systems in UAVs differ from those in traditional anti-collision methods and those inspired by conditioned reflexes. The black curves delineate the multi-step normal anti-collision procedures, which consist of eight distinct processes. These processes are designed to systematically guide the UAV through obstacle detection, analysis, and evasion, ensuring safe navigation. As indicated in the figure, each process is crucial for the UAV’s cognitive control architecture, starting from the initial sensing of the environment to the final execution of avoidance maneuvers.
The first process, ”Sensing“, involves the utilization of the UAV’s sensors to detect and identify obstacles in the vicinity. This is followed by ”Preprocessing”, where the raw sensor data is refined and organized into a usable format. ”Cognition” then interprets this information, allowing the UAV to understand and classify the obstacles within its operational space. ”Obstacle learning” is where the UAV retains and recalls information about the environment, creating a knowledge base for future navigation.
The subsequent processes, ”Behavior” and ”Strategy learning”, are responsible for developing and implementing collision-avoidance tactics. ”Pre-action” serves as the preparatory stage, where the UAV formulates a response based on its learned strategies, and ”Action” is the culmination of these processes, executing the avoidance actions with precision. These eight processes, as detailed in Table 1, form a coherent and comprehensive approach to avoiding collision, ensuring that the UAV can navigate autonomously and safely in dynamic environments.
In conditioned-reflex anti-collision (see Figure 3; the red flow), there are only two main processes, which are sensing (9) and action (10). Cognitive UAVs, like humans, can avoid obstacles directly without repetitive calculations via a conditioned reflex after training.
Figure 4 of our manuscript illustrates the detailed interplay between the trust region, flight path, and waypoints, which are crucial components in the cognitive control architecture for UAVs. The trust region is a spatial area created around the UAV, delineated by the boundaries of detected obstacles, within which the UAV is deemed to operate safely. This region forms the foundation for all subsequent navigation decisions. The flight path is derived from the trust region through a series of complex calculations that consider the UAV’s current position, the geometry of the trust region, and the intended destination. These calculations are performed by sophisticated algorithms designed to generate a trajectory that maximizes safety and efficiency. However, since the flight path is an output of algorithmic processing, there might be instances where the path includes maneuvers, such as sharp turns or abrupt changes in altitude, that exceed the UAV’s performance capabilities. For example, the flight path may include right angles that the UAV cannot physically achieve without risking a stall or control loss.
To address this, the calculated flight path undergoes further refinement to ensure that it is applicable to the UAV’s actual flight dynamics. This involves smoothing out the path to remove any extreme angles and adjusting the trajectory to fit within the UAV’s operational limitations. The end result is a flight path that is not only safe but also feasible for the UAV to follow. Once the flight path is finalized, it must be translated into a series of waypoints. These waypoints are specific points along the flight path that the UAV will navigate to, one after another, to reach its destination. The conversion of the flight path into waypoints is based on the UAV’s sensing, processing, and actuation capabilities, ensuring that each waypoint is reachable and executable.
In the graphical representation of Figure 4, filled and empty circular nodes symbolize features and pools, respectively. Features represent distinct characteristics of the environment, such as the location of obstacles, while pools are collections of possible flight paths that can be taken from one feature to another. Each pool is a composite of factors from its level and serves as a hub for the subsequent level’s features. The lateral constraints, depicted as grey rectangles, play a critical role in this model. They act as coordinators, dictating the selection among the pools they are connected to, reflecting the sequential nature of the flight-path construction process.
This comprehensive approach ensures that the UAV’s flight path is not only a product of calculations based on the trust region but also a reflection of the UAV’s cognitive understanding of its environment, resulting in a flight strategy that is both intelligent and adaptable.
Upon examining the recursive brain network (RBN) model in conjunction with the mechanisms of the human brain, a striking congruence is observed. This alignment suggests that a brain-inspired anti-collision mechanism for unmanned aerial vehicles (UAVs) can be effectively engineered to mimic cognitive processes. As depicted in Figure 5, the RBN model is divided into distinct regions that correspond to various functional areas of the brain. The blue region symbolizes the cortex function, which is responsible for higher-order cognitive processes such as planning and decision-making. The yellow region represents the thalamus function, which acts as a crucial relay center that processes and integrates sensory information before it reaches the cortex. Lastly, the purple region signifies the basal ganglia function, which plays a pivotal role in selecting and initiating appropriate actions or motor responses.
This tri-colored representation in Figure 5 not only underscores the complexity of the brain’s anti-collision response but also provides a structured framework for designing UAV systems that can exhibit cognitive-like behavior. The cortex, thalamus, and basal ganglia work in concert to enable quick and informed decisions regarding obstacle avoidance, much like a human’s innate ability to navigate through cluttered spaces without explicit calculation. By emulating these brain functions, the RBN model offers a sophisticated approach to enhancing UAV autonomy, particularly in high-stake scenarios in which rapid and accurate collision avoidance is paramount.

3. Algorithmic Design of Cognitive UAVs

There are two main parts in fulfilling an algorithmic design of a cognitive UAV. Firstly, we should build the mapping relation between obstacles and avoidance strategies (see Figure 5; the blue part), which is the learning part of cognitive UAVs. Then the corresponding obstacle pattern and avoidance strategy should be constructed so as to cognize obstacles and strategies (see Figure 5; the yellow and purple parts).

3.1. Construction of the Obstacle-Avoidance Mapping Relation

Obstacle development has a one-to-one correspondence at the beginning, which means that there is only one avoidance strategy A S l corresponding to each obstacle pattern O P k . However, there may be a similar obstacle pattern in the memory of the cortex region, which can replace the obstacle pattern O P that the UAV is currently facing, which gradually accumulates in obstacle learning. Equation (1) shows a mapping relation f that outputs a safe path S in the anti-collision zone Ω based on the surrounding environment.
S t + 1 = f S t , O t , O P t , A S t , O Ω , S Ω
An anti-collision path can be obtained via obstacle pattern mapping, but it does not mean that the obstacle-avoidance mapping relation can be constructed. Consider two obstacle patterns: O P and O P 3 (see Figure 6). The avoidance strategy A S 1 represents an anti-collision path; nevertheless, the optimal mapping relation O P 3 - A S 1 is not optimal for the obstacle pattern O P .
Unlike a normal obstacle-avoidance mapping structure, cognitive mapping relation is an incremental learning schema. It can not only output avoidance strategies online, but it also improves obstacle-avoidance mapping relations offline through Δ O t . So, Equation (1), considering incremental learning, can be improved as:
S t + 1 = f S t , O t , Δ O t , O P t , A S t
The improved structure can be described as in Figure 7. The final goal is to avoid obstacles immediately online and to find optimal paths offline. The basic algorithm is an artificial potential method, which is an algorithmic approach for collision avoidance that assigns potential values to different positions in the environment based on their proximity to obstacles. The core idea is to create a field where the UAV experiences a force that is inversely proportional to the distance from the obstacle. This force acts as a repulsive effect that guides the UAV away from obstacles. By calculating the gradient of the potential field, the UAV can determine the direction of the safest path, which is typically the path of least resistance or of lowest potential value. This method is efficient because it allows for real-time navigation decisions without the need for complex computations, making it suitable for dynamic environments where quick responses are necessary.

3.2. Construction of an Avoidance Strategy and an Obstacle Pattern

Regardless of a quad-rotor’s rotation around the center of mass, it can be described as a controllable mass. The gridding model of a UAV is constructed as shown in Figure 8, in which a UAV has eight flyable directions.
An avoidance strategy can be obtained via the sequential combination of flying actions:
A = a 1 , a 2 , , a n a i A c t i o n s
Avoidance-strategy cognition is the combination of well-organized avoidance strategies:
A B = A S 1 , A S 2 , , A S k
Suppose the k -th obstacle’s position ρ k , ϕ k and velocity v k , θ k are given in a polar coordinate system, so the definition of the k -th obstacle can be described as
o b k = ρ k , ϕ k , v k , θ k
In order to illustrate the computational process, Equation (5) is simplified to Equation (6), where A 1 to A 4 means ρ k , ϕ k , v k , θ k , respectively.
o b = A 1 , A 2 , A 3 , A 4
For the determined value x 0 of the q -th property, a subordinating degree function A q p x 0 can be obtained by the maximum-membership-degree principle. The fuzzy sets corresponding to x 0 are defined in Equation (7), where x 0 A q p   ( 1 p N ) .
A q p x 0 = q = 1 N A q p x 0
The osculating function σ i is introduced to compare whether obstacle k and obstacle l belong to same obstacle pattern.
Supposing that σ i is based on the i -th property,
σ i A i k , A i l = 1 2 A i k A i l + 1 A i k × A i l
where A i k and A i l mean the fuzzy sets based on the i -th property of obstacle k and l , respectively, and A i k A i l and A i k × A i l are the inner and outer products.
There are many obstacles in an environment, so we introduce an obstacle pattern to memorize obstacles as O P = o b 1 , o b 2 , , o b k , and the obstacle pattern cognition is a matrix set as O B = O P 1 , O P 2 , , O P k .
Suppose that the obstacle pattern around UAV is P = o b 1 , o b 2 , , o b m , and the comparing obstacle pattern knowledge now is Q = o b 1 , o b 2 , , o b n . To ensure that obstacles in Q can completely replace P , every obstacle ( o b k ) in P should find a high closed-degree function with o b l in Q , which means that o b k P , i i = 1 , 2 , 3 , 4 , o b l Q :
σ i A i k , A i l c i
where c i is the threshold value of the closed degree for the i -th property.
The similarity degree function s i m P , Q can be designed by assigning a different weight ω i to the corresponding obstacle property:
s i m P , Q = i = 1 4 σ i A i k , A i l × ω i
For the purpose of selecting the most similar obstacle pattern Q 0 , the obstacle pattern P is compared with all the obstacle patterns so that the optimal similar mapping pattern Q 0 can be chosen:
i f max Q k T B s i m P , Q k   ,   Q 0 = Q k
Finally, the knowledge incremental function f a d d is constructed so as to judge whether or not to add a new obstacle-avoidance mapping relation to the memories:
f a d d = 1   ,   i f   s i m P , Q c s i m 0   ,   i f   s i m P , Q < c s i m
where c s i m is the threshold of similarity function; the new obstacle pattern can be described and represented if s i m P , Q c s i m .

4. Simulation and Analysis

To compare the effectiveness and capability of our method, two representative algorithms were introduced (ant colony (AC) algorithm [48] and cognitive game (CG) algorithm [49]).
The AC algorithm is a simulated evolutionary algorithm derived from the biological world. It has the characteristics of distributed computing, positive feedback of information, and heuristic searching. It is one of the most effective algorithms for solving the problem of UAV path planning.
On the basis of the autonomous anti-collision responses of UAVs in a non-isolated airspace, the cognitive game algorithm transforms the collision-avoidance problem into a game model. By establishing a kinematics model of UAVs and obstacles, the CG algorithm puts forward a solution method for safe path planning.
All simulations were run under the Matlab 12.0, 3.3 GHz Inter Core i3, Windows XP system. Assume that the UAV’s detection range is 10 m and its flight speed is 2 m/s, the training scenarios involve obstacles randomly occurring within the UAV’s detection range. As shown in Figure 9a–l, 12 diverse training samples were constructed and safe avoidance strategies were obtained by a basic anti-collision algorithm. In the diagram, the black areas are designated to represent obstacles, which could be buildings, trees, or any other physical entities that may impede the flight of a drone. The blue lines delineate the collision-avoidance flight paths that the drone takes to navigate around these obstacles. The gray portions symbolize the drone itself. With this color-coding, it is evident how the drone utilizes its sensors and algorithms to intelligently plan a safe flight path based on the surrounding obstacle information. This not only showcases the drone’s autonomous navigation capabilities but also highlights its adaptability and flexibility in complex environments, successfully finding and executing an escape route from densely populated obstacle areas.
The relation between cognitive learning knowledge and the training samples after training is shown in Figure 10.
We can infer that the number of training samples is unequal to the amount of knowledge, because we introduced the concept of fuzzy matching, which combines similar obstacles and avoidance strategies so as to reduce dimensionality. The similarity-degree value of obstacles in Figure 9e,f is 0.87 (the value between the avoidance strategies in Figure 9d,e is 0.92), which means that we can use the same obstacle pattern (and avoidance strategy) to classify them. But the similarity-degree value of the obstacle patterns between Figure 9c,l is 0.3, so we have to construct two new obstacle patterns.
The environment is a 200 m × 200 m square crowded with 10 obstacles. A UAV has to fly from the start position (0,0), pass by two targets at position (60,60) and (140,140), and finally end at position (200,200).
The results are shown in Figure 11a and Figure 12a, where the gray circles are the amplified regions of three flying moments. To illustrate the safety degree of a UAV, an equivalent distance function is constructed. Supposing that there are k obstacles within sensory range, then the equivalent distance function can be defined as shown in Equation (13):
f e d = ρ 1 R ρ 2 R ρ k R = i = 1 k ρ i R ,   ρ i 0 , R
where R is the sensory range, and ρ is the distance between each obstacle and the UAV in R .
The results for the f e d are shown in Figure 11b and Figure 12b. We can infer from Figure 11a and Figure 12a that the AC algorithm and the CG algorithm cannot deal with complex obstacles. In Figure 11a, though the AC algorithm can obtain a safe path at the beginning, when it encounters many obstacles, the UAV cannot keep a safe distance from the crowded obstacles (see the local magnification of the black path). In Figure 12a, the cognitive game algorithm performs better than the ant colony algorithm, however, the UAV has a collision with obstacles midway ( f e d = 0 ). The method of the ant colony and cognitive game algorithms (see Figure 11b and Figure 12b) involves collisions with obstacles at moments 86.7 s and 152.6 s (when the value of f e d decreases to 0). In contrast, the proposed method can obtain a safe path and keep the f e d at the safe value.
This is because the ant colony algorithm requires a continuously re-planned flight path, and the cognitive game algorithm’s computing time is too long. However, our method can obtain an anti-collision path directly and dispense with repetitive calculations for similar obstacle patterns. When encountering complex obstacles, a cognitive UAV will find similar obstacle patterns based on the similarity-degree function, and then the corresponding avoidance strategy can be obtained. If there is no similar obstacle pattern, a new path can be designed, so the UAV can develop itself incrementally.
To make the results more convincing, the three algorithms were compared in the same environment repeatedly. Table 2 shows the statistical minimum value of the equivalent distance function M i n f and the corresponding time t min . We find that AC and CG often result in a collision in crowded environment, because there always exists an M i n f = 0 . As assumed above, M i n f is the minimum value of f e d ; when it equals 0, a collision occurs. In repeated results using CBN, the value of M i n f is always positive, which means that this proposed method always keeps a safe distance between obstacles and the UAV.
Finally, fast anti-collision is achieved like that of humans’ conditioned reflex. The effectiveness of the anti-collision algorithm is improved by avoiding duplicative planning in a similar environment.

5. Experiment and Analysis

As shown in Figure 13, the UAV’s perception model has eight directions. When the distance between the UAV and the obstacles is less than the threshold value, there is an obstacle occlusion in this direction.
The selected equipment and materials of the UAV are shown in Figure 14.
Figure 15 demonstrates the anti-collision experiment utilizing an unmanned aerial vehicle (UAV) in the presence of an obstacle in a single direction, where the red arrows represents the moving direction. It is observable that as the dynamic obstacle (people holding a planar board close to the UAV as the dynamic obstacle) approaches, the UAV is capable of swiftly maneuvering away from the obstacle and maintaining a safe distance. The upper right corner of the image illustrates the UAV’s trajectory in response to the obstacle’s movement. It can be discerned that the UAV consistently self-regulates to maintain a safe distance while flying, contingent upon the proximity of the obstacle to itself.
Table 3 shows the time comparison between the traditional anti-collision method and the proposed method. It can be seen that the traditional method needs more time from perceiving the obstacles to finally completing the anti-collision behavior because it needs to solve the anti-collision algorithm, convert the calculated anti-collision strategy into angle and throttle signals, and then convert it into control commands. However, the conditioned reflection method is directly mapped to the anti-collision instructions, which saves on the intermediate calculation steps, speeds up the operation time of the algorithm, and improves the anti-collision efficiency.
In addition to the practical testing of the UAV in a single-obstacle environment, this paper also conducted anti-collision experiments with obstacles in two different directions As shown in Figure 16, one was an anti-collision test in the constructed right-angle-obstacle environment, and Figure 17 features an anti-collision test in the constructed horizontal-obstacle environment, where the red arrows represents the moving direction. It can be observed that the UAV can quickly respond to these obstacles and can maintain its stability in a safe area. In Figure 16 and Figure 17, annotations made in the upper right corner indicate the UAV’s movement trend in response to the obstacles. From the top-down view, it is evident that when the UAV is too close to an obstacle, it exhibits a repulsive behavior, and when it is too far away, it exhibits an attractive behavior, thus maintaining a safe distance.

6. Conclusions

In terms of UAVs, we have argued that it can follow a cognitive method like the human conditioned reflex. After reviewing anti-collision studies, a cognitive brain-like model for UAVs has been proposed from the context of a bio-behavioral perspective. Following are some concluding remarks: Unlike traditional anti-collision algorithms, the proposed method can cognize by interacting with environment gradually. And it can learn new anti-collision mapping relations by online simulations. Similar to the human conditioned reflex, after experiencing similar obstacles, cognitive UAVs have the ability to avoid obstacles without thinking.
Still, there are several issues in need of attention and further investigations, including practical studies and algorithm optimization.

Author Contributions

Q.Z. contributed to the conceptualization and design of the study, performed the data analysis, and was instrumental in drafting the initial manuscript. R.W. played a pivotal role in the acquisition of funding, provided critical insights into the methodology, and significantly contributed to the interpretation of results. S.H. was essential in the data collection process, developed the software used for the analysis, and participated in revising the manuscript by contributing important intellectual content. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Science and Technology Innovation 2030 Key Project of “New Generation Artificial Intelligence”, China (No. 2018AAA0102403).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated or analyzed during this study are included in the manuscript. All data included in this study are available upon request by contact with the corresponding author.

Acknowledgments

We thank the anonymous reviewers for their careful review and helpful suggestions that greatly improved the manuscript. We thank Kai Zhou for suggesting improvements after reading early versions of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Amin, J.N.; Boskovic, J.D.; Mehra, K. A Fast and Efficient Approach to Path Planning for Unmanned Vehicles. In Proceedings of the AIAA Guidance, Navigation, and Control Conference, Keystone, Colorado, 21–24 August 2006. [Google Scholar]
  2. Czubenko, M.; Kowalczuk, Z.; Ordys, A. Autonomous Driver Based on an Intelligent System of Decision-Making. Cogn. Comput. 2015, 7, 569–581. [Google Scholar] [CrossRef]
  3. Wei, R.; Zhang, Q.; Xu, Z. Peers’ Experience Learning for Developmental Robots. Int. J. Soc. Robot. 2019, 12, 35–45. [Google Scholar] [CrossRef]
  4. Zhang, Q.; Wei, R. Unmanned aerial vehicle perception system following visual cognition invariance mechanism. IEEE Access 2019, 7, 45951–45960. [Google Scholar] [CrossRef]
  5. Beard, R.W.; McLain, T.W. Trajectory Planning for Coordinated Rendezvous of Unmanned Air Vehicles. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Denver, CO, USA, 14–17 August 2000. [Google Scholar]
  6. Wang, S.L.; Wei, R.X. Laguerre Diagram Construction Algorithm for Path Planning. Syst. Eng. Electron. 2013, 3, 552–556. [Google Scholar]
  7. Wei, R.X.; Wang, S.L. Self-Optimization A-Star Algorithm for UAV Path Planning Based on Laguerre Diagram. Syst. Eng. Electron. 2015, 3, 577–582. [Google Scholar]
  8. Beard, R.W.; McLain, T.W.; Goodrich, M. Coordinated Target Assignment and Intercept for Unmanned Air Vehicles. IEEE Trans. Robot. Autom. 2002, 18, 911–922. [Google Scholar] [CrossRef]
  9. Russell, S.; Norvig, P. Artificial Intelligence: A Modern Approach; Prentice Hall: London, UK, 2003. [Google Scholar]
  10. Mitchell, J.S.; Keirsey, D.M. Planning Strategic Paths through Variable Terrain Data. In Proceedings of the SPIE Conference on Applications of Artificial Intelligence, Arlington, VA, USA, 29 April–4 May 1984. [Google Scholar]
  11. Stentz, A. Optimal and Efficient Path Planning for Partially-Known Environments. In Proceedings of the 1994 International Conference on Robotics and Automation, Los Alamitos, CA, USA, 8–13 May 1994. [Google Scholar]
  12. Khatib, O. Real-Time Obstacle Avoidance for Manipulators and Mobile Robots. Int. J. Robot. Res. 1986, 5, 90–98. [Google Scholar] [CrossRef]
  13. Vanek, B.; Peni, T.; Bokor, J.; Balas, G. Practical approach to real-time trajectory tracking of UAV formations. In Proceedings of the American Control Conference, Portland, OR, USA, 28–30 June 2000. [Google Scholar]
  14. Fogel, D.B.; Fogel, L.J. Optimal Routing of Multiple Autonomous Underwater Vehicles through Evolutionary Programming. In Proceedings of the 1990 Symposium on Autonomous Underwater Vehicles Technology, Washington, DC, USA, 5–6 June 1990; pp. 44–47. [Google Scholar]
  15. Capozzi, B.J.; Vagners, J. Evolving (Semi)-Autonomous Vehicles. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Montreal, QB, Canada, 6–9 August 2001. [Google Scholar]
  16. Luo, B.; Hussain, A.; Mahmud, M.; Tang, J. Advances in Brain-Inspired Cognitive Systems. Cogn. Comput. 2016, 8, 795–796. [Google Scholar] [CrossRef]
  17. Tu, Z.; Zheng, A.; Yang, E.; Luo, B.; Hussain, A. A Biologically Inspired Vision-Based Approach for Detecting Multiple Moving Objects in Complex Outdoor Scenes. Cogn. Comput. 2015, 7, 539–551. [Google Scholar] [CrossRef]
  18. Hofstadter, D.R. Metamagical Themas: Questing for the Essence of Mind and Pattern; Basic Books; Hachette UK: London, UK, 1985. [Google Scholar]
  19. Beck, E.C.; Doty, R.W. Conditioned flexion reflexes acquired during combined catalepsy and de-efferentation. J. Comp. Physiol. Psychol. 1957, 50, 211–216. [Google Scholar] [CrossRef]
  20. Soltysik, S.; Gasanova, R. The effect of “partial reinforcement” on classical and instrumental conditioned reflexes acquired under continuous reinforcement. Acta Biol. Exp. 1969, 29, 29–49. [Google Scholar]
  21. Badia SB, I.; Pyk, P.; Verschure PF, M.J. A fly-locust based neuronal control system applied to an unmanned aerial vehicle: The invertebrate neuronal principles for course stabilization, altitude control and collision avoidance. Int. J. Robot. Res. 2007, 26, 759–772. [Google Scholar] [CrossRef]
  22. Weng, J.; McClelland, J.; Pentland, A.; Sporns, O.; Stockman, I.; Sur, M.; Thelen, E. Autonomous Mental Development by Robots and Animals. Science 2001, 291, 599–600. [Google Scholar] [CrossRef] [PubMed]
  23. Weng, J.; Hwang, W. Incremental Hierarchical Discriminant Regression. IEEE Trans. Neural Netw. 2007, 18, 397–415. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, D.; Zhang, L.; Weng, J. Spatio-Temproal Adaptation in the Unsupervised Development of Networked Visual Neurons. IEEE Trans. Neural Netw. 2009, 6, 992–1008. [Google Scholar] [CrossRef] [PubMed]
  25. Lim, M.-S.; Oh, S.-R.; Son, J.; You, B.-J.; Kim, K.-B. A Human-like Real-time Grasp Synthesis Method for Humanoid Robot Hands. Robot. Auton. Syst. 2000, 30, 261–271. [Google Scholar] [CrossRef]
  26. Asada, M. Towards Artificial Empathy: How Can Artificial Empathy Follow the Developmental Pathway of Natural Empathy? Int. J Soc. Robot. 2015, 7, 19–33. [Google Scholar] [CrossRef]
  27. Sandini, G.; Metta, G.; Vernon, D. Robotcub: An Open Framework for Research in Embodied Cognition. In Proceedings of the 4th IEEE/RAS International Conference on Humanoid Robots, Santa Monica, CA, USA, 10–12 November 2004; pp. 13–32. [Google Scholar]
  28. Simmons, R.G. Structured Control for Autonomous Robots. IEEE Trans. Robot. Autom. 1994, 10, 34–43. [Google Scholar] [CrossRef]
  29. Rosenblatt, J.K.; Thorpe, C.E. Combining Multiple Goals in a Behavior-Based Architecture. In Proceedings of the IEEE International Conference on Robotics and Automation, Pittsburgh, PA, USA, 5–9 August 1995; pp. 136–141. [Google Scholar]
  30. Brooks, R.A. A Robust Layered Control System for a Mobile Robot. IEEE Trans. Robot Autom. 1986, 2, 14–23. [Google Scholar] [CrossRef]
  31. Rescorla, R.A. Pavlovian Conditioning-It’s Not What You Think It Is. Am. Psychol. 1988, 43, 151–160. [Google Scholar] [CrossRef]
  32. Papini, M.R.; Bitterman, M.E. The role of contingency in classical conditioning. Psychol. Rev. 1990, 97, 396–403. [Google Scholar] [CrossRef] [PubMed]
  33. Von Eckardt, B. What Is Cognitive Science? MIT Press: Cambridge, MA, USA, 1996. [Google Scholar]
  34. Blomberg, O. Concepts of cognition for cognitive engineering. Int. J. Aviat. Psychol. 2011, 21, 85–104. [Google Scholar] [CrossRef]
  35. Zhou, H.; Friedman, H.S.; von der Heydt, R. Coding of border ownership in monkey visual cortex. J. Neurosci. 2000, 20, 6594–6611. [Google Scholar] [CrossRef] [PubMed]
  36. Rivara, C.B.; Sherwood, C.C.; Bouras, C.; Hof, P.R. Stereological characterization and spatial distribution patterns of Betz cells in the human primary motor cortex. Anat. Rec. Part A Discov. Mol. Cell. Evol. Biol. 2003, 270, 137–151. [Google Scholar] [CrossRef] [PubMed]
  37. Steriade, M.; Llinás, R.R. The functional states of the thalamus and the associated neuronal interplay. Physiol. Rev. 1988, 68, 649–742. [Google Scholar] [CrossRef] [PubMed]
  38. Sherman, S.M.; Guillery, R.W. Exploring the Thalamus; Academic Press: Cambridge, MA, USA, 2000. [Google Scholar]
  39. Chakravarthy, V.S.; Joseph, D.; Bapi, R.S. What do the basal ganglia do? A modeling perspective. Biol. Cybern. 2010, 103, 237–253. [Google Scholar] [CrossRef] [PubMed]
  40. Smith, Y.; Wichmann, T. Functional Anatomy and Physiology of the Basal Ganglia. In Deep Brain Stimulation in Neurological & Psychiatric Disorders; Humana Press: Totowa, NJ, USA, 2008; pp. 33–62. [Google Scholar]
  41. Stocco, A.; Lebiere, C.; Anderson, J.R. Conditional Routing of Information to the Cortex: A Model of the Basal Ganglia’s Role in Cognitive Coordination. Psychol. Rev. 2010, 117, 541–574. [Google Scholar] [CrossRef] [PubMed]
  42. Ridderinkhof, K.R.; Ullsperger, M.; Crone, E.A.; Nieuwenhuis, S. The Role of the Medial Frontal Cortex in Cognitive Control. Science 2004, 306, 443–447. [Google Scholar] [CrossRef] [PubMed]
  43. Moore, T.; Armstrong, K.M. Selective gating of visual signals by microstimulation of frontal cortex. Nature 2003, 421, 370–373. [Google Scholar] [CrossRef]
  44. Murray, E.A.; Wise, S.P.; Graham, K.S. Chapter 1: The History of Memory Systems, The Evolution of Memory Systems: Ancestors, Anatomy, and Adaptations, 1st ed.; Oxford University Press: Oxford, UK, 2016; pp. 22–24. [Google Scholar]
  45. Bouton, M.E. Learning and Beavior: A Contemporary Synthesis, 2nd ed.; Sinauer: Sunderland, MA, USA, 2016. [Google Scholar]
  46. Lee, T.S.; Mumford, D. Hierarchical Bayesian inference in the visual cortex. JOSA A 2003, 20, 1434–1448. [Google Scholar] [CrossRef]
  47. Lee, T.S. The visual system’s internal model of the world. Proc. IEEE. 2013, 103, 1359–1378. [Google Scholar] [CrossRef] [PubMed]
  48. Chen, X.; ZHAO, Y.; Han, J. An improved ant colony optimization algorithm for robotic path planning. Control Theory Appl. 2010, 27, 821–825. [Google Scholar]
  49. Ru, C.; Wei, R.; Guo, Q.; Zhang, L. Guidance control of cognitive game for unmanned aerial vehicle autonomous collision avoidance. Control Theory Appl. 2014, 31, 1555–1560. [Google Scholar]
Figure 1. Brain pathway of the conditioned reflex cycle in the human anti-collision response.
Figure 1. Brain pathway of the conditioned reflex cycle in the human anti-collision response.
Sensors 24 02790 g001
Figure 2. The cognitive control architecture of a UAV.
Figure 2. The cognitive control architecture of a UAV.
Sensors 24 02790 g002
Figure 3. The architecture of the RBN for an anti-collision response.
Figure 3. The architecture of the RBN for an anti-collision response.
Sensors 24 02790 g003
Figure 4. The hierarchical RBN model of the trust region, flight path, and waypoints.
Figure 4. The hierarchical RBN model of the trust region, flight path, and waypoints.
Sensors 24 02790 g004
Figure 5. A brain-like conditioned reflex cycle in cognitive UAVs.
Figure 5. A brain-like conditioned reflex cycle in cognitive UAVs.
Sensors 24 02790 g005
Figure 6. Structure of obstacle-avoidance mapping.
Figure 6. Structure of obstacle-avoidance mapping.
Sensors 24 02790 g006
Figure 7. Improved structure of obstacle-avoidance mapping.
Figure 7. Improved structure of obstacle-avoidance mapping.
Sensors 24 02790 g007
Figure 8. Gridding model of a UAV.
Figure 8. Gridding model of a UAV.
Sensors 24 02790 g008
Figure 9. The 12 diverse training samples.
Figure 9. The 12 diverse training samples.
Sensors 24 02790 g009
Figure 10. The relation between developmental knowledge and the training samples.
Figure 10. The relation between developmental knowledge and the training samples.
Sensors 24 02790 g010
Figure 11. Comparison of CR and AC among crowded obstacles.
Figure 11. Comparison of CR and AC among crowded obstacles.
Sensors 24 02790 g011
Figure 12. Comparison of CR and CG among crowded obstacles.
Figure 12. Comparison of CR and CG among crowded obstacles.
Sensors 24 02790 g012
Figure 13. The UAV’s perception model.
Figure 13. The UAV’s perception model.
Sensors 24 02790 g013
Figure 14. The selected equipment and materials of the UAV.
Figure 14. The selected equipment and materials of the UAV.
Sensors 24 02790 g014
Figure 15. A UAV’s anti-collision response to one obstacle.
Figure 15. A UAV’s anti-collision response to one obstacle.
Sensors 24 02790 g015
Figure 16. Anti-collision test in the constructed right-angle-obstacle environment.
Figure 16. Anti-collision test in the constructed right-angle-obstacle environment.
Sensors 24 02790 g016
Figure 17. Anti-collision test in the constructed horizontal-obstacle environment.
Figure 17. Anti-collision test in the constructed horizontal-obstacle environment.
Sensors 24 02790 g017
Table 1. Detailed description of the eight processes.
Table 1. Detailed description of the eight processes.
Number of ProcessesFunctions
1SensingUAVs use sensors onboard to detect obstacles around them.
2PreprocessingTo preprocess the mass and unclassified information of obstacles.
3CognitionIn collision avoidance, this process involves having an awareness of familiar environment and cognizing obstacles.
4Obstacle learningThis process aims to learn and reserve paths and to map the paths between trust regions and waypoints.
5BehaviorIts function is to design an anti-collision strategy and to behave safely in an environment.
6Strategy learningThis process is to memorize anti-collision strategies.
7Pre-actionThe waypoints cannot be decoded and acted by UAVs, so this process clarifies them so as to transform them into commands.
8ActionCognitive UAVs follow the commands from the RBN to avoid crowded obstacles.
Table 2. Repeated simulation results.
Table 2. Repeated simulation results.
Test No.ACCGCR
Minf/(m)tmin/(s)Minf/(m)tmin/(s)Minf/(m)tmin/(s)
1096.70169.20.2576.8
2087.5089.40.4289.1
30165.20164.30.3479.7
40148.30172.80.43267.5
5087.9094.60.26285.6
Table 3. Time comparison between the traditional anti-collision method and conditioned reflection method.
Table 3. Time comparison between the traditional anti-collision method and conditioned reflection method.
Environment1st2nd3rd4th
Traditional method6.4 s8.6 s7.1 s6.5 s
Conditioned reflection3.7 s5.2 s3.6 s4.4 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Wei, R.; Huang, S. Cognitive Control Architecture for the Practical Realization of UAV Collision Avoidance. Sensors 2024, 24, 2790. https://doi.org/10.3390/s24092790

AMA Style

Zhang Q, Wei R, Huang S. Cognitive Control Architecture for the Practical Realization of UAV Collision Avoidance. Sensors. 2024; 24(9):2790. https://doi.org/10.3390/s24092790

Chicago/Turabian Style

Zhang, Qirui, Ruixuan Wei, and Songlin Huang. 2024. "Cognitive Control Architecture for the Practical Realization of UAV Collision Avoidance" Sensors 24, no. 9: 2790. https://doi.org/10.3390/s24092790

APA Style

Zhang, Q., Wei, R., & Huang, S. (2024). Cognitive Control Architecture for the Practical Realization of UAV Collision Avoidance. Sensors, 24(9), 2790. https://doi.org/10.3390/s24092790

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop