Next Article in Journal
Smoothed Particle Hydrodynamics Modelling of Bergy Bit and Offshore Structure Interactions Due to Large Waves
Next Article in Special Issue
Ship Autonomous Berthing Strategy Based on Improved Linear-Quadratic Regulator
Previous Article in Journal
Optimal Arrangements of Renewable Energy Systems for Promoting the Decarbonization of Desalination Plants
Previous Article in Special Issue
A Review of Autonomous Berthing Technology for Ships
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method for Multi-AUV Cooperative Area Search in Unknown Environment Based on Reinforcement Learning

National Key Laboratory of Autonomous Marine Vehicle Technology, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(7), 1194; https://doi.org/10.3390/jmse12071194
Submission received: 2 April 2024 / Revised: 12 July 2024 / Accepted: 12 July 2024 / Published: 16 July 2024
(This article belongs to the Special Issue Unmanned Marine Vehicles: Perception, Planning, Control and Swarm)

Abstract

:
As an emerging direction of multi-agent collaborative control technology, multiple autonomous underwater vehicle (multi-AUV) cooperative area search technology has played an important role in civilian fields such as marine resource exploration and development, marine rescue, and marine scientific expeditions, as well as in military fields such as mine countermeasures and military underwater reconnaissance. At present, as we continue to explore the ocean, the environment in which AUVs perform search tasks is mostly unknown, with many uncertainties such as obstacles, which places high demands on the autonomous decision-making capabilities of AUVs. Moreover, considering the limited detection capability of a single AUV in underwater environments, while the area searched by the AUV is constantly expanding, a single AUV cannot obtain global state information in real time and can only make behavioral decisions based on local observation information, which adversely affects the coordination between AUVs and the search efficiency of multi-AUV systems. Therefore, in order to face increasingly challenging search tasks, we adopt multi-agent reinforcement learning (MARL) to study the problem of multi-AUV cooperative area search from the perspective of improving autonomous decision-making capabilities and collaboration between AUVs. First, we modeled the search task as a decentralized partial observation Markov decision process (Dec-POMDP) and established a search information map. Each AUV updates the information map based on sonar detection information and information fusion between AUVs, and makes real-time decisions based on this to better address the problem of insufficient observation information caused by the weak perception ability of AUVs in underwater environments. Secondly, we established a multi-AUV cooperative area search system (MACASS), which employs a search strategy based on multi-agent reinforcement learning. The system combines various AUVs into a unified entity using a distributed control approach. During the execution of search tasks, each AUV can make action decisions based on sonar detection information and information exchange among AUVs in the system, utilizing the MARL-based search strategy. As a result, AUVs possess enhanced autonomy in decision-making, enabling them to better handle challenges such as limited detection capabilities and insufficient observational information.

1. Introduction

In areas such as maritime search and rescue, intelligence reconnaissance, submarine resource exploration, and marine observation, AUVs [1] have demonstrated significant advantages with their autonomy and excellent endurance. However, as we continue to explore and develop the ocean, the area where AUVs perform search missions is also expanding, and the environment they face is becoming increasingly complex. A single AUV with limited endurance and detection capabilities is unable to cope. Therefore, when facing complex tasks, we generally consider multiple AUVs to collaboratively work [2]. The basic idea of multi-AUV cooperative control is to have multiple AUVs with relatively simple structures and functionalities work together as a team. This enhances the overall operational range and detection capabilities of the multi-AUV system. Within the multi-AUV system, each AUV can make behavioral decisions based on its own capabilities and sensor data, collaborating with other AUVs to accomplish the final mission. Additionally, AUVs operate in a marine environment with many unknowns and uncertainties. Moreover, underwater AUVs have limited sensing and communication capabilities, and these challenging conditions during operations place high demands on the AUVs’ autonomous decision-making abilities. Therefore, researching methods for multi-AUV cooperative control to enhance the intelligence, autonomy [3], and coordination among AUVs is essential. This enables AUVs to have improved decision-making capabilities and work collaboratively with other AUVs to accomplish complex tasks. This technological direction is becoming an inevitable trend in the development of multi-AUV cooperative underwater search [4].
For regional search tasks, traditional methods often employ a comb-like scanning approach [5,6,7]. In this method, AUVs follow pre-planned trajectories for step-by-step scanning. However, under this approach, AUVs lack autonomous decision-making capabilities. In complex and unknown marine environments, AUVs relying solely on pre-planned trajectories may struggle to handle unexpected situations such as encountering obstacles, making it challenging to effectively complete search tasks. Furthermore, for military area search and reconnaissance missions, traditional methods involve AUV search trajectories that exhibit strong regularity. This lack of randomness can make it easier for enemy forces to detect and potentially target our AUVs, which is not ideal for ensuring the stealth and security of our operations. When dealing with large-scale regional search tasks, the traditional approach involves dividing the area into several subregions [8,9,10] and then assigning these subregions to individual AUVs for scanning using a comb-like pattern. Therefore, before initiating AUV searches, it is necessary to pre-plan the search area, which includes area partitioning, subregion allocation, and setting AUV comb-like scanning trajectories. For different search areas, planning often needs to be adjusted, and irregular search areas can pose challenges in terms of area division and AUV comb-like scanning. Additionally, due to the difficulties associated with underwater communication [11], it is challenging for AUVs to engage in frequent and high-bandwidth information exchange. Traditional approaches often require access to information such as the positions of each AUV and the completion status of subregion searches to facilitate real-time subregion allocation. This places high demands on communication between AUVs. If AUV communication is not smooth, it can have adverse effects on the collaborative regional search tasks of the AUV cluster [12,13]. Indeed, addressing challenges posed by weak communication environments and limited local observability is crucial for improving the performance of multi-AUV systems in cooperative regional search and exploring unknown marine environments. Research in the field of multi-AUV cooperative search technology is currently focused on several key areas:
  • Intelligent algorithms: Developing intelligent algorithms that enable AUVs to make autonomous decisions in complex and uncertain underwater environments. These algorithms should allow AUVs to adapt their search strategies dynamically based on their own sensor data and available information;
  • Adaptive search strategies: Designing cooperative search strategies that can adapt to different types of search areas. AUVs should be able to perform effective searches in irregular or unpredictable environments without the need for extensive pre-planning;
  • Robustness: Enhancing the robustness of multi-AUV systems to ensure they can continue operating effectively even in the presence of individual AUV failures or communication disruptions. Redundancy and fault-tolerant control mechanisms are important in this context;
  • Collaboration of multi-AUVs in the weak communication environments: Developing communication-efficient strategies that allow AUVs to collaborate effectively with minimal information exchange. Decentralized decision-making and coordination mechanisms can be beneficial in scenarios with limited bandwidth;
  • Exploration of unknown environments: Investigating methods for AUVs to explore and map unknown or uncharted marine environments. This involves combining exploration and mapping capabilities with cooperative search strategies;
  • Optimization and planning: Researching optimization techniques and path planning algorithms that help AUVs make informed decisions about their search trajectories, considering factors such as energy efficiency, time constraints, and mission objectives.
Overall, these research areas aim to advance the capabilities of multi-AUV systems and enable them to perform efficient and adaptive cooperative search missions in challenging underwater environments. We focus on researching methods for multi-AUV cooperative area search, starting from the perspectives of adaptive search strategies and collaboration under weak communication environments. Firstly, this paper establishes a first-order kinematic and a sonar detection model for AUVs. Using Dec-POMDP, we model the cooperative area search task. Secondly, we apply the multi-agent reinforcement learning QMIX method [14] to solve the Dec-POMDP mentioned above. Building upon prior work [15], we specifically design the QMIX method based on the maximum entropy mechanism and conduct comparative analysis of the training results. Finally, we perform simulation tests and conduct a preliminary exploration of reward function settings.

2. Background

Multi-AUV cooperative regional search enables efficient coverage and detection tasks in various areas, making it a hot topic of research in both academia and industry. Some multi-AUV cooperative search systems have already found practical applications in engineering projects. One such example is the research project called Generic Ocean Array Technology System (GOATS) conducted by the Massachusetts Institute of Technology (MIT). The GOATS project utilizes multiple AUVs equipped with underwater acoustic devices to form a mobile underwater detection network for searching for underwater mines in coastal waters [16,17,18]. Building on the GOATS project, a research team composed of the National Underwater Research Center (NURC) and MIT initiated a project in 2008 known as the Generic Littoral Interoperable Networked Technology (GLINT). In the GLINT project, a multi-AUV system is equipped with various sensors to autonomously detect, locate, and track specific targets [19]. In Europe, from 2012 to 2015, various research institutions from Italy, Estonia, the United Kingdom, Spain, and Turkey collaborated on a research project called Autonomous Robotic Systems for Oceanographic and Water-Archaeological Surveys (ARROWS) [20]. The project aimed to enhance underwater scanning efficiency using multi-AUV systems and conducted research on task allocation strategies [21] and underwater communication [22].
In the academic field, there have been many successful research efforts related to multi-AUV collaborative area search. Li et al. [23] proposed a distributed dynamic predictive control (DDPC) algorithm based on predictive control principles. This algorithm predicts the states of a multi-AUV system to obtain information about the task environment and updates the AUV states as inputs for online task optimization decisions, allowing them to determine the next moment’s search area. Wang et al. [24] presented research on a multi-agent target search method for AUVs. This algorithm is based on multi-agent deep deterministic policy gradient (MADDPG) and incorporates temporal and spatial information into the reinforcement learning process. It also introduces specific rewards tailored for maritime target search scenarios. However, it does not consider the impact of sonar detection effectiveness, leading to a lack of target guidance information during the search process and subsequently affecting the efficiency of target search by AUVs. Liu et al. [25] defined the target search problem as a well-known Traveling Salesman Problem (TSP) with defined start and end points. They considered two competitive and non-communicable optimization objectives for underwater vehicles: total navigation distance and turning angles. The study also introduced mobility constraints for AUVs and utilized an improved ant colony algorithm for solving the problem. However, this research treated the search area as a known region and did not account for scenarios involving unknown search areas. Cai et al. [26] focused on the maritime search and rescue mission of multi-AUV systems and studied the multi-robot coverage path planning (MCPP) problem. They introduced a new MCPP approach, which involved transforming the MCPP problem into two sub-problems: area partition and single AUV coverage path planning, solving each individually. Although this approach considered the guidance role of prior target information, it simply divided the search area and planned coverage paths for individual AUVs. It did not account for the collaboration between AUVs. Hu et al. [27] addressed the obstacle avoidance issue encountered by underwater vehicle formations during collaborative search and target capture. They introduced an energy-optimal formation obstacle avoidance strategy and an improved self-organizing map (SOM) path planning algorithm. However, this approach relied on forming a fleet of AUVs for search, which required frequent communication between the AUVs and placed relatively high demands on underwater communication capabilities. Bai et al. [28] introduced a biologically-inspired two-layer self-organizing map algorithm for dynamic task planning involving multiple autonomous underwater vehicles. This algorithm is developed for searching multiple targets in a 3D underwater environment affected by random ocean currents and dynamic uncertain obstacles. It focuses on addressing the decoupling of task assignment and path planning in initial task planning while considering energy consumption constraints. Li et al. [29] proposed a combined approach for autonomous underwater vehicles and unmanned surface vehicles (USVs) in target search tasks in unknown marine environments with no prior information. This approach involves a local dynamic predictive control framework combined with the Lévy flight (LDPC-Lévy). The LDPC-Lévy method involves assigning subregions for AUVs to search and planning suitable positions for USVs, ensuring reasonable communication distances while exploring the environment and searching for targets.

3. Method Review

Currently, multi-AUV cooperative area search technology is still in the development stage. Many of the methods in this field draw inspiration from the more mature field of unmanned aerial vehicle (UAV) swarm search [30,31,32]. These methods can be categorized based on their search approach into formation-based search [5,6,7], partition allocation-based search [8,9,10], and swarm intelligence-based autonomous search [33,34,35].
Formation-based search involves coordinating multiple AUVs to form a configuration that maximizes the search area, and then conducting a step-by-step search in this formation [36,37,38]. This approach draws inspiration from the single AUV comb-like search method but considers multiple AUVs as a unified formation [39] to plan the search coverage path for the entire formation. Formation-based search is simple and practical, allowing for comprehensive coverage without blind spots. However, it can face challenges in complex marine environments with factors like weak communication. Maintaining formation in such conditions can be difficult. Additionally, due to the constraints of the formation, AUVs may struggle to handle unexpected situations such as obstacle avoidance. As a result, the overall autonomy of AUV formations may be limited, and the level of coordination between individual AUVs may not be very strong. Hu et al. [27] address the energy consumption challenges encountered by AUV formations during collaborative search processes. They propose an energy cost-optimal formation search strategy and an improved self-organizing map path planning algorithm. Furthermore, they introduce a Formation Comprehensive Cost (FCC) model, which considers convergence time, transformation distance, and sensor network power consumption, to facilitate obstacle avoidance within the formation. Healey et al. [7] designed a multi-underwater robots system to address shallow water mine clearance issues. In this system, a supervisor robot located outside the minefield centrally controls all swimmer robots, ensuring a high mine clearance rate. However, the most significant drawback of this system is that if the supervising robot is damaged or disabled, all other robots will be left in an uncontrolled state, which can be a critical limitation in terms of system reliability and robustness.
Partition allocation-based search methods involve dividing the search area into several subregions initially and then dynamically allocating these subregions to individual AUVs using allocation algorithms. Once an AUV receives information about its allocated subregion, it proceeds to that subregion to perform coverage and search tasks. This approach allows for a more organized and efficient distribution of search efforts among multiple AUVs. Partition allocation-based search methods offer several advantages. They divide a large, irregularly shaped search area into multiple smaller, relatively regular subregions, simplifying the complexity of the search task. Additionally, AUVs can dynamically allocate subregions through information exchange, which enhances both the autonomy of individual AUVs and the overall coordination of the multi-AUV system. However, it is important to note that this approach also increases the computational load on the AUV system and places higher demands on communication between AUVs due to the need for real-time dynamic allocation of subregions. Welling et al. [8] utilized a multi-AUV system for collaborative search and target clearance tasks. They discussed task allocation problems and compared two allocation strategies based on the nearest distance and fuzzy logic in terms of time efficiency. Hoai An Le et al. [10] researched a hierarchical search planning model that divides the search area into multiple subspaces and then performs a second-level search planning within these subspaces. This two-stage search planning approach enhances the overall efficiency and precision of the search process.
Swarm intelligence-based autonomous search methods draw inspiration from the foraging behavior of biological groups like ants [40], wolf packs [41], etc. In these methods, each AUV makes its current best decision based on observable spatial information. Through iterative steps, the AUVs work collectively to ultimately achieve complete area search coverage. This approach leverages the power of decentralized decision-making and cooperation, similar to how natural swarms of animals collaborate to achieve complex tasks. Indeed, in recent years, the rise of artificial intelligence (AI) algorithms based on machine learning has opened up new research directions for multi-agent cooperative regional search methods. These AI algorithms can enhance the decision-making, adaptability, and coordination of multi-agent systems, making them more effective in complex and dynamic environments. Researchers are increasingly exploring the integration of machine learning techniques into multi-agent systems to improve their performance and autonomy in tasks such as regional search. The multi-agent cooperative search strategies derived through swarm intelligence and machine learning methods offer several advantages. They not only enhance the autonomous decision-making capabilities of AUVs in complex marine environments but also improve the coordination between AUVs. These strategies exhibit good adaptability to different environments, making them versatile and effective tools for a wide range of scenarios. The combination of swarm intelligence and machine learning contributes to more efficient and adaptive multi-agent systems for tasks like cooperative search and exploration. Cao et al.’s work [34] focuses on target search and tracking in unknown underwater environments. They propose an integrated algorithm for a multi-autonomous underwater vehicle collaborative team. This algorithm combines Glasius bio-inspired neural network (GBNN) and bio-inspired cascaded tracking control methods. The aim is to enhance search efficiency and reduce tracking errors. GBNN plays a central role in controlling the multi-AUV team’s search for each target in this context. Liu et al.’s work [35] involves the establishment of a distributed multi-AUV cooperative search system (DMACSS) and introduces the autonomous cooperative search learning algorithm (ACSLA) integrated into DMACSS. It incorporates information fusion mechanisms and timestamp mechanisms, enabling each AUV in the system to exchange and fuse information during tasks. ACSLA is a customized learning algorithm trained using reinforcement learning methods.
In summary, the advantages and disadvantages of various search forms are shown in Table 1.
To address the challenges of limited detection capability and insufficient observational information of AUVs during area search tasks, we have developed a multi-AUV cooperative area search system (MACASS). From the perspective of enhancing AUVs’ autonomous decision-making capabilities, we integrated well-trained search strategies based on MARL into the action decision modules of each AUV within the system. The main contributions of this article are as follows:
  • Dec-POMDP environmental modeling for search tasks: In this paper, the search task is modeled as a Dec-POMDP, and relevant global states, state transition functions, and reward functions are designed. Additionally, in underwater environments where AUVs have limited sensing capabilities and cannot access global environmental state information, the paper uses sonar-based detection information as the observations available to AUVs. Furthermore, considering the challenges of underwater communication, the multi-AUV system employs a distributed control approach to minimize communication requirements;
  • Cooperative area search algorithm based on MARL: We adopt the QMIX [14] algorithm to train the search strategy for the aforementioned Dec-POMDP and then based on the design approach using the maximum entropy mechanism for QMIX proposed by Guo et al. [15], we specifically design the QMIX algorithm for the conditions of multi-AUV cooperative area search tasks. Since the multi-agent reinforcement learning algorithm based on the maximum entropy mechanism draws inspiration from the soft actor-critic (SAC) algorithm [42], this approach is referred to as SAC-QMIX;
  • Multi-AUV cooperative area search system: After modeling the multi-AUV cooperative area search task as a Dec-POMDP and training search strategy, the well-trained search strategies are integrated into the decision-making models of each AUV. Subsequently, using a distributed control approach, these AUVs form the multi-AUV cooperative area search system. During the execution of the search task, each AUV within MACASS updates the search information map in real time based on sonar detection information and information exchange among AUVs in the system. They make action decisions by the search strategies integrated into the AUVs, ultimately cooperating with each other to complete the search task.

4. Dec-POMDP Modeling

This section begins by establishing the AUV kinematic and sonar detection model. Subsequently, it models the search task as a Dec-POMDP. It specifies AUV observations based on sonar detection information and the search information map. Additionally, it takes into account the AUV’s kinematic characteristics to design corresponding actions and reward functions. This modeling approach helps create a framework for decision-making and coordination in multi-AUV cooperative search missions, considering both sensory data and the environmental context. The significance and explanations of the symbols and abbreviations used in modeling the cooperative area search task are summarized in the Appendix A.

4.1. AUV Kinematic Model

In order to provide a detailed description of the AUV’s motion, this paper establishes an inertial coordinate system E-ξηζ and a body-fixed coordinate system G-XYZ, as shown in Figure 1a. The origin of the inertial coordinate system E is located at a certain point on the water surface. The axis points north, the axis points east, and the axis points downward. Due to the relatively low cruising speed of the AUVs performing the search tasks in this paper, the effects of the Earth’s rotation can be ignored. Therefore, the coordinate system E-ξηζ is considered an inertial coordinate system. The body-fixed coordinate system, denoted as G-XYZ, is rigidly attached to the AUV itself. In this coordinate system, the origin G is located at the center of gravity of the AUV, the GX axis points along the AUV’s longitudinal axis (forward), the GY axis is perpendicular to the GX axis and points starboard (to the right), and the GZ axis is perpendicular to both the GX and GY axes and points downward.
When multiple AUVs are conducting a search in a region of size L x × L y × L z , the search area is divided along the axis into multiple subregions S i , i 1 , 2 , , I where I = L z / δ d and δ d represents the maximum sonar vertical detection distance. Multiple AUVs search the various subregions sequentially. When each AUV completes the search of a subregion S i , they immediately descend and move on to the next subregion S i + 1 for further search. This process continues until the entire region has been completely searched. When multiple AUVs are searching subregion S i , the variation in depth during AUV navigation is relatively small, and the AUVs can be considered to be operating at a fixed depth. Therefore, the problem of multiple AUVs searching subregion S i can be transformed into a two-dimensional search problem on a plane s i of size L x × L y , as shown in Figure 1b. Additionally, it is assumed that the depth at which the target is located is the same as the depth of the AUVs performing the search task. In a two-dimensional plane, the motion of AUVs can be simplified to three degrees of freedom: heave, sway, and yaw. The kinematic equations for their motion are provided in the equation below, with further details available in reference [43].
x ˙ y ˙ ψ ˙ = cos ϕ sin ϕ 0 sin ϕ cos ϕ 0 0 0 1 u v r

4.2. Sonar Detection Model

When AUVs are performing search tasks, their primary detection equipment is sonar. The detection probability of the sonar depends on the detection distance and the acoustic conditions of the operating environment. In this paper, only the detection distance between the sonar and the target is considered as the primary factor affecting sonar detection probability; the simplified sonar detection probability model [44] is represented by the following Equation (2):
P d = P D exp d D s , 0 < d D s P F , d > D s
In the equation, Pd represents the sonar detection probability, PD stands for the sonar’s detection precision, PF represents the sonar’s false alarm probability, d represents the detection distance, and Ds is the maximum detection distance of the sonar.

4.3. Search Information Map Model

From the previous information, it is evident that the problem of multiple AUVs searching subregion S i can be transformed into a two-dimensional search problem in a plane s i . Therefore, we discretize the two-dimensional plane s i which has dimensions L x × L y into an N x × N y grid map. Each grid region is denoted as grid(m, n), where m 1 , 2 , , N x , n 1 , 2 , , N y . When AUVs are performing search tasks, they update the gridded search information map using sonar detection information and information exchanged among AUVs. The search information map established in this paper consists of three main components:
  • Coverage map: This map indicates which areas have been covered or searched by the AUVs. Initially, at the start of the search task, all grid regions have a value of 0;
  • Uncertainty map: The uncertainty map represents the level of uncertainty associated with each grid region. Initial values are typically set to the max value;
  • Target presence probability map: This map estimates the probability of the presence of targets in each grid region. It provides information about how confident or uncertain the AUVs are about the presence of targets in each region. Initially, the probability of target existence in the grid region follows a uniform distribution, indicating no prior knowledge of target presence.
As the AUVs continue their search and gather more information through sonar detections and interactions with each other, these maps are updated in real time. The values in these maps evolve to reflect the changing understanding of the search environment, and AUVs use this information to make new decisions during the search process.

4.3.1. Coverage Map

At the beginning of the search task, the presence probability of targets within the sonar detection range of each AUV is generally low. Moreover, due to a lack of prior information about targets, the efficiency of AUV search efforts is often low. In order to encourage AUVs to explore unknown areas and accelerate the convergence speed of strategy learning, a coverage map is established. The coverage map is used to distinguish between grid regions that AUVs have explored and those that have never been explored. When an AUV proceeds to search a previously unexplored grid region grid(m, n), the coverage value c m n changes from 0 to 1, and it remains unchanged afterward. The update method for the coverage map is described by the following Equation (3):
c m n = 1 , if     g r i d ( m , n ) F O V 0 , else

4.3.2. Uncertainty Map

The uncertainty map is used to represent the uncertainty level within the search area. As the AUVs continue their search, the uncertainty about the environment gradually decreases. This map provides a measure of how much time AUVs have not detected those areas that have already been explored. Unlike the coverage map, the uncertainty map operates differently. When an AUV proceeds to search a grid region grid(m, n), the uncertainty value k m n for that grid region is set to its minimum value 0. As the AUV leaves that grid region, the uncertainty value k m n slowly increases over time. The specific update formula is described as shown in Equation (4). The uncertainty map primarily serves the purpose of encouraging AUVs to re-search areas that have not been searched for a long time. It does so by gradually reducing the uncertainty values in regions that have been previously explored, thereby prioritizing unexplored or less recently explored areas for further investigation by the AUVs. This strategy helps ensure comprehensive coverage and exploration of the search area. The coverage map and the uncertainty map together assist AUVs in addressing the issue of low search efficiency during the early stages of the search when there is a lack of target information. These maps help AUVs identify areas that have already been explored (coverage map) and areas where uncertainty remains high (uncertainty map). This information guides AUVs to focus their efforts on unexplored or uncertain regions, improving their search efficiency and increasing the likelihood of finding targets.
k m n ( t + 1 ) = 0 , g r i d ( m , n ) F O V ( t ) k m n ( t ) + 1 T , g r i d ( m , n ) F O V ( t )
In the equation, k m n represents the uncertainty value of the grid region grid(m, n), and FOV(t) represents the effective sonar detection range of each AUV at time t.

4.3.3. Target Presence Probability Map

The target presence probability map is used to represent the probability of the presence of a target at different locations within the search area. It helps AUVs navigate toward potential targets more efficiently. Initially, when there is no prior information about the target’s location, the probability of the target’s presence is typically assumed to be equally distributed across the entire search area. As AUVs gather information and make detections, this map is updated to reflect changes in the likelihood of a target being present in various parts of the search area. This enables AUVs to focus their search efforts on regions with a higher probability of finding a target. The target presence probability is derived using the sonar detection model and the Bayesian formula. Each AUV initially calculates the target presence probability for grid regions within its sonar detection range based on the sonar detection results, as described by the formula. Subsequently, AUVs update their target presence probabilities through information exchange between them. Then, the target presence probabilities for grid regions across the entire search area are normalized to create the final target presence probability map. The target presence probability at the detection location is described by the following equation, with detailed derivations available in reference [45]:
P j ( t + 1 ) = ( 1 P ¯ D i P ¯ F i M i 1 ) P j ( t ) k F O V i ( 1 P ¯ D i P ¯ F i M i 1 ) P k ( t ) + k F O V i ( 1 P ¯ F i M i ) P k ( t ) , if j F O V i , O ( t ) = 1 ( 1 P ¯ F i M i ) P j ( t ) k F O V i ( 1 P ¯ D i P ¯ F i M i 1 ) P k ( t ) + k F O V i ( 1 P ¯ F i M i ) P k ( t ) , if j F O V i , O ( t ) = 1 ( P ¯ D i P ¯ F i M i 1 ) P j ( t ) k F O V i ( P ¯ D i P ¯ F i M i 1 ) P k ( t ) + k F O V i ( P ¯ F i M i ) P k ( t ) ,   if j F O V i , O ( t ) = 0 ( P ¯ F i M i ) P j ( t ) k F O V i ( P ¯ D i P ¯ F i M i 1 ) P k ( t ) + k F O V i ( P ¯ F i M i ) P k ( t ) ,   if j F O V i , O ( t ) = 0
In this formula, Pj(t + 1) represents the target presence probability in grid region j at time t + 1, Pj(t) and Pk(t) represent the target presence probabilities in grid regions j and k at time t, PDi represents the sonar detection probability for AUVi, PFi represents the false alarm probability for AUVi’s sonar, Mi represents the number of grid regions within AUVi’s field of view, FOVi represents the effective sonar detection range of AUVi, O(t) is a binary variable where O(t) = 1 represents AUVi discovering a target, and O(t) = 0 represents AUVi not discovering a target at time t.

4.4. Global State and Observation Space

In this section, based on the previously established search information map, we will define the global state (S) and the observations (O) for the AUV. The global state represents the complete description of the environment and the AUV’s internal state. It includes information about the positions and states of all AUVs, the current state of the search information map (including coverage information, target presence information, and uncertainty information), the positions of any detected or suspected targets, and any other relevant variables that characterize the system at a given time. The global state S plays a significant role during the training phase of multi-agent reinforcement learning. It guides the AUVs to learn effective search strategies in situations where only local environmental information is observable. However, due to the large number of grid regions resulting from the discretization of search area Si, the dimensionality of the global state S can become excessively high, leading to the “curse of dimensionality”. To address this issue, the approach is to merge the grid regions of search area Si into subregions, where multiple grid regions g are merged into a single subregion G. We refer to the grid region g as the secondary subregion and the region G as the primary subregion This merging process is illustrated in Figure 2a.
Based on the merged grid regions in the search information map, we design the global state S as a four-tuple:
S = C , K , P , Ι
In this formula, C = c 1 , c 2 , , c N I represents the coverage information of each subregion G, K = k 1 , k 2 , , k N I represents the uncertainty information of each subregion G, P = p 1 , p 2 , , p N I represents the target existence probability information of each subregion G, N I represents the total number of subregions G, c i , k i , p i represents the coverage, uncertainty, and target existence probability information of subregion Gi, which includes the average values of coverage, uncertainty, and target existence probability. Ι = η 1 , η 2 , , η N represents the position and orientation of AUVs, N represents the total number of AUVs. η i = x i , y i , ψ i , x i , y i represents position coordinates of the i-th AUV along the ξ and η axes in the inertial coordinate system E-ξηζ. ψ i π , π represents the heading angle of the i-th AUV.
The AUV’s sonar detection range is assumed to be the grid regions within a maximum sonar detection distance of Ds, and the number of grid regions within this range is assumed to be N F O V . The AUV’s sonar detection range is illustrated in Figure 2b. We combine the detection information obtained by AUV through sonar and the communication information among AUVs to form the AUV’s observation (O). The observation for the i-th AUV, denoted as O i = C i , K i , P i , η i , H i , includes the coverage value, uncertainty value, and target existence probability within the maximum sonar detection range of AUVi. The variables relating to the relative positions and orientations of AUVs are also included in the observation, H i = γ i , 1 , γ i , 2 , , γ i , N 1 represents the relative spatial relationships between AUVi and the other AUVs, γ i , j = ψ i , j , d i , j represents the relative spatial relationship between AUVi and AUVj, ψ i , j π , π represents the relative orientation angle between AUVi and AUVj, and d i , j represents the distance between AUVi and AUVj.
The calculation formula for ψ i , j is as follows:
ψ i , j = ψ i π arctan y j y i x j x i ,   x j < x i , y j y i ψ i + π arctan y j y i x j x i ,   x j < x i , y j < y i ψ i arctan y j y i x j x i , x j > x i
The calculation formula for d i , j is as follows:
d i , j = x i x j 2 + y i y j 2

4.5. Action Space Design

In the search process of multiple AUVs in the 2D plane s, we do not consider factors such as ocean currents. We assume that the AUVs travel at a constant speed, meaning that the AUV’s longitudinal velocity u remains constant and is not affected by any lateral interference. In other words, the AUV’s lateral velocity v is set to 0. Therefore, the action set a for AUVs executing the search task in this paper can be defined as Δ ψ t . It represents the change in the heading angle of AUVs at time t. Therefore, we define the action space of AUVs at time t as follows:
A = a t 1 , a t 2 , , a t N N
In Equation (9), a t i represents the action decided by the i-th AUV at time t, and N represents the number of AUVs. Taking into account the constraints imposed by the kinematic characteristics of AUVs, we impose the following restrictions on the action set a:
90 ° Δ ψ t 90 ° a t i 90 ° , 45 ° , 0 ° , 45 ° , 90 °
In Equation (10), positive angle values indicate that the AUV is turning counterclockwise, negative angle values indicate that the AUV is turning clockwise, and an angle value of 0 indicates that the AUV is maintaining its current heading angle.
According to the AUV’s two-dimensional kinematic equation in the inertial coordinate system E-ξηζ, the AUV’s position can be updated using the following equation:
x t + 1 y t + 1 = x t y t + cos ψ t sin ψ t sin ψ t cos ψ t u t v t d t
In Equation (11), xt+1 and yt+1 represent the position coordinates of the AUV along the ξ and η axes in the inertial coordinate system at time t + 1. ψ t represents the heading angle of the AUV at time t, ut and vt represent the longitudinal velocity and lateral velocity of the AUV at time t, and dt represents the time interval.
At time t + 1, the heading angle ψ t can be updated using the following equation:
ψ t + 1 = ψ t + Δ ψ t
In Equation (12), ψ t represents the heading angle of AUV at time t + 1, and Δ ψ t represents the change in heading angle at time t.

4.6. Design of Reward Functions

The individual reward for an AUV in the multi-AUV cooperative area search problem can indeed be divided into multiple components to encourage specific behaviors. It is a common practice in reinforcement learning to create composite rewards that consider different aspects of the agent’s behavior. Based on the search information map, we have identified four components for individual reward: coverage reward (rc), uncertainty reward (rk), target discovery reward (rt), and collaboration reward (ra). Therefore, the individual reward (r) for an AUV can be expressed as follows:
r = α r c + β r k + γ r t + λ r a
In Equation (13), α , β , γ , λ represent the weight coefficients for each part of the reward.

4.6.1. Coverage Reward

When an AUV goes to a grid area grid(m, n) that has never been searched before, i.e., the coverage value cmn of grid area grid(m, n) is 0, the AUV receives a coverage reward rc, and the coverage value cmn of that grid area grid(m, n) changes from 0 to 1, indicating that it has been searched by the AUV, and its coverage value cmn will not change thereafter. The specific update formula for coverage reward is as follows:
c m n = 1 , if     g r i d ( m , n )   is   searched 0 , else
If the coverage value of the grid area that the AUV goes to is 0, then the AUV receives a coverage reward. If the coverage value is already 1, the AUV does not receive any coverage reward.

4.6.2. Uncertainty Reward

For AUVs that have already searched a grid area grid(m, n), even though its coverage value cmn no longer changes, its uncertainty value kmn continues to increase over time. Therefore, when an AUV goes to a grid area grid(m, n) that has already been searched, it can still receive uncertainty rewards rk related to its uncertainty value kmn. The expression for uncertainty rewards is as follows:
r k = k m n k max 1 ,   k m n < K k m n k max , k m n K
In Equation (15), K is the tuning coefficient. When the uncertainty value kmn of the grid area grid(m, n) that the AUV is going to is greater than K, rk is a positive reward. However, when kmn is less than K, rk is a negative reward.

4.6.3. Target Discovery Reward

The target discovery reward (rt) is designed based on the probability of target existence in the grid cells within the AUV’s sonar detection range. The probability of target existence represents the degree to which an AUV is inclined toward the presence of a target. If the grid cells within the AUV’s sonar detection range have a higher probability of target existence, it implies that the AUV is closer to the target, and there is a greater likelihood of discovering the target in those areas. Essentially, the AUV is more likely to find the target in regions with a higher probability of target presence. Therefore, based on the probability of target presence in the grid cells within the AUV’s sonar detection range, we have designed the target discovery reward (rt), and its specific expression is as follows:
r t = g r i d m , n F O V p m n N F O V
In Equation (16), FOV represents the grid cells within the AUV’s sonar detection range, and NFOV represents the number of grid cells within the AUV’s sonar detection range.

4.6.4. Collaborative Reward

In the process of multi-AUV searching within partitioned area Si, the collaborative reward is designed to encourage individual AUVs to perform distributed area searches. This aims to maximize the coverage of the multi-AUV system’s search area while avoiding the problem of AUVs clustering too closely and causing overlap in their sonar detection ranges. However, it is important to maintain a minimum distance between AUVs to ensure effective communication for sharing search information and updating the search information map. The specific formula for the collaborative reward may take into account factors related to AUV dispersion and the maintenance of communication within a certain range. The reward formula should encourage AUVs to strike a balance between spreading out to cover more area and staying close enough for efficient information sharing. Therefore, we design the collaboration reward (ra), and its specific expression is as follows:
r a = R , d i < D s k d i D s D c D s , D s < d i < D c 0 , D c < d i
In Equation (17), di represents the average distance between the i-th AUV and the other AUVs, Ds represents the maximum detection distance of the sonar, and Dc represents the communication distance between AUVs. When the average distance di between AUVi and the other AUVs is less than Ds, the AUVs are too close together, resulting in a significant overlap in their sonar detection ranges and a higher risk of collisions. This situation is unfavorable for the collaborative search of multiple AUVs, and AUVi receives a negative reward. When di is between Ds and Dc, the distance between AUVi and the other AUVs is appropriate, and AUVi receives a positive reward that is positively correlated with di. When di exceeds Dc, the average distance between AUVi and the other AUVs exceeds the communication distance, which hinders information sharing in the multi-AUV system. Therefore, AUVi does not receive any reward.

5. Algorithm Design

The QMIX algorithm [14] is a value-based algorithm in multi-agent reinforcement learning. Its core idea is to use a mixing neural network to combine the action-values Q i π τ i , a i , i 1 , 2 , , N of individual agents in a complex nonlinear manner to synthesize the joint action-value Q t o t ( s , a ) . This is used to estimate the action-value of joint actions a = a 1 , a 2 , , a N N in a multi-agent system. Indeed, it is worth noting that in the action-values Q i π τ i , a i of individual agents, the history of observed actions τ i is based on local observations o i . This means that the action-values of each agent take into account the historical actions. Additionally, the joint action-value Q t o t ( s , a ) is monotonically increasing with respect to the action-values Q i π τ i , a i of individual agents, satisfying Equation (18). This constraint ensures consistency between the centralized strategy Q t o t ( s , a ) and the individual agent strategies Q i π τ i , a i .
Q t o t s , a Q i π τ i , a i 0 , a i
The QMIX algorithm based on the maximum entropy mechanism [15] builds upon the QMIX algorithm and incorporates ideas from the soft actor-critic algorithm, which is based on maximum entropy reinforcement learning. First, each agent adopts an actor-critic network framework, enhancing the existing network structure by introducing a policy network π θ . Secondly, during the training process, a maximum entropy mechanism is introduced. Each agent not only aims to maximize the expected return but also seeks to maximize entropy. This mechanism makes the learning search strategies of agents more stochastic, encouraging agents to explore the environment more extensively and preventing the search strategy from converging prematurely.

5.1. Algorithm’s Computation Process

After incorporating the maximum entropy mechanism, the expected return for each agent during the training process is no longer solely determined by the rewards provided by the environment. Instead, it is composed of both the rewards from the environment and the entropy associated with the actions taken by each agent. Therefore, the expected return in reinforcement learning is formulated as follows:
R s t , a t + i = 1 N α i H i π | τ i , t
In Equation (19), s t represents the global state of the environment at time t, a t = a 1 , t , a 2 , t , , a N , t denotes the joint action taken by multiple agents at time t, R s t , a t represents the joint reward received by multiple agents, and α i is the entropy regularization coefficient. H i π | τ i , t represents the entropy of the action a i , t taken by agent i at time t. The formula for calculating the entropy Hi for agent i at time t is as follows:
H π a i , t | τ i , t = a A i π i a τ i , t log π i a τ i , t
The training objective is to learn the optimal policy π * that maximizes the expected return. In other words, the optimal policy π * satisfies the following equation:
π * = arg max π Ε s ~ p , a ~ π t = 0 γ t R s t , a t + i = 1 N α i H i π | τ i , t
The joint action-value function Q t o t ( s , a ) for multiple agents can be calculated through a mixing neural network using the following expression:
Q t o t s , a = F m i x q 1 τ 1 , a 1 , q 2 τ 2 , a 2 , , q   N τ N , a N
In Equation (22), q i τ i , a i represents the action-value function for the i-th agent, and the F m i x function represents the mixing neural network.
The individual state-value function v i τ i for an agent can be obtained by taking the expectation of the action-value function q i τ i , a i with respect to action a i , and the calculation formula is as follows:
v i τ i = Ε a i ~ π q i τ i , a i
The overall state-value function V t o t s for multiple agents can be obtained by taking the expected value of the joint action-value function Q t o t ( s , a ) with respect to the joint action a. The calculation formula is expressed as follows:
V t o t s = E a π Q t o t s , a = E a π F m i x q 1 τ 1 , a 1 , q 2 τ 2 , a 2 , , q N τ N , a N
Due to the complexity of taking the expected value of the joint action-value function Q t o t s , a with respect to the joint action a, to reduce computational complexity, Equation (24) can be approximated as follows:
V t o t s = E a π F m i x q 1 τ 1 , a 1 , q 2 τ 2 , a 2 , , q N τ N , a N F m i x E a 1 π q 1 τ 1 , a 1 , E a 2 π q 2 τ 2 , a 2 , , E a N π q N τ N , a N F m i x v 1 τ 1 , v 2 τ 2 , , v N τ N
After incorporating the maximum entropy mechanism, the overall state-value V s based on the maximum entropy mechanism can be expressed as a function of the individual state-values v i τ i of agents and the entropy H i π a i | τ i . The specific formula is as follows:
V π s = V t o t s + i = 1 N α i H i π | τ i , t = Ε a ~ π F m i x q 1 τ 1 , a 1 , q 2 τ 2 , a 2 , , q N τ N , a N + i = 1 N α i H i π | τ i , t F m i x Ε a 1 π q 1 τ 1 , a 1 , Ε a 2 π q 2 τ 2 , a 2 , , Ε a N π q N τ N , a N + i = 1 N α i H i π | τ i , t F m i x v 1 τ 1 , v 2 τ 2 , , v N τ N + i = 1 N α i H i π | τ i , t
The Bellman equation for the action-value function Q π s , a regularized by entropy can be expressed as follows:
Q π s , a = Ε a ~ π R s , a + γ Q t o t s , a + i = 1 N α i H i π | τ i = R s , a + γ V t o t s + i = 1 N α i H i π | τ i = R s , a + γ V π s
Based on Equation (27), we can compute the target Q-values. Therefore, the loss function for the critic network can be represented as follows:
L Q = Ε s , a , s ~ D Q t o t s , a R s , a + γ V ¯ s 2
In Equation (28), Q t o t s , a represents the joint action-value function for multiple agents. It can be computed from individual action-value functions q i τ i , a i of each agent using the mixing neural network as defined in Equation (22). R s , a denotes the joint reward obtained from the environment. V ¯ s represents the target state-value function, which can be calculated based on the individual target state-value functions of each agent and the entropies of the agents, as defined in Equation (26).
The training objective of the actor network is to maximize the overall state-value V π s . Therefore, the loss function of the actor network can be represented as follows:
L π = Ε s ~ D , a ~ π V π s
The above can be summarized, and the algorithm’s computational flowchart is illustrated in the following Figure 3. The significance and explanations of symbols and abbreviations used in the design process of the algorithm are summarized in the Appendix A.

5.2. Algorithm Framework Design

The SAC-QMIX algorithm consists of networks for multiple agents and a mixing neural network. The specific network architecture is designed as shown in the Figure 4. The design involves the following components:
  • Individual agent networks: Each agent has its own actor-critic neural network architecture. This network takes local observations as input and outputs action probabilities (actor) and action-value estimates (critic);
  • Mixing neural network (QMIX): The mixing neural network takes the individual agents’ action-values and combines them in a complex, non-linear manner to estimate the joint action-values for the multi-agent system;
  • Maximum entropy mechanism: During training, a maximum entropy mechanism is incorporated into the agent networks, which encourages exploration and introduces randomness into the policies. This mechanism helps agents to learn more robust and diverse strategies;
According to the diagram in Figure 4, the agent’s network includes both a policy network π θ and a value network Q ϕ . The mixing neural network consists of a hypernetwork and a feedforward neural network. The hypernetwork generates the weights w1, w2 and biases b1, b2 for the feedforward network based on the global state (s). The feedforward network combines the action-values Q a of each agent to produce the joint action-value Q t o t . It is important to note that the weights w1, w2 of the feedforward network must remain positive to satisfy the condition in Equation (18).
Figure 4. (a) The structure of the mixing neural network. (b) The overall structure SAC-QMIX consists of. (c) Network structure of the agents.
Figure 4. (a) The structure of the mixing neural network. (b) The overall structure SAC-QMIX consists of. (c) Network structure of the agents.
Jmse 12 01194 g004
The training approach used in this paper for the algorithm is called centralized training with decentralized execution (CT-DE). This means that during the training process, the mixing neural network synthesizes the joint action-value function Q t o t s , a and the global state-value function V s based on the individual agents’ action-value functions q i τ i , a i and state-value functions v i τ i as well as the global environment state s. However, after training is completed, each agent makes real-time action decisions based solely on observations o when executing tasks. The training framework of the algorithm is illustrated in Figure 5. This training method utilizes the global state s during the training process to compensate for the adverse effects caused by the insufficient observational information of the individual agent. The global state s contains information from all agents, which helps alleviate the instability issues that can arise during multi-agent reinforcement learning training.
Incorporating the computational process of the above algorithm with the algorithmic framework, the training process of the SAC-QMIX algorithm is as shown in Algorithm 1.
Algorithm 1. SAC-QMIX algorithm training process.
1 :   Input   initial   policy   parameters   θ , Q - functions   parameters   , ϕ 1 , ϕ 2 , empty replay buffer D
2 :   Set   target   parameters   equal   to   main   parameters ,   ϕ targ , 1 ϕ 1 ,   ϕ targ , 2 ϕ 2
3: for steps = 1:M do:
4:    Observe the state s, obtain observations oi, for i = 1, 2, …, n
5 :   Select   action   a i ~ π θ · | τ i , for i = 1, 2, …, n
6 :   Execute   action   a i in the environment, for i = 1, 2, …, n
7 :   Observe   next   state   s ,   next   observation   o ,   reward   r s , a , done signal
8 :   Store   s , o , a , r , d in episode experience exp
9 :   if   s is terminal, done is true then
10:    Supplement the episode experience exp to the maximum episode experience
11:    Reset environment
12:   end if
13:   Store episode experience exp in replay buffer D
14:   if it is time to update then
15:    Randomly sample a batch of episode experience exp from replay buffer D
16 :   Compute   individual   action - value   q i : q   i k = Q ϕ , i τ k , a k ,   q   targ , i k = Q ϕ targ , i τ k , a k ,
               for i = 1, 2, for k = 1, 2, …, n
17 :   Compute   individual   state - value   v i :
                v i k τ i = Ε a i π q i k τ k , a k for i = 1, 2 for k = 1, 2, …, n
                v targ , i k τ i = Ε a i π q targ , i k τ k , a k for i = 1, 2 for k = 1, 2, …, n
18: Compute united action-value function:
                Q t o t , i s , a = F m i x q i 1 τ 1 , a 1 , q i n τ n , a n for i = 1, 2
19 :   Compute   the   overall   state - value   function   V t o t , i s = F m i x v i 1 τ 1 , , v i n τ n
               for i = 1, 2
20: Compute target Q-function y:
                y r , o , d = r + γ min i = 1 , 2 F m i x v targ , i 1 τ 1 , , v targ , i n τ n + i = 1 n α i H i π τ i
21: Update Q-network parameters using gradient descent:
                ϕ i 1 B exp B Q t o t , i s , a y r , o , d 2 for i = 1, 2
22: Update policy network parameters using gradient ascent:
                θ 1 B exp B min i = 1 , 2 V t o t , i s + i = 1 n α i H i π τ i
23: Update the entropy regularization coefficient using gradient descent:
                α 1 B exp B i = 1 n H π θ a i | τ i H ¯
24: Soft update target network:
25 :   ϕ targ , 1 = τ ϕ 1 + 1 τ ϕ targ , 1
26 :   ϕ targ , 2 = τ ϕ 2 + 1 τ ϕ targ , 2
27:  end if
28: end for

6. Simulation Test

This paper sets the collaborative search area for multiple AUVs in two different scenarios. Scenario 1 is a square area with dimensions of 420 m × 420 m, and the Cartesian coordinate system’s origin is located at the bottom left corner of the search area. The search area is divided into 7 × 7 primary subregions, with each primary subregion measuring 60 m × 60 m in size. The corresponding secondary subregions are 14 × 14. Scenario 2 is an irregular octagonal area with sides measuring 180 m. This scenario is primarily used to test the adaptability of the search strategy to irregular search areas. We used the QMIX algorithm as well as an improved version of the QMIX algorithm to train in both of these scenarios. Initially, all grid cells have a coverage mark of 0, an uncertainty value of the maximum value kmax equal to 1 and an uncertainty growth factor T equal to 50. In the corresponding uncertainty reward module, the tuning coefficient K is set to 0.5. The configuration includes three AUVs, each with a maximum sonar detection range (Ds) of 90 m, a sonar detection accuracy (PD) of 0.8, a sonar false alarm probability (PF) of 0.2, an effective communication distance between AUVs of 240 m, an AUV cruising speed (v) of 1 m per second, and an AUV action execution time (Δt) of 60 s. We established a neural network framework based on PyTorch in the Python platform and implemented the algorithm. The neural network parameters and learning training parameters are detailed in Table 2. Finally, we conducted simulation testing in a reinforcement learning gym environment.
Our simulation Scenario 1 is set up as follows: AUV 1 starts at (30 m, 30 m), AUV 2 starts at (90 m, 30 m), and AUV 3 starts at (150 m, 30 m). Each AUV has an initial heading of 90°. The target is located at (330 m, 270 m) and remains stationary.
Figure 6 shows the simulation diagram based on the reinforcement learning gym environment. Figure 6a represents the search trajectory map based on the QMIX algorithm, Figure 6c represents the search trajectory map based on the SAC-QMIX algorithm, and Figure 6b,d represent the corresponding uncertainty maps of the search areas. In Figure 6, the first AUV is represented in blue, the second AUV is represented in yellow, the third AUV is represented in green, and the golden pentagon star represents the target. The search areas of AUVs are assigned different colors based on the magnitude of uncertainty values, with uncertainty values increasing from red to light blue.
Our simulation Scenario 2 is set up as follows: the first AUV’s initial position is at (30 m, 270 m), the second AUV’s initial position is at (30 m, 210 m), and the third AUV’s initial position is at (30 m, 150 m). The initial heading for each AUV is 0°. The target position is at (330 m, 270 m), and it remains constant. It is worth noting that, compared to the regular square search area in Scenario 1, the four corners of the search area in Scenario 2 are restricted zones. Therefore, the boundary conditions for the search area in Scenario 2 will be more complex. The final simulation results for Scenario 2 are shown in Figure 7. In Figure 7a, the search trajectory is based on the QMIX algorithm, and in Figure 7c, it is based on the SAC-QMIX algorithm. Figure 7b,d represent the corresponding uncertainty maps of the search area.
Based on the simulation results of the two scenarios, we can observe that both the search strategies using the QMIX algorithm and the SAC-QMIX algorithm achieve high area coverage rates and successfully locate the target. However, the trajectory pattern of the search based on the QMIX algorithm lacks regularity, indicating lower coordination among AUVs during the search process. In contrast, the search trajectory based on the SAC-QMIX algorithm appears more organized, suggesting better coordination among AUVs. The latest research progress of the multi AUVs cooperative area search system (MACASS) in unknown environments is shown in the Supplementary Materials.
We also conducted simulation tests on the fault tolerance performance of the MARL-based multi-AUV cooperative area search system. To simulate the potential single AUV breaks down in the multi-AUV cooperative area search system, we set a limited number of decision times for AUV 2. When AUV 2 reaches this limited number of decision times, it ceases to perform the search task, while the remaining AUVs continue with the search task. In the simulation test, we applied the search area from Scenario 1. The multi-AUV cooperative area search system utilized the search strategy based on the SAC-QMIX algorithm. We present the simulation results in Figure 8, where Figure 8a represents the initial stage of the search task, Figure 8b represents the stage when each AUV has made five decisions, Figure 8c represents the stage when AUV 2 breaks down, Figure 8d represents the stage when the remaining AUVs have made 15 decisions, and Figure 8e represents the stage when the remaining AUVs have made 25 decisions.
From the simulation results, it can be observed that even when one AUV experiences a failure and anchors, the remaining AUVs can still achieve a high area coverage rate in completing the area search task. This indicates that the MARL-based multi-AUV cooperative area search system possesses a certain degree of fault tolerance.
Furthermore, since the target information in the search area is initially unknown to each AUV, they can only obtain observational information through sonar detection and communication between AUVs during the search process and use this information to make action decisions based on the policy network. We illustrate the changes in the target probability map for AUVs during the search process in Figure 9. As shown in Figure 9, at the beginning of the search task, due to the lack of a priori information about the target, the target follows a uniform distribution in the search area. However, as the search task officially begins, each AUV calculates the target’s probability of existence within the sonar detection range using the Bayesian formula based on sonar detection results. The target probability map is updated accordingly. Therefore, as the AUVs continue the search, the target’s probability of existence in the grid areas without targets gradually approaches zero, and eventually, the target probability map converges to the target location.
We compared the training results of the SAC-QMIX algorithm with the QMIX algorithm graph. The training results of both algorithms are shown in Figure 10, where Figure 10a represents the training reward graph and Figure 10b represents the search area coverage. Besides, we adopt the classic reinforcement learning method DQN (Deep Q-Network) to train the collaborative area search strategy network for multiple AUVs, and we use the training results as a contrast. The training method is also distributed training.
From the training results, we can observe that the SAC-QMIX algorithm converges faster than the QMIX algorithm. SAC-QMIX achieves a high level of search strategy in approximately 100,000 iterations. This is because the addition of the maximum entropy mechanism in the SAC-QMIX algorithm encourages agents to explore more during the training process, preventing them from getting stuck in local optima, thus speeding up the overall convergence process. The collaborative search strategy based on DQN did not converge to satisfactory results, and the reward and area detection rate remained consistently low.
Finally, we conducted ablation experiments on the coverage reward and uncertainty reward to explore the effects of different reward modules on the algorithm training results and the final test results. The algorithm used during the training process was the QMIX algorithm. The training results of different reward modules are shown in the Figure 11.
From Figure 11, it can be observed that although the final training results under different reward settings are consistent, the AUVs can complete the area search task with a higher degree of exploration. However, under the initial reward setting, the algorithm converges faster, starting to converge around 600,000 steps. The algorithm without the coverage reward module converges slightly slower, starting to converge around 700,000 steps. On the other hand, the algorithm without the uncertainty reward module converges slowest, starting to converge around 1.1 million steps.
Corresponding to the ablation experiment mentioned above, this article conducts simulation tests based on the policy networks trained with different reward modules to validate the performance of different policy networks in searching for targets. The simulation area is a 600 m × 600 m square region, with the origin of the Cartesian coordinate system located at the bottom left corner of the search area. The initial positions of AUVs are as follows: AUV 1 starts at (30 m, 30 m), AUV 2 starts at (90 m, 30 m), and AUV 3 starts at (150 m, 30 m). The initial heading of each AUV is set to 90°, and the cruising speed of AUVs is 1 m/s. The sonar detection model used is the one established in Section 4.2. In each simulation test, each AUV can make a maximum of 40 decisions. The target’s position (xt, yt) is randomly generated in each simulation test, and the simulation is conducted 300 times, where x t 0 , 600 , y t 0 , 600 ; the results are shown in the Table 3.
From Table 3, it can be seen that policy networks trained with different reward modules can enable multiple AUVs to achieve a high success rate in completing the target search task. However, the success rate of searches without the coverage reward and without the uncertainty reward modules is significantly lower than that of searches under the initial reward module.
To compare and analyze the target search efficiency under different reward modules, this paper randomly selected 80 simulated tests from the successful target search simulations mentioned above and recorded the number of steps required to search for the target. The statistical results are shown in the Figure 12.
From Figure 12, we can see that the average number of decisions for the initial reward is 10, with a total variance of 9.49. For the network without coverage reward, the average number of decisions is 13, with a total variance of 40.23. Similarly, for the network without uncertainty reward, the average number of decisions is also 13, with a total variance of 74.32. The policy network based on the initial reward has the fewest average decision times and the smallest total variance, indicating the best performance. The policy network based on the reward without coverage and the reward without uncertainty have the same average decision times, but the total variance of the former is smaller, indicating better stability compared to the latter.
To analyze the training results under different reward modules as described above, we recorded the proportion of global rewards attributed to each module throughout the training process. The results are depicted in Figure 13. From the figure, it is evident that during the initial stages of training, the rewards for exploration and uncertainty accounted for a significant proportion. Specifically, coverage rewards constituted 30% of the global rewards, uncertainty rewards comprised 40%, while target discovery rewards represented only 3%. As training progressed, the proportions of coverage and uncertainty rewards decreased somewhat, whereas the proportion of target discovery rewards gradually increased and eventually stabilized around 30%. This indicates that each AUV began to focus its search on areas with a higher probability of target presence.

7. Conclusions

In this paper, we have transformed the problem of multi-AUV collaborative area search into a two-dimensional plane search problem and modeled the search process as a decentralized partially observable Markov decision process (POMDP). To address the challenge of limited perception capabilities of AUVs in underwater environments, we have constructed search information maps, including coverage maps, uncertainty maps, and target existence probability maps. Each AUV updates these maps through sonar sensing and communication with other AUVs to make action decisions during the search process. Furthermore, we have employed multi-agent reinforcement learning to solve the Markov decision process we established earlier. We first used the classical QMIX algorithm to solve the Markov decision process mentioned above, verifying the feasibility of multi-agent reinforcement learning in the problem of multi-AUV cooperative area search. Subsequently, we further analyzed and specifically designed the improved algorithm SAC-QMIX based on the QMIX algorithm and conducted a comparison between the two algorithms based on simulation test results and training outcomes. Finally, we validated the effectiveness of the established model and multi-agent reinforcement learning method through simulation tests. From the simulation test results, it can be observed that each AUV can complete the search task through action decisions generated by the policy network. Lastly, from the reward training curve, it is evident that the SAC-QMIX algorithm converges faster compared to the QMIX algorithm, but the final absolute performance is the same as the QMIX algorithm. This is a limitation of the SAC-QMIX algorithm and an area for improvement in the future. In the future, we can further research the multi-AUV cooperative area search system from the perspectives of target search time and searching moving targets. Additionally, we can compare the performance of more different algorithms and explore local path planning methods for the execution module in the MACASS.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/jmse12071194/s1, Video S1: Presentation of Multi-AUV Cooperative Area Search Task in Unknown Environments. The latest research progress of the multi AUVs cooperative area search system (MACASS) that we have established has been provided. The video showcased the latest research progress of MACASS that we have established. From the video, it can be seen that each AUV has preliminarily achieved a high area detection rate in completing area search tasks in unknown obstacle environments while ensuring safety conditions.

Author Contributions

Conceptualization, Y.L. and M.M.; methodology, Y.L.; software, M.M.; validation, Y.L., M.M., and D.W.; formal analysis, Y.L.; investigation, M.M.; resources, G.L.; data curation, W.C.; writing—original draft preparation, M.M.; writing—review and editing, Y.L.; visualization, D.W.; supervision, J.C.; project administration, J.C.; funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the National Natural Science Foundation of China (52071099, 52071104) and the National Key Laboratory of Autonomous Marine Vehicle Technology: 2023-SXJQR-SYSJJ01.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article and Supplementary Materials.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

In Section 4, the significance and explanations of the symbols and abbreviations used in modeling the cooperative area search task are summarized in the following table:
Table A1. The significance and explanations of symbols and abbreviations used in Section 4.
Table A1. The significance and explanations of symbols and abbreviations used in Section 4.
Symbol and AbbreviationSignificance and Explanation
L x Length of the search area (m)
L y Width of the search area (m)
L z Height of the search area (m)
S i The i-th search subzone
δ d Height of the subzone (m)
s i Simplified subzone
P d Sonar detection probability
P D Sonar detection precision
P F Sonar’s false alarm probability
D s Maximum sonar detection range (m)
GPrimary subregion
grid(m, n)/gThe secondary subregion
c m n Coverage value of grid cell grid(m, n)
k m n Uncertainty value of grid cell grid(m, n)
p m n Target presence probability of grid cell grid(m, n)
η i Pose information of the i-th AUV
FOVThe effective sonar detection range of each AUV
NFOVThe number of grid cells within the AUV’s sonar detection range
SGlobal state
OiObservation for the i-th AUV
AJoint actions of AUVs
ψ i , j Relative orientation angle between AUVi and AUVj
d i , j Distance between AUVi and AUVj
H i Relative spatial relationships between AUVi and the other AUVs
rIndividual reward of AUV
rcCoverage reward
rkUncertainty reward
rtTarget discovery reward
raCollaboration reward
In Section 5, this paper summarizes the significance and explanations of symbols and abbreviations used in the design process of the QMIX algorithm based on the maximum entropy mechanism as shown in the following table.
Table A2. The significance and explanations of symbols and abbreviations used in Section 5.
Table A2. The significance and explanations of symbols and abbreviations used in Section 5.
Symbol and AbbreviationSignificance and Explanation
s t Global state of the environment at time t
a t Joint action taken by multiple agents at time t
R s t , a t Joint reward received by multiple agents
π θ Policy network
Q ϕ Value network
H i π | τ i , t The entropy of the action a i , t taken by agent i at time t
q i τ i , a i Action-value function for the i-th agent
F m i x Mixing neural network
Q t o t ( s , a ) Joint action-value function
v i τ i Individual state-value function
V t o t s Overall state-value function
Q π s , a Action-value function regularized by entropy
V s Entropy-regularized state-value function
V ¯ s Target state-value function

References

  1. Sahoo, A.; Dwivedy, S.K.; Robi, P.S. Advancements in the field of autonomous underwater vehicle. Ocean Eng. 2019, 181, 145–160. [Google Scholar] [CrossRef]
  2. Hadi, B.; Khosravi, A.; Sarhadi, P. A Review of the Path Planning and Formation Control for Multiple Autonomous Underwater Vehicles. J. Intell. Robot. Syst. 2021, 101, 67. [Google Scholar] [CrossRef]
  3. Gafurov, S.A.; Klochkov, E.V. Autonomous unmanned underwater vehicles development tendencies. In Proceedings of the 2nd International Conference on Dynamics and Vibroacoustics of Machines (DVM), Samara, Russia, 15–20 September 2015; pp. 141–148. [Google Scholar]
  4. Wang, L.; Zhu, D.; Pang, W.; Zhang, Y. A survey of underwater search for multi-target using Multi-AUV: Task allocation, path planning, and formation control. Ocean. Eng. 2023, 278, 114393. [Google Scholar] [CrossRef]
  5. Zhang, J.Y.; Ning, X.; Ma, S.C. An improved particle swarm optimization based on age factor for multi-AUV cooperative planning. Ocean. Eng. 2023, 287, 115753. [Google Scholar] [CrossRef]
  6. Yoon, S.; Qiao, C. Cooperative Search and Survey Using Autonomous Underwater Vehicles (AUVs). IEEE Trans. Parallel Distrib. Syst. 2011, 22, 364–379. [Google Scholar] [CrossRef]
  7. Healey, A.J. Application of formation control for multi-vehicle robotic minesweeping. In Proceedings of the 40th IEEE Conference on Decision and Control (CDC), Orlando, FL, USA, 4–7 December 2001; pp. 1497–1502. [Google Scholar]
  8. Welling, D.M.; Edwards, D.B. Multiple autonomous underwater crawler control for mine reacquisition. In Proceedings of the ASME International Mechanical Engineering Congress and Exposition, Orlando, FL, USA, 5–11 November 2005; pp. 257–262. [Google Scholar]
  9. Miao, R.; Pang, S.; Jiang, D. Development of an Inexpensive Decentralized Autonomous Aquatic Craft Swarm System for Ocean Exploration. J. Mar. Sci. Appl. 2019, 18, 343–352. [Google Scholar] [CrossRef]
  10. Hoai An Le, T.; Duc Manh, N.; Tao Pham, D. A DC programming approach for planning a multisensor multizone search for a target. Comput. Oper. Res. 2014, 41, 231–239. [Google Scholar] [CrossRef]
  11. Yang, Y.; Xiao, Y.; Li, T.S. A Survey of Autonomous Underwater Vehicle Formation: Performance, Formation Control, and Communication Capability. IEEE Commun. Surv. Tutor. 2021, 23, 815–841. [Google Scholar] [CrossRef]
  12. Yan, Z.P.; Zhang, C.; Tian, W.D.; Zhang, M.Y. Formation trajectory tracking control of discrete-time multi-AUV in a weak communication environment. Ocean. Eng. 2022, 245, 110495. [Google Scholar] [CrossRef]
  13. Chen, S.; Ho, D.W. Consensus control for multiple AUVs under imperfect information caused by communication faults. Inf. Sci. 2016, 370, 565–577. [Google Scholar] [CrossRef]
  14. Rashid, T.; Samvelyan, M.; de Witt, C.S.; Farquhar, G.; Foerster, J.; Whiteson, S. QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning. In Proceedings of the 35th International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  15. Guo, F.; Wu, Z. Learning maximum entropy policies with QMIX in cooperative MARL. In Proceedings of the 2nd IEEE International Conference on Electronic Technology, Communication and Information, ICETCI 2022, Changchun, China, 27–29 May 2022; pp. 357–361. [Google Scholar]
  16. Edwards, J.R. Real-time classification of buried targets with teams of unmanned vehicles. In Proceedings of the Ocean’s 2002 Conference and Exhibition, Biloxi, MS, USA, 29–31 October 2002; pp. 316–319. [Google Scholar]
  17. Liu, T.-C.; Schmidt, H. AUV-based seabed target detection and tracking. In Proceedings of the Ocean’s 2002 Conference and Exhibition, Biloxi, MS, USA, 29–31 October 2002; pp. 474–478. [Google Scholar]
  18. Bovio, E.; Cecchi, D.; Baralli, F. Autonomous underwater vehicles for scientific and naval operations. Annu. Rev. Control 2006, 30, 117–130. [Google Scholar] [CrossRef]
  19. Schneider, T.; Schmidt, H. Unified command and control for heterogeneous marine sensing networks. J. Field Robot. 2010, 27, 876–889. [Google Scholar] [CrossRef]
  20. Allotta, B.; Costanzi, R.; Magrini, M.; Monni, N.; Moroni, D.; Pascali, M.A.; Reggiannini, M.; Ridolfi, A.; Salvetti, O.; Tampucci, M. Towards a robust system helping underwater archaeologists through the acquisition of geo-referenced optical and acoustic data. In Proceedings of the 10th International Conference on Computer Vision Systems, ICVS 2015, Copenhagen, Denmark, 6–9 July 2015; pp. 253–262. [Google Scholar]
  21. Tsiogkas, N.; Frost, G.; Monni, N.; Lane, D. Facilitating multi-AUV collaboration for marine archaeology. In Proceedings of the MTS/IEEE OCEANS 2015, Genova, Italy, 18–21 May 2015. [Google Scholar]
  22. Maurelli, F.; Saigol, Z.; Insaurralde, C.C.; Petillot, Y.R.; Lane, D.M. Marine world representation and acoustic communication: Challenges for multi-robot collaboration. In Proceedings of the 2012 IEEE/OES Autonomous Underwater Vehicles, AUV 2012, Southampton, UK, 24–27 September 2012. [Google Scholar]
  23. Li, J.; Li, C.; Zhang, H. Distributed Dynamic Predictive Control for Multi-AUV Target Searching and Hunting in Unknown Environments. Machines 2022, 10, 366. [Google Scholar] [CrossRef]
  24. Wang, G.; Wei, F.; Jiang, Y.; Zhao, M.; Wang, K.; Qi, H. A Multi-AUV Maritime Target Search Method for Moving and Invisible Objects Based on Multi-Agent Deep Reinforcement Learning. Sensors 2022, 22, 8562. [Google Scholar] [CrossRef]
  25. Yan, Z.; Liu, W.; Xing, W.; Herrera-Viedma, E. A Multi-Objective Mission Planning Method for AUV Target Search. J. Mar. Sci. Eng. 2023, 11, 144. [Google Scholar] [CrossRef]
  26. Cai, C.; Chen, J.; Yan, Q.; Liu, F. A Multi-Robot Coverage Path Planning Method for Maritime Search and Rescue Using Multiple AUVs. Remote Sens. 2023, 15, 93. [Google Scholar] [CrossRef]
  27. Hu, X.; Shi, Y.; Bai, G.; Chen, Y. Collaborative Search and Target Capture of AUV Formations in Obstacle Environments. Appl. Sci. 2023, 13, 16. [Google Scholar] [CrossRef]
  28. Bai, G.; Chen, Y.; Hu, X.; Shi, Y.; Jiang, W.; Zhang, X. Multi-AUV dynamic trajectory optimization and collaborative search combined with task urgency and energy consumption scheduling in 3-D underwater environment with random ocean currents and uncertain obstacles. Ocean. Eng. 2023, 275, 113841. [Google Scholar] [CrossRef]
  29. Li, C.; Li, J.; Zhang, G.; Chen, T. IROA-based LDPC-Lévy method for target search of multi AUV-USV system in unknown 3D environment. Ocean. Eng. 2023, 286, 115648. [Google Scholar] [CrossRef]
  30. Hou, K.; Yang, Y.; Yang, X.; Lai, J. Distributed Cooperative Search Algorithm With Task Assignment and Receding Horizon Predictive Control for Multiple Unmanned Aerial Vehicles. IEEE Access 2021, 9, 6122–6136. [Google Scholar] [CrossRef]
  31. Fei, B.; Bao, W.; Zhu, X.; Liu, D.; Men, T.; Xiao, Z. Autonomous Cooperative Search Model for Multi-UAV With Limited Communication Network. IEEE Internet Things J. 2022, 9, 19346–19361. [Google Scholar] [CrossRef]
  32. Xiao, J.; Wang, G.; Zhang, Y.; Cheng, L. A Distributed Multi-Agent Dynamic Area Coverage Algorithm Based on Reinforcement Learning. IEEE Access 2020, 8, 33511–33521. [Google Scholar] [CrossRef]
  33. Rajnarayan, D.G.; Ghose, D. Multiple Agent Team Theoretic Decision-Making for Searching Unknown Environments. In Proceedings of the 42nd IEEE Conference on Decision and Control, Maui, HI, USA, 9–12 December 2003; pp. 2543–2548. [Google Scholar]
  34. Cao, X.; Sun, H.B.; Jan, G.E. Multi-AUV cooperative target search and tracking in unknown underwater environment. Ocean Eng. 2018, 150, 1–11. [Google Scholar] [CrossRef]
  35. Liu, Y.; Wang, M.; Su, Z.; Luo, J.; Xie, S.R.; Peng, Y.; Pu, H.Y.; Xie, J.J.; Zhou, R. Multi-AUVs Cooperative Target Search Based on Autonomous Cooperative Search Learning Algorithm. J. Mar. Sci. Eng. 2020, 8, 843. [Google Scholar] [CrossRef]
  36. Huang, B.; Zhou, B.; Zhang, S.; Zhu, C. Adaptive prescribed performance tracking control for underactuated autonomous underwater vehicles with input quantization. Ocean. Eng. 2021, 221, 108549. [Google Scholar] [CrossRef]
  37. Zhou, B.; Huang, B.; Su, Y.; Wang, W.; Zhang, E. Two-layer leader-follower optimal affine formation maneuver control for net-worked unmanned surface vessels with input saturations. Int. J. Robust Nonlinear Control 2023, 34, 3631–3655. [Google Scholar] [CrossRef]
  38. Huang, B.; Zhang, S.; He, Y.; Wang, B.; Deng, Z. Finite-time anti-saturation control for Euler–Lagrange systems with actuator failures. ISA Trans. 2022, 124, 468–477. [Google Scholar] [CrossRef]
  39. Huang, B.; Song, S.; Zhu, C.; Li, J.; Zhou, B. Finite-time distributed formation control for multiple unmanned surface vehicles with input saturation. Ocean. Eng. 2021, 233, 109158. [Google Scholar] [CrossRef]
  40. Dorigo, M.; Birattari, M.; Stuetzle, T. Ant colony optimization–Artificial ants as a computational intelligence technique. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  41. Liu, C.; Yan, X.; Liu, C.; Wu, H. The Wolf Colony Algorithm and Its Application. Chin. J. Electron. 2011, 20, 212–216. [Google Scholar]
  42. Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor. In Proceedings of the 35th International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  43. Fossen, T.I. Handbook of Marine Craft Hydrodynamics and Motion Control; John Wiley and Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  44. Jia, Q.; Xu, H.; Feng, X.; Gu, H.; Gao, L. Research on cooperative area search of multiple underwater robots based on the prediction of initial target information. Ocean. Eng. 2019, 172, 660–670. [Google Scholar] [CrossRef]
  45. Shem, A.G.; Mazzuchi, T.A.; Sarkani, S. Addressing uncertainty in UAV navigation decision-making. IEEE Trans. Aerosp. Electron. Syst. 2008, 44, 295–313. [Google Scholar] [CrossRef]
Figure 1. (a) AUV coordination. (b) AUV projection on the horizontal plane.
Figure 1. (a) AUV coordination. (b) AUV projection on the horizontal plane.
Jmse 12 01194 g001
Figure 2. (a) Grid region division. (b) Sonar detection range.
Figure 2. (a) Grid region division. (b) Sonar detection range.
Jmse 12 01194 g002
Figure 3. The computational flowchart.
Figure 3. The computational flowchart.
Jmse 12 01194 g003
Figure 5. Training framework diagram.
Figure 5. Training framework diagram.
Jmse 12 01194 g005
Figure 6. (a) Search trajectory map based on QMIX. (b) Uncertainty map based on QMIX. (c) Search trajectory map based on SAC-QMIX. (d) Uncertainty map based on SAC-QMIX.
Figure 6. (a) Search trajectory map based on QMIX. (b) Uncertainty map based on QMIX. (c) Search trajectory map based on SAC-QMIX. (d) Uncertainty map based on SAC-QMIX.
Jmse 12 01194 g006
Figure 7. (a) Search trajectory map based on QMIX. (b) Uncertainty map based on QMIX. (c) Search trajectory map based on SAC-QMIX. (d) Uncertainty map based on SAC-QMIX.
Figure 7. (a) Search trajectory map based on QMIX. (b) Uncertainty map based on QMIX. (c) Search trajectory map based on SAC-QMIX. (d) Uncertainty map based on SAC-QMIX.
Jmse 12 01194 g007
Figure 8. (a) Initial stage of the search task (b) Stage of decision-making with 5 decisions. (c) Stage when AUV 2 breaks down (d) Stage of decision-making with 15 decisions (e) Stage of decision-making with 25 decisions.
Figure 8. (a) Initial stage of the search task (b) Stage of decision-making with 5 decisions. (c) Stage when AUV 2 breaks down (d) Stage of decision-making with 15 decisions (e) Stage of decision-making with 25 decisions.
Jmse 12 01194 g008
Figure 9. (a) Probability map for initial stages. (b) Probability map for 5 decisions. (c) Probability map for 8 decisions. (d) Probability map for 15 decisions (e) Probability map for 25 decisions.
Figure 9. (a) Probability map for initial stages. (b) Probability map for 5 decisions. (c) Probability map for 8 decisions. (d) Probability map for 15 decisions (e) Probability map for 25 decisions.
Jmse 12 01194 g009
Figure 10. (a) Training curve of reward. (b) Training curve of coverage rate.
Figure 10. (a) Training curve of reward. (b) Training curve of coverage rate.
Jmse 12 01194 g010
Figure 11. Training results under different reward settings.
Figure 11. Training results under different reward settings.
Jmse 12 01194 g011
Figure 12. Number of target search decisions under different reward modules.
Figure 12. Number of target search decisions under different reward modules.
Jmse 12 01194 g012
Figure 13. The proportion of global rewards attributed to different reward modules.
Figure 13. The proportion of global rewards attributed to different reward modules.
Jmse 12 01194 g013
Table 1. Advantages and disadvantages of different search forms.
Table 1. Advantages and disadvantages of different search forms.
Collaborative Search FormAdvantagesDisadvantages
Formation-based search
  • Low method complexity, simple, and feasible
  • High area coverage rate
  • Suitable for large-scale, no-blind-zone searches
  • During the search process, the autonomy is low
  • Collaboration among AUVs is not strong
  • Challenging to deal with complex environments
Partitioned allocation search
  • Simplifying the search difficulty in complex areas
  • Good autonomy and inter-group coordination of AUVs
  • Strong dependence on communication
  • There may be delays if the computational load is high
Swarm intelligent autonomous search
  • High autonomy in AUV decision-making
  • Strong adaptability to different environments
  • Weak dependence on communication
  • Prone to getting stuck in local optima
Table 2. Training parameters table.
Table 2. Training parameters table.
ParameterValue
Actor learning rate lr1 1 × 10 4
Critic learning rate lr2 1 × 10 4
α learning rate lr3 1 × 10 4
Target entropy H ¯ 0.2
Discount factor γ 0.99
Soft update coefficient τ 0.1
RNN dimension64
Batch size32
Max episodes M30,000
Max steps T 1.5 × 10 6
Table 3. Table of test results with different reward modules.
Table 3. Table of test results with different reward modules.
Reward ModuleNumber of TestsSuccessful Search TimesSuccess Search Rate
Initial reward30029999.7%
Reward without coverage30027892.7%
Reward without uncertainty30026287.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Ma, M.; Cao, J.; Luo, G.; Wang, D.; Chen, W. A Method for Multi-AUV Cooperative Area Search in Unknown Environment Based on Reinforcement Learning. J. Mar. Sci. Eng. 2024, 12, 1194. https://doi.org/10.3390/jmse12071194

AMA Style

Li Y, Ma M, Cao J, Luo G, Wang D, Chen W. A Method for Multi-AUV Cooperative Area Search in Unknown Environment Based on Reinforcement Learning. Journal of Marine Science and Engineering. 2024; 12(7):1194. https://doi.org/10.3390/jmse12071194

Chicago/Turabian Style

Li, Yueming, Mingquan Ma, Jian Cao, Guobin Luo, Depeng Wang, and Weiqiang Chen. 2024. "A Method for Multi-AUV Cooperative Area Search in Unknown Environment Based on Reinforcement Learning" Journal of Marine Science and Engineering 12, no. 7: 1194. https://doi.org/10.3390/jmse12071194

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop