Next Article in Journal
Tsallis Entropy-Based Complexity-IPE Casualty Plane: A Novel Method for Complex Time Series Analysis
Previous Article in Journal
Comparative Analysis of Deterministic and Nondeterministic Decision Trees for Decision Tables from Closed Classes
Previous Article in Special Issue
Improving Existing Segmentators Performance with Zero-Shot Segmentators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Possibilistic Formulation of Autonomous Search for Targets

School of Engineering, RMIT University, 376-392 Swanston Street, Melbourne, VIC 3000, Australia
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(6), 520; https://doi.org/10.3390/e26060520
Submission received: 21 May 2024 / Revised: 13 June 2024 / Accepted: 14 June 2024 / Published: 17 June 2024
(This article belongs to the Special Issue Advances in Uncertain Information Fusion)

Abstract

:
Autonomous search is an ongoing cycle of sensing, statistical estimation, and motion control with the objective to find and localise targets in a designated search area. Traditionally, the theoretical framework for autonomous search combines sequential Bayesian estimation with information theoretic motion control. This paper formulates autonomous search in the framework of possibility theory. Although the possibilistic formulation is slightly more involved than the traditional method, it provides a means for quantitative modelling and reasoning in the presence of epistemic uncertainty. This feature is demonstrated in the paper in the context of partially known probability of detection, expressed as an interval value. The paper presents an elegant Bayes-like solution to sequential estimation, with the reward function for motion control defined to take into account the epistemic uncertainty. The advantages of the proposed search algorithm are demonstrated by numerical simulations.

1. Introduction

Search is a repetitive cycle of sensing, estimation (localisation), and motion control, with the objective to find and localise one, or as many as possible targets, inside the search volume, in the shortest possible time. The searching platform (agent) is assumed to be mobile and capable of sensing, while the detection process is typically imperfect [1], in the sense that the probability of detection is less than one, with a small (but non-negligible) probability of false alarms. Autonomous search refers to the search by an intelligent agent without human intervention.
Search techniques have been used in many situations. Examples include rescue and recovery operations, security operations (e.g., search for toxic or radioactive emissions), understanding animal behaviour, and military operations (e.g., anti-submarine warfare) [2]. Search techniques have also become increasingly important in field robotics [3,4,5,6] for the purpose of carrying out dirty and dangerous missions. A formal search theory has roots in the works by Koopman [7], and has since been expanded and extended to different problems and applications. It can be categorised into a static versus a moving target search, a reactive versus a non-reactive target search, a single versus multiple target search, a cooperative versus a non-cooperative target search, etc. [2].
In this paper, we focus on an area search for an unknown number of static targets using a realistic sensor on a single searching platform, conceptually similar to the problems discussed in [8,9,10]. The searching agent can be a drone equipped with a sensor capable of detecting targets on the ground with a certain probability of detection (as a function of range) as well with some false alarm probability.
The dominant theoretical framework for the formulation of search is probability theory, where Bayesian inference is used to sequentially update the posterior probability distribution of target locations as new measurements are collected over time [9,11,12,13,14]. Sensor motion control is typically formulated as a partially observed Markov decision process (POMDP) [15]. The information state in the POMDP formulation is represented by the posterior probability distribution of targets. The set of sensor motion controls (actions), which determine where the searching agent should move next, can be made a single step or multiple steps ahead. The reward function in POMDP maps the set of admissible actions to a set of positive real numbers (rewards) and is typically formulated as a measure of information gain (e.g., the reduction in entropy, Fisher information gain) [16].
Statistical inference is based on mathematical models. In the target search context, we need a model of sensing which incorporates the uncertainty with regards to the probability of true and false detection as well as the statistics of target (positional) measurement errors. This uncertainty in the Bayesian framework is expressed by probability functions—in particular, the probability of detection, the probability of false alarm, and the probability density function (PDF) of a positional measurement, given the true target location. The key limitation of the Bayesian approach, however, is that these probabilistic models must be known precisely. Unfortunately, in many practical situations it is difficult or even impossible to postulate precise probabilistic models. Consider, for example, the probability of detection. It typically depends on the (unknown) size and reflective characteristics of the target, and hence, at best can be specified as a confidence interval (rather than a precise probability value), for a given distance to the target. Thus, we need to deal with epistemic uncertainty, which incorporates both randomness and partial ignorance.
In order to deal with epistemic uncertainty, an alternative mathematical framework for inference is required. Such theories involve non-additive probabilities [17] for the representation and processing of uncertain information. They include, for example, possibility theory [18], Dempster–Shafer theory [19], and imprecise probability theory [20]. Because the last two theories are fairly complicated, and at present applicable only to discrete state spaces, we focus on possibility theory [21,22]. Recent research in nonlinear filtering and target tracking [23,24,25,26,27] has demonstrated that possibility theory provides an effective tool for uncertain knowledge representation and reasoning.
The main contributions of this paper include a theoretical formulation of autonomous search in the framework of possibility theory and a demonstration of its robustness in the presence of epistemic detection uncertainty. The paper presents an elegant Bayes-like solution to sequential estimation, with a definition of the reward function for motion control which takes into account the epistemic uncertainty. Evaluation of the proposed search algorithm considers scenarios with a large number of targets and for two cases for the probability of detection as a function of range: (i) the case when it is known precisely; (ii) the case when it is known only as an interval value.
The paper is organised as follows. Section 2 introduces the autonomous search problem. Section 3 reviews the standard probabilistic formulation of autonomous search and presents the theoretical framework for estimation using possibility functions. Section 4 formulates the new possibilistic solution to autonomous search. Numerical results with a comparison are presented in Section 5, while the conclusions are drawn in Section 6.

2. Problem Formulation

Consider a search area S . A surveillance drone, flying at a fixed altitude, has a mission to autonomously search and localise the ground-based static targets in S , as in [10]. The number and locations of the targets are unknown. Following [9,10], the search area S is discretized into n c 1 cells of equal size. The presence or absence of a target in the nth cell at a discrete time k = 0 , 1 , 2 , can be modelled by a Bernoulli random variable (r.v.) X k , n { 0 , 1 } , where by convention X k , n = 1 denotes that a target is present, (i.e., 0 denotes target absence) and n = 1 , , n c is the cell index.
Suppose the search agent is equipped with a sensor (e.g., a radar) which illuminates a region L k S at time k and collects a set of detections Z k within L k . Each detection reports the Cartesian coordinates of a possible target. However, the sensing process is uncertain in two ways: (1) the reported target coordinates are affected by measurement noise; (2) the measurement set Z k may include false detections and also may miss some of the true target detections. The probability of true target detection is a (monotonically decreasing) function of range and is specified as an interval value for a given range.
The objective is to detect and localise as many targets as possible in the shortest possible time.

3. Background

3.1. Probabilistic Search

Autonomous search in the Bayesian probabilistic framework is typically information driven. The information state at time k is represented by the posterior probability of target presence in each cell of the discretised search area. This posterior probability at time k is denoted by P k , n = P r { X k , n = 1 | Z 1 : k } , where Z 1 : k : = Z 1 , , Z k is the sequence of measurement sets up to the current time k. The posterior probability of target absence is then simply P ¯ k , n = P r { X k , n = 0 | Z 1 : k } = 1 P k , n , and therefore, is unnecessary to compute.
The target or threat map is defined as the array P k = [ P k , n ] . Initially, at time k = 0 , the map is specified as P 0 , n = 1 2 , for all n = 1 , , n c , thus expressing the initial ignorance. As time progresses and the search agent collects measurements, the threat map is sequentially updated using Bayes’s rule. Consequently, the information content of the threat map P k increases with time. The information content of the threat map is measured by its entropy, defined as
H k = 1 n c n = 1 n c P k , n log 2 P n , k + ( 1 P n , k ) log 2 ( 1 P n , k ) .
Note that at k = 0 , H 0 = 1 and that entropy decreases with time.
In order to explain how the threat map is updated using Bayes’s rule, let us introduce another Bernoulli r.v. Y n , k { 0 , 1 } , where Y n , k = 1 represents the event that a detection from the set Z k has fallen inside the nth cell ( Y n , k = 0 represents the opposite event). Bayes’s rule is given by
P r { X = i | Y = j } = P r { Y = j | X = i } P r { X = i } = 0 , 1 P r { Y = j | X = } P r { X = }
where the subscript ( n , k ) is temporarily removed from X n , k and Y n , k in order to simplify the notation, and i , j { 0 , 1 } .
Note that P r { Y = 1 | X = 1 } = D and P r { Y = 1 | X = 0 } = F represent the probability of detection and the probability of false alarm, respectively. Then, P r { Y = 0 | X = 1 } = 1 D and P r { Y = 0 | X = 0 } = 1 F .
Given P k 1 , if none of the detections in Z k falls into the nth cell (i.e., Y n , k = 0 ), the probability P n , k is updated according to (2), as
P k , n = ( 1 D k , n ) P k 1 , n ( 1 D k , n ) P k 1 , n + ( 1 F k , n ) ( 1 P k 1 , n )
where D k , n is the probability of detection and F k , n is the probability of false alarm in the nth cell of search area S at time k.
If Z k contains a detection in the nth cell (i.e., Y n , k = 1 ), then the update equation according to (2) is
P k , n = D k , n P k 1 , n D k , n P k 1 , n + F k , n ( 1 P k 1 , n )
After collecting the measurement set Z k 1 , the searching agent must decide on its subsequent action, that is, where to move (and sense) next. Suppose the set of possible actions (for movement) is A k . This set can be formed by considering one or more motion steps ahead (in the future). The reward function associated with every action α A k is typically defined as the reduction in entropy of the threat map [10], that is,
R k ( α ) = H k 1 E { H k ( α ) }
Note the expectation operator E with respect to the (future) detection set Z k ( α ) . In practical implementation, in order to simplify computation, we typically adopt an approximation that circumvents E in (5). This approximation involves the assumption that a single realisation for Z k ( α ) is sufficient: the one which results in hypothetical detection(s) at those cells which are characterised by a high probability of target presence, i.e., such that P k 1 , n > ζ , where ζ is a threshold close to 1. The searching agent chooses the action which maximises the reward, i.e.,
α k * = arg max α A k R k ( α ) .

3.2. The Possibilistic Estimation Framework

Possibility theory is developed for quantitative modelling of epistemic uncertainty. The concept of the uncertain variable in possibility theory, plays the same role as the random variable in probability theory. The main difference is that the quantity of interest is not random, but simply unknown, and our aim is to infer its true value out of a set of possible values. The theoretical basis of this approach can be found in [28,29,30]. Briefly, the uncertain variable is a function X : Ω X , where Ω is the sample space and X is the state space (the space where the quantity of interest lives). Our current knowledge about X can be encoded in a function π X : X 0 , 1 , such that π X ( x ) is the possibility (credibility) for the event X = x . Function π X is not a density function, it is referred to as a possibility function, being the primitive object of possibility theory [22]. It can be viewed as a membership function determining the fuzzy restrictions of minimal specificity (in the sense that any hypothesis not known to be impossible cannot be ruled out) about x [18]. Normalisation of π X is sup x X π X ( x ) = 1 if X is uncountable, and max x X π X ( x ) = 1 if X is finite and countable.
In the formulation of the search problem, we will deal with two binary uncertain variables, corresponding to r.v.s X k , n and Y k , n . Hence, let us focus on a discrete uncertain variable X and its state space X = { x 1 , , x N } . The possibility measure of an event A X is a mapping Π X : 2 X [ 0 , 1 ] , where 2 X is the set of all subsets of X . Mapping Π X satisfies three axioms: (1) Π X ( ) = 0 ; (2) Π X ( X ) = 1 ; and (3) the possibility of a union of disjoint events A 1 and A 2 is given by Π X ( A 1 A 2 ) = max [ Π X ( A 1 ) , Π X ( A 2 ) ] . Possibility measure Π X is related to the possibility function π X as follows:
Π X ( A ) = max x A π X ( x )
for every A X . There is also a notion of the necessity of an event N X ( A ) , which is dual to Π X ( A ) in the sense that
N X ( A ) = 1 Π X ( A c ) ,
where A c is the complement of A in X . One can interpret the necessity–possibility interval [ N X ( A ) , Π X ( A ) ] as the belief interval, specified by the lower and upper probabilities in the sense of Willey [20]. Note that for a binary variable X { 0 , 1 } , this interval can be expressed for event A = { 1 } as P r { X = 1 } [ N X ( 1 ) , Π X ( 1 ) ] = [ 1 Π X ( 0 ) , Π X ( 1 ) ] , where, due to normalisation, the following condition must be satisfied: max { Π X ( 0 ) , Π X ( 1 ) } = 1 .
Bayes-like updating in possibility theory is described next. Suppose π ( x ) is the prior possibility function over the state space X = { x 1 , , x N } . Let γ ( z | x ) be the likelihood of receiving measurement z Z if x X is true. Then, the posterior possibility of x X is given by [28,31,32]
π ( x | z ) = γ ( z | x ) π ( x ) max x X [ γ ( z | x ) π ( x ) ] .

4. Theoretical Formulation of Possibilistic Search

4.1. Information State

The information state at time k in the framework of possibility theory will be represented by two posteriors:
1.
The posterior possibility of target presence Π k , n 1 = Π X k , n ( { 1 } | Z 1 : k ) ;
2.
The posterior probability of target absence Π k , n 0 = Π X k , n ( { 0 } | Z 1 : k ) .
We need both of them, because Π k , n 0 cannot be worked out from Π k , n 1 . Consequently, during the search two posterior possibility maps need to be updated sequentially over time, Π k 1 = [ Π k , n 1 ] and Π k 0 = [ Π k , n 0 ] , where n = 1 , , n c .
Suppose now that the probability of detection is specified by an interval value, that is,
D k , n [ D ̲ k , n , D ¯ k , n ]
where D ̲ k , n and D ¯ k , n represent the lower and upper probability of this interval, respectively. Because a detection event is a binary variable, due to the reachability constraint for probability intervals [33], (9) implies that the probability of non-detection is in interval [ 1 D ¯ k , n , 1 D ̲ k , n ] . Then, via normalisation we can express the possibility of detection D k , n 1 and the possibility of non-detection D k , n 0 (in cell n at time k) as
D k , n 1 = D ¯ k , n max { 1 D ̲ k , n , D ¯ k , n }
D k , n 0 = 1 D ̲ k , n max { 1 D ̲ k , n , D ¯ k , n } .
satisfying max { D k , n 0 , D k , n 1 } = 1 . Interval [ 1 D k , n 0 , D k , n 1 ] represents the necessity–possibility interval for the probability of detection. Note that specification of a possibility function from a probability mass function expressed by probability intervals is not unique; for example, another more involved method for this task is via the maximum specificity criterion [34].
In general, the probability of detection D k , n by a sensor, as well as the two possibilities D k , n 0 and D k , n 1 , are typically dependent on the distance d n , k between the nth grid cell and the searching agent’s position at time k.
In a similar manner, we can also assume that the probability of false alarm is specified by an interval value, that is, F k , n [ 1 F k , n 0 , F k , n 1 ] , where F k , n 0 and F k , n 1 represent the possibility of no false alarm and the possibility of false alarm (in cell n at time k), respectively.
Next, we explain how to sequentially update, during the search, the two posterior possibility maps Π k 1 (for target presence) and Π k 0 (for target absence). The proposed update equations follow from (3) and (4), when we apply the Bayes-like update rule (8).
Given Π k 1 1 and detection set Z k , if none of the detections in Z k falls into the nth cell, the possibility of target presence in the nth cell is updated as follows:
Π k , n 1 = D k , n 0 Π k 1 , n 1 max { D k , n 0 Π k 1 , n 1 , F k , n 0 Π k 1 , n 0 } ,
for n = 1 , , n c . Similarly, in this case Π k , n 0 is updated according to
Π k , n 0 = F k , n 0 Π k 1 , n 0 max { D k , n 0 Π k 1 , n 1 , F k , n 0 Π k 1 , n 0 } .
If a detection from Z k falls into the nth cell, then the update equation for Π k , n 1 can be expressed as
Π k , n 1 = D k , n 1 Π k 1 , n 1 max { D k , n 1 Π k 1 , n 1 , F k , n 1 Π k 1 , n 0 } .
And finally, in this case the update equation for Π k , n 0 is given by
Π k , n 0 = F k , n 1 Π k 1 , n 0 max { D k , n 1 Π k 1 , n 1 , F k , n 1 Π k 1 , n 0 }
Note that the probability of target presence in each cell of the search area, using the described possibilistic approach, is expressed by a necessity–possibility interval, i.e.,
P k , n [ 1 Π k , n 0 , Π k , n 1 ]
for n = 1 , , n c , where max { Π k , n 0 , Π k , n 1 } = 1 . Initially, at time k = 0 (before any sensing action), the posterior possibility maps are set to
Π 0 , n 0 = Π 0 , n 1 = 1 ,
meaning that P 0 , n [ 0 , 1 ] , for n = 1 , , n c . This is an expression of initial ignorance about the probability of target presence in the nth cell.

4.2. Epistemic Reward

Let us first quantify the amount of uncertainty contained in the information state, represented by two posterior possibility maps: Π k 1 and Π k 0 . Various uncertainty (and information) measures in the context of non-additive probabilistic frameworks have been proposed in the past [35,36,37]. We adopt the principle that epistemic uncertainty corresponds to the volume under the possibility function [25,37]. For a possibility function π over a discrete finite state space X = { x 1 , , x N } , epistemic uncertainty equals the sum i = 1 N π ( x i ) . The possibilistic entropy G k , contained in the information state, represented by Π k 1 and Π k 0 , is then defined as
G k = 1 n c n = 1 n c [ Π k , n 1 + Π k , n 0 ] 1
Equation (18) can be interpreted as the average volume of possibility functions of all binary variables X n , k , for n = 1 , , n c . Subtraction by 1 on the right-hand side of (18) ensures that G k [ 0 , 1 ] . Thus, at k = 0 , when Π 0 , n 0 = Π 0 , n 1 = 1 , we have G 0 = 1 . This means that initially (at the start of the search), the amount of information contained in the information state is zero (representing total ignorance). As the searching agent moves and collects measurements it gains knowledge, and as a result either Π k , n 0 or Π k , n 1 will reduce its value in some cells (keeping in mind that max { Π k , n 0 , Π k , n 1 } = 1 ), thus reducing the possibilistic entropy G k . Finally, G k = 0 if either Π k , n 0 = 0 (and due to normalisation Π k , n 1 = 1 ) or Π k , n 1 = 0 (and Π k , n 0 = 1 ) for all cells n = 1 , , n c .
Note that (18) can also be expressed as
G k = 1 n c n = 1 n c [ Π k , n 1 ( 1 Π k , n 0 ) ]
which gives another interpretation of possibilistic entropy G k : it represents the average necessity–possibility interval over all cells in the search area. This interpretation does not mean that G k is a measure of uncertainty only due to imprecision, because (18) and (19) are equivalent.
Similar to (5), we define the reward function as the reduction in possibilistic entropy of the information state, expressed by maps Π k 1 and Π k 0 . Mathematically, this is expressed as
R k ( α ) = G k 1 E { G k ( α ) }
where, as before, α A k is an action from the set of admissible actions at time k and E is the expectation with respect to the (random) measurement set Z k ( α ) . Again, in order to simplify the computation, we make the same assumption described in relation to (5): a single realisation for Z k ( α ) consisting of hypothetical detection(s) at those cells which are characterised by Π k 1 , n 1 Π k 1 , n 0 > ζ . Finally, the searching agent chooses the action which maximises the reward, as in (6).
The search mission is terminated when the reduction in possibilistic entropy falls below a specified threshold, i.e., when G k 1 G k < ξ .

5. Numerical Results

5.1. Simulation Setup and a Single Run

We use a simulation setup similar to [10]. The search area S is a rectangle of size 100 km × 90 km, discretised into n c = 100 × 90 resolution cells of size 1 km2. A total of 80 targets are placed at (a) uniformly random locations across the search area; (b) two squares in diagonal corners of the search area. A typical scenario with a uniform distribution of targets is shown in Figure 1, where cyan coloured asterisks indicate where the targets are placed.
The probability of detection D is modelled as a function of the distance between the nth grid cell and the searching agent’s position at time k. The following mathematical model is adopted for this purpose:
D ( d ; μ , σ ) = 1 1 σ 2 π d e ( t μ ) 2 2 σ 2 d t
where d 0 is the distance, while μ > 0 and σ > 0 are modelling parameters. Figure 2 illustrates this model; it displays the imprecise model of the probability of detection D as a function of distance d, using (21) with two sets of parameters μ and σ (the orange-coloured area). The search algorithm described in Section 4 is using this imprecise model for its search mission. The model provides the upper and lower probabilities D ̲ k , n and D ¯ k , n for a given range, from which we can work out D k , n 1 and D k , n 0 , via (10) and (11), respectively. The true value of the probability of detection, which is used in the generation of simulated measurements (but which is unknown to the search algorithm), is plotted with the solid blue line in Figure 2. The truth is also based on model (21), using one particular pair of μ and σ values (The actual values used for the orange-coloured area in Figure 2 are μ 1 = 8000 , σ 1 = 2200 , μ 2 = 18,000, and σ 2 = 2200 . The true probability of detection (blue line in Figure 2) is obtained using μ = 9000 and σ = 2200 ).
With this specification, the probability of detecting a target located more than a certain distance ρ max from the searching agent is practically zero. Assuming 360 coverage, the sensing area L k is a circular area of radius ρ max . The spatial distribution of false alarms is assumed to be uniform over L k , with probability F k , n = 0.005 (per cell of L k ). For simplicity, we will assume that this parameter is known as the precise value to the search algorithm of Section 4. The threshold parameter ζ is set to 0.8 .
Sensor measurements are affected by additive Gaussian noise with the standard deviation in range and azimuth of 100 m and 1 , respectively. An additional assumption is that there is at most one target per cell and one detection per cell.
The searching agent’s motion is modelled by the coordinated turn (CT) model [38], with the turning rate taking values from the set
Ψ = { 0.4 , 0.3 , 0.2 , 1 , 0 , 0.1 , 0.2 , 0.3 , 0.4 }
(the units are °/s). We consider one-step ahead path planning, with action space A k defined as a Cartesian product A k = Ψ × Δ . Here, Δ is the set of time intervals of CT motion (with the selected turning rate), adopted as Δ = { 60 , 120 } seconds.
The results of a single run of the possibilistic search at time k = 140 for a uniform placement of targets is shown in Figure 1, Figure 3, and Figure 4. Figure 1 displays the search path (blue dotted line). The searching agent enters the search area S in the bottom left corner, and follows an inward-spiral path, in accordance with the probabilistic search [10]. Figure 3 shows the two posterior possibilistic maps: (a) target presence Π k 1 ; and (b) target absence Π k 0 . The colour coding is as follows: white cells of the maps indicate zero possibility, while black cells denote the possibility is equal to 1. Figure 3a indicates that the area around the travelled path in Π k 1 is mainly white, with occasional black cells where targets are possibly located. In those cells of the search area S where Π k 1 is high (black colour) and Π k 0 is low (white colour), there is a high chance that a target is placed. Therefore, the presence of a target in each cell of the search area is declared if the difference Π k , n 1 Π k , n 0 > 0.8 .
The output of the search algorithm at k = 140 is shown in Figure 4, which represents a map of estimated target positions: each red asterisk indicates a cell where the search algorithm declared a target. We can visually compare Figure 4 (estimated target positions at k = 140 ) with Figure 1 (true target positions).
If the search were to be continued beyond k = 140 , the full spiral path would be completed at about k = 200 (for an average run). After that, the rate of reduction in possibilistic entropy would significantly drop and the search algorithm would automatically stop (according to the termination criterion).

5.2. Monte Carlo Runs

Next, we compare the average search performance of the possibilistic search versus the probabilistic search. The adopted metric for search performance is the optimal sub-pattern assignment (OSPA) error, because it expresses in a mathematically rigorous manner the error both in the target position estimate and in the target number (cardinality error) [39]. The parameters of OSPA error used are cut-off c = 50 km and order p = 1 . The mean OSPA error is estimated by averaging over 100 Monte Carlo runs, with a random placement of targets on every run. Because the search duration is random, for the sake of averaging the OSPA error, we fixed the duration to k = 201 time steps.
In order to apply the probabilistic search for the problem specified in Section 5.1, we must adopt a precise (rather than an interval-valued) probability of detection. For comparison’s sake, we will consider two cases: (a) when the true probability of detection versus range (i.e., the blue line in Figure 2) is known; (b) given the interval-valued probability of detection (orange area in Figure 2), we choose the mid-point of the interval at a given range as the true value. Case (a) is ideal and is expected to result in the best performance, whereas case (b), because it uses an incorrect value of the probability of detection, is expected to perform worse.
The resulting three mean OSPA errors are presented in Figure 5 for two different target placements: (i) uniformly random target locations across the search area; (ii) random placement in two squares positioned in diagonal corners of the search area. The mean OSPA line colours in Figure 5 are as follows: black for possibilistic search; blue for probabilistic using true D (i.e., ideal case (a) above); red for probabilistic using wrong D (i.e., case (b)). All three mean OSPA error curves follow the same trend: they reduce steadily from the initial value of c as the searching agent traverses the area along the spiral path and discovers the targets. Of the three compared methods, as expected, the best performance (i.e., the smallest OSPA error) is achieved using the probabilistic with true D (ideal case). The possibilistic solution, which operates using the available interval-valued probability of detection, is fairly close to the ideal case. Finally, the probabilistic using the wrong value of D is the worst. The difference in performance is particularly dramatic when the placement of targets is non-uniform.

6. Conclusions

This paper formulated a solution to autonomous search for targets in the framework of possibility theory. The main rationale for the possibilistic formulation is its ability to deal with epistemic uncertainty, expressed by partially known probabilistic models. In this paper, we focused on the interval-valued probability of detection (as a function of range). The paper presented Bayes-like update equations for the information state in the possibilistic framework, as well as an epistemic reward function for motion control. The numerical results demonstrated that the proposed possibilistic formulation of search can deal effectively with epistemic uncertainty in the form of interval-valued probability of detection. As expected, the (conventional) probabilistic solution performs (sightly) better when the correct precise model of the probability of detection is known (the ideal model-match case). However, the probabilistic solution can result in dramatically worse performance if an incorrect precise model is adopted.

Author Contributions

Conceptualisation, B.R.; investigation, Z.C.; methodology, Z.C., B.R. and D.Y.K.; software, Z.C.; supervision, B.R. and D.Y.K.; validation, Z.C.; writing—original draft, Z.C.; writing—review and editing, B.R. and D.Y.K. All authors have read and agreed to the published version of the manuscript.

Funding

Zhijin Chen is supported by an Australian Government Research Training Program (RTP) Scholarship.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shlesinger, M.F. Search research. Nature 2006, 443, 281–282. [Google Scholar] [CrossRef] [PubMed]
  2. Stone, L.D.; Royset, J.O.; Washburn, A.R. Optimal Search for Moving Targets; Springer: Cham, Switzerland, 2016. [Google Scholar]
  3. Marjovi, A.; Marques, L. Multi-robot olfactory search in structured environment. Robot. Auton. Syst. 2011, 59, 867–881. [Google Scholar] [CrossRef]
  4. Pitre, R.R.; Li, X.R.; Delbalzo, R. UAV route planning for joint search and track missions—An information-value approach. IEEE Trans. Aerosp. Electron. Syst. 2012, 48, 2551–2565. [Google Scholar] [CrossRef]
  5. Burgués, J.; Marco, S. Environmental chemical sensing using small drones: A review. Sci. Total Environ. 2020, 748, 141172. [Google Scholar] [CrossRef] [PubMed]
  6. Park, M.; An, S.; Seo, J.; Oh, H. Autonomous source search for UAVs using Gaussian mixture model-based infotaxis: Algorithm and flight experiments. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 4238–4254. [Google Scholar] [CrossRef]
  7. Koopman, B.O. Search and screening. OEG Rep. 1946. Available online: https://www.loc.gov/resource/gdcmassbookdig.searchscreening56koop/?st=gallery (accessed on 13 June 2024).
  8. Bakut, P.A.; Zhulina, Y.V. Optimal control of cell scanning sequence in search for objects. Eng. Cybern. 1971, 9, 740–746. [Google Scholar]
  9. Krout, D.W.; Fox, W.L.J.; El-Sharkawi, M.A. Probability of target presence for multistatic sonar ping sequencing. IEEE J. Ocean. Eng. 2009, 34, 603–609. [Google Scholar] [CrossRef]
  10. Angley, D.; Ristic, B.; Moran, W.; Himed, B. Search for targets in a risky environment using multi-objective optimisation. IET Radar Sonar Navig. 2019, 13, 123–127. [Google Scholar] [CrossRef]
  11. Kendall, R.J.; Presley, S.M.; Austin, G.P.; Smith, P.N. Advances in Biological and Chemical Terrorism Countermeasures; CRC Press: Boca Raton, FL, USA, 2008. [Google Scholar]
  12. Furukawa, T.; Mak, L.C.; Durrant-Whyte, H.; Madhavan, R. Autonomous Bayesian search and tracking, and its experimental validation. Adv. Robot. 2012, 26, 461–485. [Google Scholar] [CrossRef]
  13. Ristic, B.; Skvortsov, A.; Gunatilaka, A. A study of cognitive strategies for an autonomous search. Inf. Fusion 2016, 28, 1–9. [Google Scholar] [CrossRef]
  14. Haley, K. Search Theory and Applications; Springer Science & Business Media: New York, NY, USA, 2012; Volume 8. [Google Scholar]
  15. Chong, E.K.P.; Kreucher, C.M.; Hero, A.O. POMDP approximation using simulation and heuristics. In Foundations and Applications of Sensor Management; Springer: Cham, Switzerland, 2008; pp. 95–119. [Google Scholar]
  16. Hero, A.O., III; Kreucher, C.M.; Blatt, D. Information theoretic approaches to sensor management. In Foundations and Applications of Sensor Management; Springer: Boston, MA, USA, 2008; pp. 33–57. [Google Scholar]
  17. Hampel, F. Nonadditive probabilities in statistics. J. Stat. Theory Pract. 2009, 3, 11–23. [Google Scholar] [CrossRef]
  18. Zadeh, L.A. Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets Syst. 1978, 1, 3–28. [Google Scholar] [CrossRef]
  19. Yager, R.R.; Liu, L. Classic Works of the Dempster-Shafer Theory of Belief Functions; Springer: Cham, Switzerland, 2008; Volume 219. [Google Scholar]
  20. Walley, P. Statistical Reasoning with Imprecise Probabilities; Springer: Cham, Switzerland, 1991; Volume 42. [Google Scholar]
  21. Dubois, D.; Prade, H.; Sandri, S. On possibility/probability transformations. In Fuzzy Logic: State of the Art; Springer: Dordrecht, The Netherlands, 1993; pp. 103–112. [Google Scholar]
  22. Dubois, D.; Prade, H. Possibility theory and its applications: Where do we stand? In Springer Handbook of Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2015; pp. 31–60. [Google Scholar]
  23. Ristic, B.; Houssineau, J.; Arulampalam, S. Robust target motion analysis using the possibility particle filter. IET Radar Sonar Navig. 2019, 13, 18–22. [Google Scholar] [CrossRef]
  24. Ristic, B.; Houssineau, J.; Arulampalam, S. Target tracking in the framework of possibility theory: The possibilistic Bernoulli filter. Inf. Fusion 2020, 62, 81–88. [Google Scholar] [CrossRef]
  25. Chen, Z.; Ristic, B.; Houssineau, J.; Kim, D.Y. Observer control for bearings-only tracking using possibility functions. Automatica 2021, 133, 109888. [Google Scholar] [CrossRef]
  26. Houssineau, J.; Zeng, J.; Jasra, A. Uncertainty modelling and computational aspects of data association. Stat. Comput. 2021, 31, 1–19. [Google Scholar] [CrossRef]
  27. Ristic, B. Target tracking in the framework of possibility theory. ISIF Perspect. Inf. Fusion 2023, 6, 47–48. [Google Scholar]
  28. Houssineau, J.; Bishop, A.N. Smoothing and filtering with a class of outer measures. SIAM/ASA J. Uncertain. Quantif. 2018, 6, 845–866. [Google Scholar] [CrossRef]
  29. Bishop, A.N.; Houssineau, J.; Angley, D.; Ristic, B. Spatio-temporal tracking from natural language statements using outer probability theory. Inf. Sci. 2018, 463, 56–74. [Google Scholar] [CrossRef]
  30. Houssineau, J. A linear algorithm for multi-target tracking in the context of possibility theory. IEEE Trans. Signal Process. 2021, 69, 2740–2751. [Google Scholar] [CrossRef]
  31. Boughanem, M.; Brini, A.; Dubois, D. Possibilistic networks for information retrieval. Int. J. Approx. Reason. 2009, 50, 957–968. [Google Scholar] [CrossRef]
  32. Ristic, B.; Gilliam, C.; Byrne, M.; Benavoli, A. A tutorial on uncertainty modeling for machine reasoning. Inf. Fusion 2020, 55, 30–44. [Google Scholar] [CrossRef]
  33. Campos, L.M.D.; Huete, J.F.; Moral, S. Probability intervals: A tool for uncertain reasoning. Intern. J. Uncertain. Fuzziness-Knowl.-Based Syst. 1994, 2, 167–196. [Google Scholar] [CrossRef]
  34. Masson, M.-H.; Denoeux, T. Inferring a possibility distribution from empirical data. Fuzzy Sets Syst. 2006, 157, 319–340. [Google Scholar] [CrossRef]
  35. Klir, G.J.; Smith, R.M. On measuring uncertainty and uncertainty-based information: Recent developments. Ann. Math. Artif. Intell. 2001, 32, 5–33. [Google Scholar] [CrossRef]
  36. Abellán, J. Uncertainty measures on probability intervals from the imprecise dirichlet model. Intern. J. Gen. Syst. 2006, 35, 509–528. [Google Scholar] [CrossRef]
  37. Pota, M.; Esposito, M.; Pietro, G.D. Transforming probability distributions into membership functions of fuzzy classes: A hypothesis test approach. Fuzzy Sets Syst. 2013, 233, 52–73. [Google Scholar] [CrossRef]
  38. Bar-Shalom, Y.; Li, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation: Theory Algorithms and Software; John Wiley & Sons: New York, NY, USA, 2004. [Google Scholar]
  39. Schuhmacher, D.; Vo, B.-T.; Vo, B.-N. A consistent metric for performance evaluation of multi-object filters. IEEE Trans. Signal Process. 2008, 56, 3447–3457. [Google Scholar] [CrossRef]
Figure 1. Simulation setup: the cyan stars indicate the true targets; the blue dotted line is the trajectory of the searching agent up to k = 140 steps; the red dots indicate detections at k = 140 .
Figure 1. Simulation setup: the cyan stars indicate the true targets; the blue dotted line is the trajectory of the searching agent up to k = 140 steps; the red dots indicate detections at k = 140 .
Entropy 26 00520 g001
Figure 2. The imprecise model of the probability of detection D used in simulations. The true D is plotted with the blue solid line.
Figure 2. The imprecise model of the probability of detection D used in simulations. The true D is plotted with the blue solid line.
Entropy 26 00520 g002
Figure 3. Posterior possibility maps at time k = 140 : (a) Target presence map Π k 1 ; (b) target absence map Π k 0 (white colour implies zero possibility).
Figure 3. Posterior possibility maps at time k = 140 : (a) Target presence map Π k 1 ; (b) target absence map Π k 0 (white colour implies zero possibility).
Entropy 26 00520 g003
Figure 4. Output of the search algorithm: Estimated target locations (indicated by red asterisks).
Figure 4. Output of the search algorithm: Estimated target locations (indicated by red asterisks).
Entropy 26 00520 g004
Figure 5. Mean OSPA errors obtained from 100 Monte Carlo runs. The scenario involves 80 targets placed at (a) uniformly random locations; (b) uniformly random locations of two diagonal squares in the search area.
Figure 5. Mean OSPA errors obtained from 100 Monte Carlo runs. The scenario involves 80 targets placed at (a) uniformly random locations; (b) uniformly random locations of two diagonal squares in the search area.
Entropy 26 00520 g005aEntropy 26 00520 g005b
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Z.; Ristic, B.; Kim, D.Y. A Possibilistic Formulation of Autonomous Search for Targets. Entropy 2024, 26, 520. https://doi.org/10.3390/e26060520

AMA Style

Chen Z, Ristic B, Kim DY. A Possibilistic Formulation of Autonomous Search for Targets. Entropy. 2024; 26(6):520. https://doi.org/10.3390/e26060520

Chicago/Turabian Style

Chen, Zhijin, Branko Ristic, and Du Yong Kim. 2024. "A Possibilistic Formulation of Autonomous Search for Targets" Entropy 26, no. 6: 520. https://doi.org/10.3390/e26060520

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop