Next Article in Journal
CTHNet: A CNN–Transformer Hybrid Network for Landslide Identification in Loess Plateau Regions Using High-Resolution Remote Sensing Images
Previous Article in Journal
Multi-Objective Optimization for Resource Allocation in Space–Air–Ground Network with Diverse IoT Devices
Previous Article in Special Issue
Activities of Daily Living Object Dataset: Advancing Assistive Robotic Manipulation with a Tailored Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficacy of Segmentation for Hyperspectral Target Detection

by
Yoram Furth
* and
Stanley R. Rotman
Department of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer Sheva blvd 1, Beer-Sheva 84105, Israel
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(1), 272; https://doi.org/10.3390/s25010272
Submission received: 4 December 2024 / Revised: 1 January 2025 / Accepted: 2 January 2025 / Published: 6 January 2025
(This article belongs to the Special Issue Vision Sensors for Object Detection and Tracking)

Abstract

:
Algorithms for detecting point targets in hyperspectral imaging commonly employ the spectral inverse covariance matrix to whiten inherent image noise. Since data cubes often lack stationarity, segmentation appears to be an attractive preprocessing operation. Surprisingly, the literature reports both successful and unsuccessful segmentation cases, with no clear explanations for these divergent outcomes. This paper elucidates the conditions under which segmentation might improve detector performance. Focusing on a representative algorithm and assuming a target additive model, the study examines all influential factors through theoretical analysis and extensive simulations. The findings offer fundamental insights and practical guidelines for characterizing segmented datasets, enabling a thorough evaluation of segmentation’s utility for detector performance. They outline the range of target scenarios and parameters where segmentation may prove beneficial and help assess the potential impact of proposed segmentation strategies on detection outcomes.

1. Introduction

Assuming the presence of a subpixel target within a hyperspectral image, traditional algorithms frequently fail in detection despite matching the target spectral signature. These algorithms typically employ the inverse covariance matrix for noise reduction. Our inquiry is directed towards elucidating the specific conditions under which using image segmentation can enhance the detector performance.
Historically, there have been conflicting reports on how beneficial segmentation is for detection. For instance, the generally positive results of [1,2,3] contrast with the negative findings reported by Pieper et al. [4]. These inconsistencies raised questions about the true value of segmentation. However, recent studies [5,6,7] have demonstrated that segmentation preprocessing can significantly improve detection performance, emphasizing that its success often depends on specific methods and conditions. These findings, combined with the earlier conflicting results, highlight the need for a deeper understanding of the factors that influence segmentation’s efficacy.
Focusing on the Normalized Segmented Matched Filter (NSMF) with a target additive model, this study explores segmentation’s role from foundational concepts to practical applications. It determines the conditions under which segmentation is likely to enhance detection efficacy for a given algorithm and data properties. We demonstrate how various parameters—such as the target’s strength, spectral signature, and the acceptable false alarm rate—impact segmentation’s contribution. Building on these insights, we aim to establish a predictive framework that outlines, for a given background, in which range of parameters segmentation is likely to improve detection.
In a previous study, we introduced the effect of the target’s spectral signature on segmentation efficacy [8]. Here, we extend this analysis to consider all relevant factors influencing segmentation’s efficacy in detection. We propose that the observed lack of significant improvement in some cases arises from intrinsic data properties and other detector-related factors. This study provides practical guidelines to predict what possible benefit can be gained if the detector uses the segmented data.
The introductory section outlines the foundational principles. Section 2 identifies key rules that explain how each factor simultaneously influences segmentation efficacy. Section 3 presents simulations and various tests on a standard dataset. Finally, Section 4 presents practical conclusions regarding the conditions under which segmentation improves detection efficacy.

1.1. Matched-Filter Target Detection

A standard algorithm for detecting small targets is the Matched Filter (MF), chosen in this study as the representative method for evaluating segmentation efficacy due to its widespread use and reliance on covariance-based whitening. Other algorithms, such as the Adaptive Cosine Estimator (ACE), also depend on covariance properties, suggesting that the findings here will generalize to ACE and similar methods. Ongoing research is testing this hypothesis.
Assuming the background data fit a multivariate normal distribution (MVN) once the background power has been mitigated and assuming an additive target model, it is essentially defined by t T Φ 1 x m . For each pixel vector x , it subtracts the background m , whitens the residual by the inverse covariance matrix Φ, and matches the result to the desired target signature t . Using the Generalized Likelihood Ratio Test (GLRT), m and Φ are estimated from the data itself [9]. This provides a score to each pixel, related to the probability it contains a target. The target is then detected by testing the hypothesis that x contains only “background” against the hypothesis that it contains a target. In this study, we assume an additive target model of x x   + p t , where p is the target power.
Consider the hypothesis that there is no target present in the dataset. While collecting statistics of such scores “without target” being present can be conducted directly on the original dataset (due to the overwhelming number of no-target pixels), robust evaluation for the “with target” hypothesis becomes possible using an implantation procedure, as explained in [10]. This procedure involves implanting desirable targets alternately into the dataset, enabling systematic analysis of algorithm performance across a range of controlled target-to-background interactions.
The distributions of scores for the two hypotheses (“with target” and “without target”) typically overlap, making it challenging to define an optimal detection threshold that balances detection rates with false alarms. The ROC curve is frequently used to optimize this threshold, as it provides a graphical representation of the trade-off. Additionally, the ROC curve serves as a metric of algorithm quality, with higher curves indicating greater cumulative detections for a given false alarm rate.
These two domains derive two different thresholds’ notations used in this paper: ( i ) “ η ”, indicating the decision threshold in the distributions’ domain, versus ( i i ) “ t h ”, indicating the number of false alarms in the integral domain, called the error threshold. The relationship between these thresholds is:
t h = η f N z d z ,  
where f N z is the scores’ distribution with no targets ( N ) present.

1.2. Normalized SMF

Since data are often not stationary and the variance varies throughout the data, estimating one global covariance from the whole image might be wrong [1]. On the other hand, calculating it locally is problematic due to insufficient statistics [11]. A common approach to address this challenge is segmentation: dividing the image into distinct areas based on the similarity of clustered points, estimating the covariance for each area individually, and applying the detector locally using the corresponding covariance. This results in two variations of the MF algorithm: a local version and a global version [12]:
S M F t T Φ s 1 x m ,   G M F t T Φ G 1 x m   ,
where SMF and GMF refer to the segmented and the global version, respectively, Φ s and Φ G are the segmented and the global covariance, respectively, and the other parameters are as defined above.
This work focuses on the NSMF version, where each MF is normalized by its standard deviation. Additionally, its superior performance [8,10], it has the advantage of having its t component scaling invariant, as can be shown by rewriting the expression as follows:
N S M F t T Φ s 1 x m t T Φ s 1 t = t ~ T Φ ~ s 1 x m ;   t ~ = t t ,   Φ ~ s = Φ s t ~ T Φ s 1 t ~ .
We can thus examine the target direction effects separately from its power. Therefore, in the sequel, the symbol t will assume a normalized vector.

1.3. Problem Statement and Prior Work

The effectiveness of segmentation in hyperspectral target detection remains a subject of ongoing debate, with outcomes often varying based on the target and data characteristics. For instance, in this study, segmentation applied to the same dataset and algorithm yielded divergent results: one target showed significant performance improvements with a higher NSMF ROC-Curve than the global one (NGMF), while another target showed no improvement despite identical conditions (Figure 1).
These observations suggest that segmentation efficacy is influenced not only by the data structure but also by other critical algorithmic factors. Indeed, the literature reflects this ambiguity: while segmentation sometimes improves detection [1,2,3], in other cases, it does not [4]. This paper builds on the foundational work of Ben Yakar [12] and provides an in-depth analysis of the simultaneous effects of key factors, including the target power ( p ), the covariance matrix ( Φ ), the target spectrum ( t ), and the error threshold ( t h ). A comprehensive report, including mathematical analyses and experimental details, is available in the Supplementary Materials.
Segmentation of hyperspectral datasets has been an active area of research for both classification and target detection processes [13,14,15,16,17], driven by the non-stationary nature of hyperspectral imagery in both spatial and spectral domains. Recent studies closely aligned with our focus have demonstrated the benefits of segmentation preprocessing for subpixel target detection. For example, Liang et al. [5,6] employed segmentation across spectral and spatial regions to detect both real and implanted targets, showing improved detection probabilities (see Figures 7 and 8 in [5]). Similarly, Stalley [7] compared various segmentation methods, including k-means, Gaussian mixture models, and subspace clustering, demonstrating that all methods outperformed non-segmented detection algorithms. These findings underscore the relevance and potential of segmentation preprocessing, further motivating its exploration in this study.

1.4. Evaluation Metrics

To quantify segmentation impact, we use two key metrics. The first metric is A t h , defined as:
A t h 0 t h P D τ d τ 0.5   t h 2 t h 0.5   t h 2 ,
where P D represents the probability of detection, and t h denotes the error threshold. Introduced by Caefer et al. [10], this metric measures the area under the ROC curve as a function of the maximum false alarm rate. It produces a scalar value between 0 and 1, with higher values indicating better average ROC performance.
The second metric is B t h , defined as:
B t h A L t h A G t h ,
where A G and A L refer to A t h for a Matched Filter using a global or segmented estimated covariance matrix, respectively. Originally introduced by Ben-Yakar [12], this metric quantifies the improvement in detection performance attributable to segmentation. Higher values indicate greater benefits provided by segmentation.

1.5. Datacubes

This study employs a progressive reduction and deduction approach, starting from controlled synthetic scenarios and advancing towards fully real-world data. This methodology ensures robust and reliable conclusions applicable to practical hyperspectral target detection. This consists of three types of data, as depicted in Figure 2:
(a)
Synthetic dataset: This dataset consists of two standard Gaussian distributions with two covariance matrices (Φ1, Φ2) in two-dimensional space and a target t with power p Each Gauss represents the spectral cross-section of the second moment of the residual noise (xm) of the corresponding segment and has 3 degrees of freedom: aspect ratio, scaling factor, and rotation angle. The target is represented as a 2D unit vector with two degrees of freedom—its magnitude and its rotation angle—corresponding to the target’s power and spectrum, accordingly. This minimalistic setup efficiently covers the range of possible target-to-background interactions while avoiding the complexities of high-dimensional data. This dataset serves as the fundamental building block for comprehensive synthetic simulations.
(b)
Simple In-house Cube (SIC): This 91-channel dataset contains stationary data with only two distinct areas. Its non-MVN distributions enable realistic simulations, bridging the gap between synthetic and real-world data. By manipulating the covariance matrices of two areas, we can simulate different types of inhomogeneity as interacted with any desirable target. This allows gradually increasing the complexity while retaining control over key parameters. The SIC serves as a baseline for exploring what can be practically deduced in more complex scenarios.
(c)
RIT Cube: This hyperspectral dataset depicts a scene around Cooke City, Montana, provided by the Rochester Institute of Technology (RIT) [18]. It contains highly non-stationary data with a mix of natural and manmade materials. The data were collected using a high-resolution imaging spectrometer under controlled conditions, with details on the camera, lens, and weather provided in Snyder et al. [18]. Real-world data from the RIT Cube closes the gap entirely by addressing the final layer of complexity, including multiple segments with natural spatial variability and diverse noise characteristics. These complexities allow for testing the full applicability of the guidelines and insights derived from the simpler datasets.
In addition, a second standard dataset, the Via Reggio dataset, provided by [19], was tested. This dataset features significantly different noise characteristics and environmental conditions and was used to validate the practical conclusions and guidelines presented in this paper. However, its results are omitted here due to redundancy.
A key aspect of the experimental framework in this study is the implantation of targets into datasets. This approach, described in Section 1.1, allows for systematic evaluation of algorithms under controlled conditions, enabling flexible testing of various target-to-background interactions. Unlike traditional methods relying on fixed, real-world targets, our strategy ensures that the interaction between targets and their backgrounds can be comprehensively analyzed. This methodology has been used and justified by other researchers [20].
For segmentation we utilized the K-Means algorithm [21], chosen for its simplicity and widespread use. This approach was sufficient for comprehensively studying the indirect effects of the various factors influencing segmentation efficacy. As a pixelwise classifier, K-Means does not account for spatial properties, focusing solely on spectral similarity. Considering the trade-off regarding the number of segments [22], we divided the SIC dataset into two segments and the RIT dataset into five segments, as shown in Figure 3.

2. Factors of Influence

This section analyzes how each of the involved factors affects the segmentation success in the order of the algorithm (Equation (2)): it starts with the effect of p , the target power, added to the pixel assuming an additive model. It continues with the effect of Φ , the covariance matrix directly affected by the data inhomogeneity. It continues with the effect of t , the target direction in the spectral domain. It then ends with the effect of t h , the error threshold in terms of the number of false alarms.

2.1. Influence of the Target Power

This section analyzes the influence of the target power on segmentation’s efficacy.

2.1.1. Effect on the Distributions

The expectation of the Matched Filter without a target present is zero. But when a target is added, a bias occurs:
μ i = p t T Φ i 1 t ,
where p and t refer to the target power and direction, respectively, and Φ i refers to one of the global or local covariance matrices. Therefore, increasing the target power ( p ) increases the expectation proportionally, shifting the “with target” distributions to the right, providing more detections ( P D ). Since this is true for any selected threshold, the performance ( A ) improves as well (see Equation (4)). Additionally, since Equation (6) holds for any kind of covariance, both the global and the local performance improve.
However, it is unclear what happens to the benefit ( B ), defined by the local to global performance ratio (Equation (5)), whether it remains constant or it grows in some way. Answering this requires delving into the role of p .

2.1.2. Statistical Influence of the Target Power

Figure 2a may represent a spectral cross-section of two segments’ covariances, their composed covariance ( Φ G ), and a target vector ( t ) with some power (p). As such, the illustration shows that as p increases, the Signal-to-Noise Ratio (SNR) increases relative to any covariance, whether global or local. Equation (7) shows that this increase happens proportionally regardless of the domain:
S N R e x p e c t e d   s i g n a l s t a n d a r d   d e v i a t i o n = ( 2 ) μ σ = ( 3 ) μ ( 6 ) p .
This insight leads to the intuition that there are two extremes and one intermediate state:
  • When the target is weak, its SNR is bad even locally. There is, therefore, no point in segmenting since the resulting performance would remain bad in any case.
  • When the target is strong, its SNR is already good globally, and segmentation is redundant.
  • However, an intermediate p might exist, as in the example of Figure 2, where the SNR is still bad globally but already good locally. This case is where segmentation is worthwhile since it provides good local performance, which is, at the same time, significantly better than the global one.
Therefore, the response B p is expected to be unimodal, low at the extrema, and maximal in the middle. Still, some characteristics of this shape are unclear, such as the location of the optimum, the maximal performance improvement, and the efficacy range of p values where segmentation is beneficial.

2.1.3. Range of Effective Target Powers

A synthetic simulation, illustrated in Figure 4, enabled characterizing the shape of B p . Increasing p gradually from a tiny value showed that although the performance improves in both domains, a gap occurring between the global ( A G ) and local ( A L ) performance function causes the benefit function ( B ), given by their ratio (Equation (5)), to become bell-shaped. Below, the origin of this gap is analyzed through the performance functions.
It can be seen in Figure 4c that the local A L is the one that rises first, and therefore B initially rises (Figure 4d). This rising starts to become significant when the highest segment expectation in the local domain (Figure 4b) just crosses its decision threshold ( η L ). This cross is related to the fastest increase in P D , causing the local almost steepest ascent of A L . Segmentation starts to be worthwhile.
As the target keeps getting stronger, B rises a bit more but reaches its maximum rapidly and begins to fall (Figure 4d). This fall happens due to its denominator A G that is starting to grow, thanks to the global distributions in which detection is beginning to be feasible (Figure 4a).
Thereafter, the benefit B keeps decreasing until becoming so low that segmentation is no longer worthwhile. This decrease comes from the steep ascent of A G (Figure 4c), originated in the global distribution (Figure 4a) that crosses its decision threshold ( η G ).
This analysis reveals that the range of target powers in which segmentation might be effective is well defined by locating, both globally and locally, where the with-target expectation meets the decision threshold. We call these domain boundaries: p 1 after the leading local expectation and p G after the global expectation. Substituting Equation (6) gives explicit expressions for these indicators:
p 1 η L max s t T Φ s 1 t   ,   p G η G t T Φ G 1 t .
These values could be viewed as anchors regarding p , since its impact on segmentation efficacy does not depend on its absolute value but only on its position relative to these values. As p is a multiplier, the degree of segmentation efficacy is indicated by the ratio of these anchors, that is,
Δ p p G p 1 .
This index is called the “efficacy range’s width”.

2.1.4. The Optimal Target Power

It can be seen in Figure 4d above that the benefit function B ( p ) becomes optimal shortly after p 1 . Further experiments, summarized in Figure 5, show that this behavior is consistent for any type of data structure and that in the pure Gauss case, the optimal target power ( p m a x ) satisfies approximately p m a x 2 p 1 .
Figure 5 also shows that the maximal benefit ( B m a x ) increases together with the efficacy range ( Δ p ). Their connection can be explained mathematically. At the p m a x point, A L is always 0.5, so the major effect on B comes from A G (Equation (5)). Since μ G keeps decreasing, the numerator in Equation (4) tends to zero. Numerical approximations such as Taylor’s can show that this tendency decreases proportionally to μ G . Substituting Equation (8) and the p m a x approximation into Equation (6) gives μ G 2 p 1 η G p G 1 Δ p . Compiling all together gives:
B m a x 1 A G 1 μ G Δ p .
The logic in this result is that the more the data structure challenges the detection due to its lack of homogeneity, the broader the range of targets to which segmentation can contribute. As such, relatively weaker targets become detectable, and segmentation brings a higher benefit.

2.1.5. Principles

The target power ( p ) affects the segmentation benefit ( B ) through the SNR. It has a unimodal response that has a range of efficacy bounded by well-defined p 1 and p G anchors. The optimal target power goes with p 1 and the respective benefit is proportional to these anchors’ ratio ( Δ p ).

2.2. Influence of the Data Inhomogeneity

This section analyzes the influence of the covariance matrix ( Φ ), mainly affected by the degree of inhomogeneity of the data from which it is estimated.

2.2.1. A New Measure, “Kb”

The previous section concludes that the maximal segmentation benefit depends on the efficacy range ( Δ p ) , as defined in Equation (9). Substituting the anchors’ explicit expressions from Equation (8) gives a two-part expression, with the error threshold affecting only the first part, where the covariances and the target direction affect mainly the second part:
Δ p = η G η L max s t T Φ s 1 t t T Φ G 1 t K b max s t T Φ s 1 t t T Φ G 1 t .
The latter part is called K b after “benefit factor” since this scalar factor has a proportional effect on Δ p , hence also on the maximal benefit.
Therefore, when the data structure changes and the covariances layout is affected, the segmentation performance is influenced by how this change is reflected through that key factor K b . Therefore, it is essential to understand the significance of this new intermediate factor.

2.2.2. Meaning of “Kb”

Substituting the Matched Filter expectation (Equation (6)) into K b definition (Equation (11)) gives a ratio between the maximal local expectation and the global expectation. However, K b might be better comprehended in the spectral domain where Equation (6)’s expression signifies “Mahalanobis Distance” along the target direction ( t ), as illustrated in Figure 6a:
Such a distance means “the length of the full target vector p t , relative to its intersection point with the covariance matrix’s ellipsoid”, a point denoted here as X S . Hence there exists an inverse relationship between this intersection point ( X S ) and the expectation ( μ ) above. This connection derives that the K b expression can be rewritten as a ratio of intersection points:
K b = max s μ s μ G = X S G min s X S s .
This expression gives, in a layout like the one illustrated in Figure 6b: the intersection with the global Φ G relative to the local Φ 1 , or simply, X S G / X S 1 .
Two types of inhomogeneity are illustrated in Figure 7, with their corresponding K b annotated. Type (a) is “scaling”, where the segments’ covariances differ just by a scalar factor, and type (b) is “angular”, where the covariances differ by some planar rotation. In this illustration, despite the different inhomogeneity types, the K b ratio obtains a completely identical value. However, this identity sustains only in the specific illustrated angle of t , where in other angles, the K b ratio becomes different. This follows that K b reflects the degree of the data inhomogeneity, though not in the general sense, but rather in how it acts along the specific target direction.

2.2.3. Principles

A newly introduced factor called K b was found to be a key in understanding the impact of variations in the data. That is since when the data structure varies, the covariances change by an influence that, to a specific target, is condensed into that scalar K b , which has an immediate connection to Δ p and affects the optimal benefit B m a x proportionally.
This new factor is given by a closed-form expression, much easier to calculate than the benefit that involves complex optimization over integrations. This expression is equivalently a ratio of expectations or a ratio of intersection points, indicating the degree of “directional inhomogeneity” of the data. Moreover, this expression implies that inhomogeneity does not determine the segmentation impact in general but rather its highest possible impact, where the specific impact depends on the specific target of interest.

2.3. Influence of the Target Direction

This section analyzes the influence of t , the target direction in the spectral domain.

2.3.1. A Rule of Thumb: Theoretical Perspective

As discussed in Section 2.2.2, analyzing the target direction impact on the segmentation benefit is possible by tracking its effect on the intermediate factor K b . Figure 8 below illustrates a case of angular inhomogeneity and a rotating target over some cross-section plane. The illustration shows that when t rotates, the ratio of the intersection points does not change significantly. Hence the maximal benefit B m a x is not changing significantly as a function of the target direction.
However, this is not the only possibility. Additional examples are shown in Figure 9. Specifically, similar to case (a), where K b is inherently rotation invariant, and case (b), which we have just analyzed, case (c) is also nearly invariant, as it represents a linear transformation of case (b); hence, proportions are preserved, and the K b ratio remains intact. However, cases (d) and (e) are different. In case (d), the covariances also have a scaling difference on top of just rotation. This scaling causes the target to achieve a K b ratio much smaller at 90° than at 0°. A similar phenomenon occurs in case (e), where only the major eigenvectors are rescaled.
This phenomenon reveals that it is likely to find the optimal target direction close to one of the strongest eigenvectors where the most significant gap between the global and the local intersections usually occurs.

2.3.2. Estimating the Optimal Direction

The rule of thumb valid on two segments is not necessarily true on multiple segments. Fortunately, however, an analytical method for finding such a direction has been discovered. Formally, the optimal direction should be around the maximal K b t , that is, in the direction where the highest ratio between X S G and the minor X S s occurs (Equation (12)). But since that minor value might jump from one local segment to another, it is easier to first focus on solving one single segment, Φ 1 , as in Figure 10a, in which the remaining objective is to maximize X S G / X S 1 .
It turns out that, in this reduced case, whitening the system by Φ 1 provides an interesting result. As illustrated in Figure 10, the whitening transforms X S 1 to 1 . In consequence, by Equation (12), X S G becomes none other than K b (Figure 10b). Since this property holds for any direction t , the whole ellipsoid of the whitened Φ G represents K b in all possible directions. The biggest K b is therefore obtained where the largest radius occurs, which is simply on the major eigenvector (Figure 10c). All that remains, then, is to transform this vector back to the original domain, thus obtaining the optimal target direction in terms of K b (Figure 10d). In the general case, this process is repeated for every segment, and the highest K b is selected, as defined mathematically in the following equations:
t ^ max argmax t K b t = Φ s ^ 0.5   K b m a x s ^ ,   s ^ = argmax s K m a x s ,   K b m a x s max t K b s = ± major Φ s 0.5 Φ G Φ s 0.5   .
It is worth noting that the proposed process relies solely on covariance matrices and involves highly efficient operations, making it computationally fast. Any optimal direction t ^ max is only defined up to a sign ( ± t ).
On real data, maximizing K b provides only an estimation of the optimal direction. Where this is generally reasonable in predicting the maximal efficacy width (Δp), due to their direct relationship (Equation (11)), it is less accurate in predicting the maximal benefit. Not only that, the actual benefit at ± t ^ max may vary significantly due to asymmetries in the data structure, but the real maximal benefit ( B m a x ) usually does not occur exactly there, as this estimation is based on the second moment and does not fully capture the real data distribution. Similarly, the range of effective target directions can be approximated by solving arg t K b t > β for a desirable threshold β .

2.3.3. Principles

The target direction t interacts with the covariances, affecting segmentation efficacy through the key factor K b . The optimal direction can be estimated analytically by maximizing K b , which gives a closed-form solution, easy-to-calculate. Likewise, the directions’ efficacy range can be estimated by comparing K b to some desirable threshold.

2.4. Influence of the Decision Threshold

This section analyzes the influence of t h , the error threshold, in terms of the number of false alarms.

2.4.1. The Joint Impact of the Error Threshold

According to Equation (1), permitting fewer false alarms by a lower error threshold ( t h ) derives a higher decision threshold ( η ). Since this property holds for both the global and the segmented MF, both p 1 and p G anchors increase proportionally (Equation (8)). Hence, while the optimal power ( p m a x ), which goes with the first anchor, should increase (Section 2.1.4), the efficacy width ( Δ p ), given by the ratio of these anchors (Equation (9)), is not expected to change significantly. However, the target direction ( t ) operates orthogonally, since its impact on the performance passes mainly through the K b factor, which is independent of the thresholds (Equation (11)).
A synthetic simulation, shown in Figure 11, demonstrates the effect of reducing t h threshold. Indeed, the optimal target power ( p m a x ) shifts right along with both anchors, while the relative width ( Δ p ) remains constant. Surprisingly, yet, along with them, the optimal benefit ( B m a x ) increases, instead of remaining constant after its proportional relation to Δ p (Equation (10)).

2.4.2. Impact on the Maximal Benefit

Understanding the root of B m a x ’s rising was possible using a synthetic simulation, illustrated in Figure 12. In that simulation, t h was reduced gradually from a high value, causing η to shift right in both the global and the local distributions. Simultaneously, for tracking the impact on B m a x , the target power was adjusted as to sit on its optimal point ( p m a x ). Since such an adjustment causes the highest local expectation to sit consistently just after the local threshold (Figure 12b), P D barely changes, and the local performance ( A L ) remains almost constant (Figure 12c). Contrarily, globally, while the threshold shifts similarly, the with-target distribution lags behind (Figure 12a), P D decreases consistently, the global performance ( A G ) drops monotonically (Figure 12c), and the benefit function grows endlessly (Figure 12d). Thus, unlike B ( p ) , which responds unimodally (Figure 4), t h has a monotonic response.
Intuitively, this behavior indicates that as the global detection becomes more challenging due to a stricter threshold, the impact of segmentation becomes more pronounced. Nevertheless, it remains intriguing why the behavior differs so significantly between the global and local domains. After all, shifting a threshold is equivalent to inversely shifting expectations, so one would expect the change in the number of detections to change similarly in both domains.
Delving into the above theory reveals that due to Equation (1), reducing the error threshold shifts both the global and the local decision thresholds in similar proportions. Whereas, as a target power increases, each global or local expectation grows relative to the intersection point in its domain (Figure 6), whose ratio is nothing but K b (Equation (12)). The simulation above (Figure 12) used K b = 8 , which caused the global expectation to barely move, although the respective threshold kept shifting. Thus, unlike expectations, changing a target power is inequivalent to an opposite shift of the respective decision threshold. This insight means that the global domain, where the target is weaker relative to the reference noise, requires more relative efforts to improve the SNR.

2.4.3. Impact per Domain

The definite trend of the optimal point can be extracted from mathematics. The optimal power ( p m a x ) goes with the first anchor (Section 2.1.4), which substituting Equation (9)’s components with Equation (11) and Equation (8) gives a proportion of η L / K b for any given global SNR. The corresponding benefit ( B m a x ) evolves with A G , due to the constant A L (Figure 12c), hence it depends on an integration over P D in the global domain (Equation (4)). Since P D is based on the with-target distribution ( f T ), which is a shifted version of the without-target distribution ( f N ), its integration gives a similar response as P F A which gives 0.5 t h 2 [10]. This implies that A G in Equation (4) is proportional to t h 2 / t h 0.5 t h 2 which tends to t h as t h goes to zero. Thus, B m a x tends to the inverse of t h . Additionally, Equation (11) gives Δ p K b which, substituting into (10) gives B m a x K b . Combining all together gives, asymptotically, the following proportions:
p m a x   , p 1 η L K b ,   B m a x K b t h .
Interestingly, the threshold effect is related to the domain of each factor: the target power factors vary with η that is in the scores’ domain, where the optimal benefit varies with t h which acts in the integral domain.

2.4.4. Principles

The more challenging the detection becomes due to a tighter threshold, the more the p 1 and p G anchor points increase, as well as the position and intensity of the optimal benefit. The response to such a change is monotonic and tends to evolve proportionally, unlike the target power that responds unimodally. The threshold has a negligible effect on the target’s efficacy range and has an orthogonal effect to K b and thus to any of the target direction factors.

2.5. Guidelines

This section analyzes how each of the NSMF factors affects the segmentation success:
  • The target power ( p ) affects the efficacy of segmentation through the SNR. The segmentation’s benefit varies unimodally with respect to two key anchors called p 1 and p G . The optimal p is proportional to the small anchor ( p 1 ), while the width of the efficacy range is determined by the ratio of p G to p 1 (Equation (9)).
  • The covariances ( Φ s ), reflecting variations in the data structure, affect segmentation’s efficacy through a scalar key factor named K b . This factor qualifies “directional inhomogeneity” along the target direction and has a proportional influence on both the targets’ efficacy range ( Δ t ) and the maximal benefit ( B m a x ).
  • The target direction ( t ) interacts with the covariances, affecting efficacy through K b factor. Maximizing K b enables estimating the optimal direction using a closed-form expression that is computationally straightforward.
  • Threshold ( t h ): Tightening the threshold raises both p 1 and p G anchors, resulting in a monotonic increase in the optimal segmentation benefit. Such a change triggers proportional relationships: the target power factors vary with the local decision threshold ( η L ), while the optimal benefit varies with the error threshold ( t h ).
These properties provide a framework for characterizing practical cases, allowing for informed decision-making regarding the range of targets and thresholds where segmentation is worthwhile in the detection process.

3. Experiments

This section presents the experimental results conducted to validate the theoretical conclusions. It begins by introducing the preparations made for these experiments. Subsequently, it delves into the influences of the data structure on the results. Finally, it explores the influence of the target signature on real datasets.

3.1. Experiments Considerations

For our experiments, we utilized the three types of data described in detail in Section 1.5: the Synthetic Dataset, the Simple In-House Cube (SIC), and real-world data, represented by the RIT Cube. Each dataset offered unique characteristics and challenges for experimentation. To ensure the reproducibility of our results, this section outlines the special preprocessing steps applied to prepare these datasets for experimentation.
Each of these three datasets posed unique preprocessing challenges: The Synthetic Dataset required adjustments to maintain analytical consistency. The SIC dataset involved targeted manipulations to introduce controlled inhomogeneities. Finally, the RIT Cube required addressing significant boundary estimation biases caused by the non-convexity of real-world segments.
The principal methods used to address these challenges are outlined below. A more detailed description is available in the comprehensive study report referenced in the Supplementary Materials.

3.1.1. Preprocessing

Several preprocessing steps were necessary for SIC data simulations. One critical step involved ensuring that data manipulations were based on a neutral starting point. First, we addressed the imbalance in segment sizes by cropping the image from 50 × 50 to 34 × 47, effectively reducing the populations imbalance from 40 % to 0.2 % . Next, we addressed stationarity by balancing the segment covariances using a unique whitening transform, defined as Φ G 0.5 Φ s 0.5 . This process whitens segment s by its own covariance and then unwhitens it by the global covariance. As a result, this manipulation effectively reduced the covariances’ deviation from 5 ° to 0 ° .
The SIC data simulations were carried out over cross-sections of interest, for example, by the plane spanned by the major and minor Φ G ’s eigenvectors. Such a simulation includes linear transformations for each segment across planar sections by transforming the data by V T V T , where V is a basis, such as the eigenvectors matrix, and T is a desirable transform matrix. The latter is the identity matrix, excluding four cells determining a 2D transform over the two selected dimensions.

3.1.2. Practical Aspects

Each type of data had its challenges after its specific characteristics and its respective processing.
In some extreme cases, the synthetic simulations encountered numerical errors. To address this, a more accurate method was developed, incorporating three key principles: 1. maintaining pure analytic forms consistently, 2. employing standard numeric solutions for Gaussian distributions, and 3. applying approximation such as L’Hôpital’s theorem.
In the SIC realistic simulations, the presence of long-tailed distributions impacting performance, coupled with data quantization leading to high volatility, posed challenges in deriving reliable conclusions. To address this issue, we compared each result with a corresponding synthetic simulation, enabling us to characterize distribution shape properties and meticulously isolate stochastic phenomena.
In RIT’s standard dataset, background estimation errors occurring along the boundaries of segments introduced a notable bias in the per-segment average noise ( x m ). This challenge was successfully resolved by developing a unique mirror padding technique, specifically designed for non-convex borders, as elaborated in Appendix A.

3.1.3. Consistency

Special attention was paid to isolating the primal sources of influence. While target power exclusively affects the SNR (Equation (7)), data structure and target direction simultaneously affect both homogeneity (Equation (11)) and SNR (Equation (6)). To understand their roles thoroughly, we fixed the SNR while varying the data structure and the target direction. This was obtained by consistently normalizing the vector t so that the global expectation ( μ G ) remains constant.

3.2. Influence of the Data Structure

This section presents results regarding the influence of inhomogeneity due to variations in the data structure. It presents representative results for each of the three types of datacubes.

3.2.1. Synthetic Simulation

Using the synthetic dataset, we gradually rescaled two identical segments. In consequence, the inhomogeneity increased and K b grew, respectively. Figure 13a shows the effect of K b on the optimal target power ( p m a x ), in the upper graph, and on the respective benefit ( B m a x ), in the lower graph, both on the logarithmic scale. The results show that for a high enough K b , p m a x decreases proportionally and B m a x increases proportionally. Additionally, as the error threshold ( t h ) decreases, both p m a x and B m a x grow consistently.
Although these results match the theoretical conclusions of Equation (14) asymptotically, it seems that at K b < 10 the benefit grows faster. For example, at a threshold of 0.01, there is a quadratic growth rate of K b 2 . An alternative model that approximately satisfies the plots might be B m a x     H 10 K b K b / 10 t h + H ¯ 10 K b t h log K b , where H 10 ( ) and H ¯ 10 ( ) refer to a composition around 10 , such as the Heaviside unit function u ( K b 10 ) and its complementary, respectively. However, such a model would only be applicable for data that follows normal distributions, which is not typically the case [11].

3.2.2. SIC Real Simulation

Using the SIC dataset, we repeated the simulation for real data. Starting with a purely stationary state, we gradually rescaled the two segments by some multiplicative factor so that K b increases, regardless of the target direction. Figure 13b shows the effect on the optimal target power p m a x and on the respective benefit.
The results closely resemble those in the synthetic case above, with the primary distinction being that B m a x converges to a constant value. The origins of this difference can be traced back to quantization and the presence of long-tailed distributions in real data. In extreme K b , the optimum ( p m a x ) is obtained at a super weak target that shifts the global expectation to almost zero. Then, due to quantization, the number of detections becomes like the number of false alarms, the global performance stops falling, and the benefit ( B m a x ) stabilizes at a fixed value.

3.2.3. RIT Standard Dataset

In the RIT standard dataset, the underlying structure is inherent. However, since K b is related to the interaction with the target direction (Equation (11)), the data can be examined by selecting significant directions. We, therefore, chose the directions of the strongest eigenvectors from each of the five covariances of this data. For each direction, we computed K b , determined the optimal p along with its respective benefit ( B m a x ), and plotted it on scattergrams as shown in Figure 13c.
The results exhibit behavior consistent with the previous findings: as K b rises, p max drops, and B m a x grows proportionally. The progression of B m a x follows an approximately quadratic rate in this specific case, as K b is less than 10, and as we chose a threshold ( t h ) of 0.01.

3.3. Influence of the Target Signature

This section presents results regarding the influence of the target signature on real datasets. It starts with a representative two-segment case. It then continues with a standard dataset and examines diverse types of target spectrum, striving to deduce ways to improve the performance further.

3.3.1. A Rule of Thumb: Practical Perspective

To test the target direction impact in the reduced two-segment case discussed above (Section 2.3.1), we used the SIC real dataset and a cross-section plane spanned by the major-to-minor Φ G ’s eigenvectors (Section 3.1.1). Over that plane, we simulated angular inhomogeneity and rotated a target from 0 ° to 90 ° (Figure 14a). We then measured the optimal benefit obtained at each angle. The resulting upper graph in Figure 14c shows almost no rotation impact.
Thereafter, we gradually reduced the major eigenvector of Φ 2 and rechecked the rotation impact (Figure 14b). The two lower graphs in Figure 14c exhibit a decline as they approach 90°, in accordance with the decreasing factor. This decline happens due to the reduction in the ratio of K b intersections as explained in Section 2.3.1. This reveals that since such asymmetries are common, it is likely to find the optimal target direction close to one of the most dominant eigenvectors.

3.3.2. Estimating the Global Optimum

To examine the impact of different target directions in the standard RIT dataset, we started with searching for the global optimal target ( t m a x ). We first estimated this optimum by maximizing K b using Equation (13). To get a sense of how good this estimate is, we crossed the spectral space with a plane spanned by the estimated t ^ m a x vector and Φ 5 ’s major eigenvector (Figure 15a). Over this plane, we rotated a target and measured K b and B m a x per angle.
The results appear in Figure 15b. The lower graph shows K b , which is indeed maximal at 0°, where t ^ m a x sits. The upper graph shows the actual B m a x , whose true maximum is not much higher than its value at t ^ m a x , but still, its position is somewhat shifted to the left. Fortunately, however, the curve between these two points looks convex, implying that the global maximum might be achieved using standard optimization methods initialized at the estimated point.
Table 1 column (a) shows the performance values for the estimated best direction. For a 0.01 error threshold, segmentation improves performance from 0.006 to 0.252, which is 40.6 times better. By comparison, the best target found in Ben Yakar’s work improved the performance by only 4.1 times (see Table 1 in [12]). Table 1 also shows that a smaller error threshold improves even more dramatically. So, our method seems to enable finding a direction with a benefit from the highest possible, which is also close enough to the absolute maximum.

3.3.3. Limiting the Optimum to the Positive Cone

Unfortunately, the just found global t ^ m a x cannot be a target spectrum since some of its components are negative. In Figure 16a, the area outside the positive cone is grayed out, showing that t ^ m a x remains outside, while Φ 4 ’s eigenvector remains inside the cone. Figure 16b shows that the optimal direction along the rotation towards that eigenvector sits precisely on the cone border, making it relatively easy to find.
Table 1 column (b) shows that even under the positivity constraint, segmentation improves performance by 36.4 times, which is still remarkable. So, rotating the global best direction towards a target that sits within the valid area enables finding a direction that meets the required constraint but is still close enough to the absolute maximum.

3.3.4. Limiting the Optimum to the Image Pixels

Since forcing positivity is not enough to promise a valid signature, we additionally narrowed the constraint to the set of the image pixels so that the target would be undoubtedly valid. We searched for the pixel whose direction provides the maximal benefit among all the pixels. Figure 17a shows a map of all the pixels that give K b   >   2 . We analyzed each of these pixels using our rules and found the pixel that offers the maximal benefit. This pixel is annotated “ × ” in three different views in Figure 17. Notably, this pixel belongs to segment #5, which boasts the highest benefit among the five segments, where the pixel itself is positioned on the border of that segment and visually differs from most other pixels within this segment.
Table 1 column (c) shows that despite the extreme constraint, segmentation enhances performance by 13.2 times, which is still three times better than Ben-Yakar’s best result [12]. Hence, one might get a sense of the maximal benefit of valid targets through image self-analysis of the kind presented here.

3.3.5. Efficacy with the Provided Targets

We further constrained the selected target by narrowing the possible set to only the twelve targets provided as part of the dataset. The best performer was a laboratory sampled blue cotton (F3). The result, shown in Table 1 column (d), gives that for a 0.01 error threshold, segmentation improves by only three times. This result indicates that segmentation might be redundant for such targets.
Overall, the results in Table 1 show that although segmentation provides a high impact in some directions—while narrowing the set of valid targets, this impact drops until becoming negligible. This rapid degradation implies that the efficacy range of valid directions might be relatively tight.

3.3.6. Directions’ Efficacy Range

One option for getting an impression of the efficacy range is sampling targets from the dataset itself. Figure 17a depicts such an example, showing a map of K b > 2 pixels. It was found to be 97.5% accurate (measured by the relative number of true hits) equivalent to Δ p > 3.98 and 95.6% accurate equivalent to B m a x > 3.98 , except that its computation lasts only 1.5 s instead of more than 100 h. Curiously, in this resulted map, the lit pixels occupy only 7.4% of the image. This result means that not many valid directions exist where segmentation is beneficial for this specific dataset.
One of the reasons for the limited domain of targets is the high sensitivity to the exact direction. It can be shown by comparing Φ 5 ’s major eigenvector to the pixel with the most correlated signature, which, despite the considerable similarity (Figure 18a), tracking their cross-section reveals that B m a x drops from 17 to 4.7 in just 2.3° (Figure 18b). This phenomenon might be explained by the above insight that K b , from which B m a x evolves, is represented by the radius of a multidimensional ellipsoid in the whitened domain ( Φ ~ G in Figure 10c). It turns out that, like in the data [1], the first eigenvectors of this ellipsoid are relatively strong and then weaken rapidly. Therefore, when the target rotates between strong eigenvectors, K b barely changes. Whereas, when it rotates towards most of the other weak dimensions, then a rapid degradation occurs. This behavior implies that the efficacy range of target directions is built from a manifold wide in few dimensions but narrow in most others.

3.4. Efficacy with RIT Data

According to the results, discerning the worthiness of segmentation for the RIT dataset poses an initial challenge. On one hand, the observed benefits are pronounced in a limited range of targets; however, on the other hand, this range broadens as the required threshold diminishes, amplifying the overall segmentation’s benefit. Yet, a pivotal consideration in this example may stem from the utilization of a basic “K-Means” segmentation approach, which might not inherently prioritize the creation of more homogeneous areas, contrary to our initial objective. Therefore, a transition to a more suitable technique designed to enhance stationarity might potentially result in a substantial enhancement to the detection across a broader range of targets, even under stringent threshold conditions.

4. Conclusions

This study aimed to identify the conditions under which segmentation improves the detection of subpixel targets in hyperspectral images. While data inhomogeneity was initially hypothesized as the primary factor, our comprehensive analysis revealed that multiple algorithmic factors also significantly impact detection performance.
Our analysis of the NSMF algorithm identified the matched target as a critical factor in segmentation performance, specifically through its influence on “directional inhomogeneity”. We introduced a novel factor, “ K b ”, representing the maximal benefit, and developed a closed-form estimator for determining the optimal target direction. These findings were distilled into practical rules describing the influence of each factor on segmentation performance.
We validated these insights through extensive simulations across a broad range of conditions. Additionally, we applied these findings to two standard datasets, evaluating segmentation performance for various targets and characterizing the conditions under which segmentation is beneficial. This enabled deducing properties about the setup and proposing ways to enhance the detection. Further investigations into additional datasets and scenarios are ongoing.
In contrast to prior studies that primarily demonstrated segmentation’s potential for improving detection performance [5,6,7], this work focused on dissecting the underlying factors influencing segmentation efficacy. By analyzing these factors systematically, this study bridges the gap between observed benefits and the conditions that enable them.
Additionally, the theoretical insights, this work introduces practical analysis tools that can support future research and real-world applications, including:
  • p -rules: Essential for characterizing segmentation’s effects on detection performance.
  • K b : A key factor for efficiently predicting segmentation benefit for any target.
  • t ^ m a x : A closed-form estimator for the optimal target direction, easily computable.
These tools can be adapted for use with different models and algorithms. Unlike our prior work [8], which focused only on the influence of the matched target, this study provides an in-depth analysis of all contributing factors, supported by theoretical proofs and comprehensive experimental validations.
This study lays the groundwork for completing prior research on predicting segmentation benefits a priori. While an earlier study proposed linking data inhomogeneity to segmentation efficacy, we lacked a practical framework due to the complexity of segmentation-dependent factors. The K b m a x metric introduced here simplifies this challenge by providing a single, interpretable measure of inhomogeneity for any given segmentation. This enables systematic comparisons with segmentation-independent inhomogeneity metrics, allowing evaluation of their ability to predict segmentation’s potential benefit.
This study focused on the Normalized Segmented Matched Filter (NSMF) algorithm to analyze segmentation efficacy. Ongoing research is testing whether segmentation is similarly effective in other algorithms, such as the Adaptive Cosine Estimator (ACE), which also relies on covariance-based whitening. Additionally, future work will explore alternative segmentation strategies, including hierarchical clustering, to assess their impact on the conclusions drawn here. These efforts aim to validate and expand the applicability of the proposed framework across different detection algorithms and segmentation methods.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25010272/s1, Related code and data are available at: https://github.com/yoramfurth/spie2022_figures, accessed on 1 January 2025. Supplementary documentation of the research can be accessed at: https://drive.google.com/drive/folders/1AnwbHInJEoEpGemREbNQFz3s9-USge6c, accessed on 1 January 2025.

Author Contributions

Conceptualization, Y.F. and S.R.R.; Methodology, Y.F.; Software, Y.F.; Validation, Y.F.; Formal analysis, Y.F.; Investigation, Y.F.; Resources, S.R.R.; Data curation, Y.F.; Writing—original draft, Y.F.; Writing—review & editing, Y.F. and S.R.R.; Visualization, Y.F.; Supervision, S.R.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The synthetic data used in this study are generated programmatically using deterministic parameters, and the code for their generation is provided in the repository associated with this article. The RIT dataset is publicly available at http://dirsapps.cis.rit.edu/blindtest/, while the Via Reggio dataset is available upon request from its authors. The SIC dataset, a cropped hyperspectral image, can be provided upon request from the authors. All results are reproducible using the provided code, with no intermediate data stored.

Acknowledgments

We would like to express our gratitude to the Rochester Institute of Technology (RIT) for providing the publicly accessible hyperspectral dataset used in this study. We also thank the authors of the Via Reggio dataset for granting us access to their valuable data.

Conflicts of Interest

The authors declare that they have no competing interests or conflicts of interest related to the content of this manuscript.

Appendix A. Unbiased Background Estimation

Estimating a covariance matrix, as in Equation (2), is typically performed by averaging x   -   m x     m T over a set of pixels [23]. Since m itself is estimated with a spatial filter [2,10], a bias might occur, resulting in a nonzero expectation of x m . Appending a mirror padding to the image borders overcomes such a global bias algebraically. However, this does not hold while applying it per segment.
Swapping the view from local filtering to spreading a pixel’s energy to its neighbors (Figure A1a) underlies the foundation of this solution and unveils a more fitting generalization. For a square frame, the regular mirror padding compensates for any missing contribution with a respective mirrored value (the arrows in Figure A1b). In arbitrary segments, however, non-convex borders suffer from an ambiguous definition of the mirroring operation and an unbalanced redistribution of the energy within the image. In the example of an 8 -neighbors averaging filter [10], illustrated in Figure A1c, a naïve mirror padding cause x j to contribute back 11 times instead of 8 times.
Figure A1. The concept of unbiased filtering: (a) the energy of a pixel x j spreaded by a 5   ×   1 ones filter, (b) the effect of mirror padding, (c) the origin of unbalanced contribution, and (d) a corrected contribution. The white area represents the set of active pixels to be padded, while the gray area represents the outer region where the padding occurs. The arrows represent the spread energy.
Figure A1. The concept of unbiased filtering: (a) the energy of a pixel x j spreaded by a 5   ×   1 ones filter, (b) the effect of mirror padding, (c) the origin of unbalanced contribution, and (d) a corrected contribution. The white area represents the set of active pixels to be padded, while the gray area represents the outer region where the padding occurs. The arrows represent the spread energy.
Sensors 25 00272 g0a1
An alternative approach might be to mirror using an operation inverse to the following filtering so that the filtering will distribute the whole energy spread to the padded area back inwards. In the example of Figure A1d, each padded pixel contributes cumulatively exactly x j , so that x j is contributed precisely 8 times.
This idea can be generalized to mirroring the data using an adaptive flipped filter instead of simple direct padding. Each pixel on the margins would then get the weighted average of the pixels it sees within the image. If m is estimated by correlating x with a kernel h , then the mirror padding would be formed using the same kernel, flipped, and normalized by summing its in-segment elements. Formally, this means convolving x h / M s h per segment, where “ / ” denotes a pixel-wise divisor, “ ” denotes convolution, and M s is a binary mask of segment number s .

References

  1. Funk, C.; Theiler, J.; Roberts, D.A.; Borel, C.C. Clustering to Improve Matched Filter Detection of Weak Gas Plumes in Hyperspectral Thermal Imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 1410–1420. [Google Scholar] [CrossRef]
  2. West, J.E.; Messinger, D.W.; Schott, J.R. Comparative Evaluation of Background Characterization Techniques for Hyperspectral Unstructured Matched Filter Target Detection. J. Appl. Remote Sens. 2007, 1, 013520. [Google Scholar] [CrossRef]
  3. Bajorski, P. Generalized detection fusion for hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2011, 50, 1199–1205. [Google Scholar] [CrossRef]
  4. Pieper, M.; Manolakis, D.; Truslow, E.; Cooley, T.; Lipson, S. Performance Evaluation of Cluster-Based Hyperspectral Target Detection Algorithms. In Proceedings of the 19th IEEE International Conference on Image Processing, Orlando, FL, USA, 30 September–3 October 2012; pp. 2669–2672. [Google Scholar] [CrossRef]
  5. Liang, Y.; Markopoulos, P.P.; Saber, E. Spatial–spectral segmentation of hyperspectral images for subpixel target detection. SPIE J. Appl. Remote Sens. 2019, 13, 036502. [Google Scholar] [CrossRef]
  6. Liang, Y. Object Detection in High Resolution Aerial Images and Hyperspectral Remote Sensing Images. Ph.D. Thesis, Rochester Institute of Technology, New York, NY, USA, 2019. [Google Scholar]
  7. Stalley, S.O. Clustered Hyperspectral Target Detection. Master’s Thesis, Portland State University, Portland, OR, USA, 2020. [Google Scholar]
  8. Furth, Y.; Rotman, S.R. Effective Segmentation for Point Target Detection. In Proceedings of the SPIE Algorithms, Technologies, and Applications for Multispectral and Hyperspectral Imaging XXIX, Orlando, FL, USA, 13 June 2023; pp. 227–236. [Google Scholar] [CrossRef]
  9. Kelly, E.J. Performance of an Adaptive Detection Algorithm; Rejection of Unwanted Signals. IEEE Trans. Aerosp. Electron. Syst. 1989, 25, 122–133. [Google Scholar] [CrossRef]
  10. Caefer, C.E.; Stefanou, M.S.; Nielsen, E.D.; Rizzuto, A.P.; Raviv, O.; Rotman, S.R. Analysis of False Alarm Distributions in the Development and Evaluation of Hyperspectral Point Target Detection Algorithms. Opt. Eng. 2007, 46, 076402. [Google Scholar] [CrossRef]
  11. Manolakis, D.; Marden, D.; Shaw, G.A. Hyperspectral image processing for automatic target detection applications. Linc. Lab. J. 2003, 14, 79–116. [Google Scholar]
  12. Ben-Yakar, S.; Blumberg, D.G.; Rotman, S.R. Advantages and limitations of segmentation for point target detection in hyperspectral imagery. In Proceedings of the 6th IEEE Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lausanne, Switzerland, 24–27 June 2014; pp. 1–4. [Google Scholar] [CrossRef]
  13. Dong, S.; Quan, Y.; Feng, W.; Dauphin, G.; Gao, L.; Xing, M. A pixel cluster CNN and spectral-spatial fusion algorithm for hyperspectral image classification with small-size training samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4101–4114. [Google Scholar] [CrossRef]
  14. García, J.L.; Paoletti, M.E.; Jiménez, L.I.; Haut, J.M.; Plaza, A. Efficient semantic segmentation of hyperspectral images using adaptable rectangular convolution. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  15. Xie, F.; Wang, R.; Jin, C.; Wang, G. Hyperspectral image classification based on superpixel merging and broad learning system. Photogramm. Rec. 2024, 39, 435–456. [Google Scholar] [CrossRef]
  16. Swarupa, V.V.S.; Devanathan, M. Hyperspectral Image Acquisition Methods and Processing Techniques Based on Traditional and Deep Learning Methodologies—A Study. In Proceedings of the IEEE Third International Conference on Smart Technologies in Computing, Electrical and Electronics (ICSTCEE), Bengaluru, India, 16–17 December 2022; pp. 1–13. [Google Scholar] [CrossRef]
  17. Fuding, X.I.E.; Xu, L.I.; Dan, H.U.A.G.N.; Cui, J.I.N. Superpixel Merging-Based Hyperspectral Image Classification. J. Syst. Sci. Math. Sci. 2021, 41, 3268. [Google Scholar]
  18. Snyder, D.; Kerekes, J.; Fairweather, I.; Crabtree, R.; Shive, J.; Hager, S. Development of a web-based application to evaluate target finding algorithms. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGRASS), Boston, MA, USA, 7–11 July 2008; Volume 2, pp. II-915–II-918. [Google Scholar] [CrossRef]
  19. Acito, N.; Matteoli, S.; Rossi, A.; Diani, M.; Corsini, G. Hyperspectral airborne “Viareggio 2013 Trial” data collection for detection algorithm assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2365–2376. [Google Scholar] [CrossRef]
  20. Theiler, J.; Matteoli, S.; Ziemann, A. Bayesian detection of solid subpixel targets. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Brussels, Belgium, 11–16 July 2021; pp. 3213–3216. [Google Scholar] [CrossRef]
  21. Hartigan, J.A.; Wong, M.A. A K-Means Clustering Algorithm. J. R. Stat. Soc. Ser. C-Appl. Stat. 1979, 28, 100–108. [Google Scholar] [CrossRef]
  22. Carlotto, M.J. A cluster-based approach for detecting man-made objects and changes in imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 374–387. [Google Scholar] [CrossRef]
  23. Kelly, E.J. An Adaptive Detection Algorithm. IEEE Trans. Aerosp. Electron. Syst. 1986, 22, 115–127. [Google Scholar] [CrossRef]
Figure 1. NSMF results on real data for two targets: the top row shows a real case of improved detection with segmentation, while the bottom row shows no improvement despite using the same data and segmentation. Column (a) displays global distributions, column (b) shows local distributions, and column (c) presents the ROC curves. Subscripts G / L denote global/local distributions, and superscripts N / T represent distributions without/with a target, respectively.
Figure 1. NSMF results on real data for two targets: the top row shows a real case of improved detection with segmentation, while the bottom row shows no improvement despite using the same data and segmentation. Column (a) displays global distributions, column (b) shows local distributions, and column (c) presents the ROC curves. Subscripts G / L denote global/local distributions, and superscripts N / T represent distributions without/with a target, respectively.
Sensors 25 00272 g001
Figure 2. The three datasets used for this study: (a) synthetic data of Gaussian distributions, (b) one band from a Simple In-house Cube (SIC), and (c) one band from a data-cube, provided by Rochester Institute of Technology (RIT).
Figure 2. The three datasets used for this study: (a) synthetic data of Gaussian distributions, (b) one band from a Simple In-house Cube (SIC), and (c) one band from a data-cube, provided by Rochester Institute of Technology (RIT).
Sensors 25 00272 g002
Figure 3. Segmentation maps used for (a) SIC datacube and (b) RIT datacube. The labels’ indexes correspond to the local covariances used in the sequel, that is, Φ 1 corresponds to segment #1, Φ 2 to segment #2, etc.
Figure 3. Segmentation maps used for (a) SIC datacube and (b) RIT datacube. The labels’ indexes correspond to the local covariances used in the sequel, that is, Φ 1 corresponds to segment #1, Φ 2 to segment #2, etc.
Sensors 25 00272 g003
Figure 4. Range of effective target powers through different domains: (a) the global distributions, (b) the local distributions, (c) the respective performance, and (d) the resulted segmentation benefit. The purple grayed-out region represents the powers that are out of the efficacy range.
Figure 4. Range of effective target powers through different domains: (a) the global distributions, (b) the local distributions, (c) the respective performance, and (d) the resulted segmentation benefit. The purple grayed-out region represents the powers that are out of the efficacy range.
Sensors 25 00272 g004
Figure 5. Evolution of B ( p ) as a function of inhomogeneity. The global SNR is fixed, hence, p G remains constant.
Figure 5. Evolution of B ( p ) as a function of inhomogeneity. The global SNR is fixed, hence, p G remains constant.
Sensors 25 00272 g005
Figure 6. The spectral meaning of K b : (a) expectation as a Mahalanobis Distance and (b) ratio of intersection points.
Figure 6. The spectral meaning of K b : (a) expectation as a Mahalanobis Distance and (b) ratio of intersection points.
Sensors 25 00272 g006
Figure 7. Comparing K b on two kinds of inhomogeneities: (a) scaling and (b) angular.
Figure 7. Comparing K b on two kinds of inhomogeneities: (a) scaling and (b) angular.
Sensors 25 00272 g007
Figure 8. Angular inhomogeneity through a cross-section plane and a target at three angles: (a) 0°, (b) 23°, and (c) 45°.
Figure 8. Angular inhomogeneity through a cross-section plane and a target at three angles: (a) 0°, (b) 23°, and (c) 45°.
Sensors 25 00272 g008
Figure 9. Target direction impact on different layouts: (a) scaling, (b) Φ 1 rotation; upon it, (c) appends system rotation and scaling, (d) appends rescaling of both ellipsoids, and (e) appends rescaling of the major axis of both ellipsoids.
Figure 9. Target direction impact on different layouts: (a) scaling, (b) Φ 1 rotation; upon it, (c) appends system rotation and scaling, (d) appends rescaling of both ellipsoids, and (e) appends rescaling of the major axis of both ellipsoids.
Sensors 25 00272 g009
Figure 10. Analytic process of maximizing K b t : (a) the objective with respect to one segment, (b) the effect of whitening, (c) the optimal K b in the whitened domain, and (d) the equivalent in the original domain.
Figure 10. Analytic process of maximizing K b t : (a) the objective with respect to one segment, (b) the effect of whitening, (c) the optimal K b in the whitened domain, and (d) the equivalent in the original domain.
Sensors 25 00272 g010
Figure 11. Evolution of B ( p ) as a function of different thresholds.
Figure 11. Evolution of B ( p ) as a function of different thresholds.
Sensors 25 00272 g011
Figure 12. The influence of the error threshold in different domains: (a) the global distributions, (b) the local distributions, (c) the respective performance, and (d) the resulted segmentation benefit.
Figure 12. The influence of the error threshold in different domains: (a) the global distributions, (b) the local distributions, (c) the respective performance, and (d) the resulted segmentation benefit.
Sensors 25 00272 g012
Figure 13. Influence of inhomogeneity for different thresholds on different types of datasets: (a) synthetic, (b) SIC cube, (c) RIT cube.
Figure 13. Influence of inhomogeneity for different thresholds on different types of datasets: (a) synthetic, (b) SIC cube, (c) RIT cube.
Sensors 25 00272 g013
Figure 14. Influence of the target direction in the two-segment case on a representative layout: (a) the layout, (b) after reducing one major eigenvector by half, and (c) the resulting maximal benefit in three cases.
Figure 14. Influence of the target direction in the two-segment case on a representative layout: (a) the layout, (b) after reducing one major eigenvector by half, and (c) the resulting maximal benefit in three cases.
Sensors 25 00272 g014
Figure 15. Segmentation performance for linear combinations of t ^ m a x with Φ 5 ’s major eigenvector ( ν 1 ): (a) a planar cross-section for the selected pair and (b) segmentation performance for a range of angles.
Figure 15. Segmentation performance for linear combinations of t ^ m a x with Φ 5 ’s major eigenvector ( ν 1 ): (a) a planar cross-section for the selected pair and (b) segmentation performance for a range of angles.
Sensors 25 00272 g015
Figure 16. Segmentation optimum for constraint combinations of t ^ m a x with Φ 4 ’s major eigenvector ( ν 1 ): (a) a planar cross-section for the selected pair and (b) segmentation performance for different angles.
Figure 16. Segmentation optimum for constraint combinations of t ^ m a x with Φ 4 ’s major eigenvector ( ν 1 ): (a) a planar cross-section for the selected pair and (b) segmentation performance for different angles.
Sensors 25 00272 g016
Figure 17. Several views with the best performing pixel, annotated “⊗”: (a) a map of all K b   >   2 pixels, (b) the segmentation map cropped at (a)’s dashed region, and (c) a PCA-based visualization of the cropped datacube.
Figure 17. Several views with the best performing pixel, annotated “⊗”: (a) a map of all K b   >   2 pixels, (b) the segmentation map cropped at (a)’s dashed region, and (c) a PCA-based visualization of the cropped datacube.
Sensors 25 00272 g017
Figure 18. The performance of Φ 5 ’s major eigenvector compared to its most similar pixel: (a) comparison of their two signatures and (b) performance analysis.
Figure 18. The performance of Φ 5 ’s major eigenvector compared to its most similar pixel: (a) comparison of their two signatures and (b) performance analysis.
Sensors 25 00272 g018
Table 1. RIT cube performance obtained with the optimal target under different constraints: (a) no constraint, (b) within the positive cone, (c) belongs to the data, and (d) belongs to a given set of real targets.
Table 1. RIT cube performance obtained with the optimal target under different constraints: (a) no constraint, (b) within the positive cone, (c) belongs to the data, and (d) belongs to a given set of real targets.
Constraint ( a )   t m a x ( b )   t   0 ( c )   t data ( d )   t targets
t h 0.0010.010.10.0010.010.10.0010.010.10.0010.010.1
K b 4.033.922.911.63
Δ p 4.524.063.584.183.9133.322.522.441.71.61.53
p max 0.1040.050.0180.170.060.0160.220.050.0234 × 10−413 × 10−46 × 10−4
NGMF ( A G )6 × 10−40.0060.0396 × 10−40.0060.0337 × 10−40.0120.0840.0050.0370.149
NSMF ( A L )0.2730.2520.2410.2570.2180.2010.2240.1540.2020.1350.1230.208
Benefit ( B )45340.66.1145136.46.133413.22.4226.83.321.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Furth, Y.; Rotman, S.R. Efficacy of Segmentation for Hyperspectral Target Detection. Sensors 2025, 25, 272. https://doi.org/10.3390/s25010272

AMA Style

Furth Y, Rotman SR. Efficacy of Segmentation for Hyperspectral Target Detection. Sensors. 2025; 25(1):272. https://doi.org/10.3390/s25010272

Chicago/Turabian Style

Furth, Yoram, and Stanley R. Rotman. 2025. "Efficacy of Segmentation for Hyperspectral Target Detection" Sensors 25, no. 1: 272. https://doi.org/10.3390/s25010272

APA Style

Furth, Y., & Rotman, S. R. (2025). Efficacy of Segmentation for Hyperspectral Target Detection. Sensors, 25(1), 272. https://doi.org/10.3390/s25010272

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop