Next Article in Journal
Deep Learning-Based Detection of Oil Spills in Pakistan’s Exclusive Economic Zone from January 2017 to December 2023
Previous Article in Journal
CVTNet: A Fusion of Convolutional Neural Networks and Vision Transformer for Wetland Mapping Using Sentinel-1 and Sentinel-2 Satellite Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Characterization of Water Composition with UAV-Based Hyperspectral Imaging and Generative Topographic Mapping

Hanson Center for Space Sciences, University of Texas at Dallas, Richardson, TX 75080, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(13), 2430; https://doi.org/10.3390/rs16132430
Submission received: 24 May 2024 / Revised: 25 June 2024 / Accepted: 29 June 2024 / Published: 2 July 2024
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)

Abstract

:
Unmanned aerial vehicles equipped with hyperspectral imagers have emerged as an essential technology for the characterization of inland water bodies. The high spectral and spatial resolutions of these systems enable the retrieval of a plethora of optically active water quality parameters via band ratio algorithms and machine learning methods. However, fitting and validating these models requires access to sufficient quantities of in situ reference data which are time-consuming and expensive to obtain. In this study, we demonstrate how Generative Topographic Mapping (GTM), a probabilistic realization of the self-organizing map, can be used to visualize high-dimensional hyperspectral imagery and extract spectral signatures corresponding to unique endmembers present in the water. Using data collected across a North Texas pond, we first apply GTM to visualize the distribution of captured reflectance spectra, revealing the small-scale spatial variability of the water composition. Next, we demonstrate how the nodes of the fitted GTM can be interpreted as unique spectral endmembers. Using extracted endmembers together with the normalized spectral similarity score, we are able to efficiently map the abundance of nearshore algae, as well as the evolution of a rhodamine tracer dye used to simulate water contamination by a localized source.

1. Introduction

Inland water bodies present a unique challenge to characterization by remote sensing imagery due to their complex spectral characteristics and small-scale spatial variability. The broad bands of multispectral imagers coupled with the irregular shape of lakes and rivers result in pixels with highly mixed signals that are easily dominated by reflectance from shore and nearshore vegetation sources [1,2]. Recently, the combination of hyperspectral imaging with unmanned aerial vehicles (UAVs), such as drones, has emerged as a powerful approach to simultaneously address the spectral, spatial, and temporal limitations of traditional high-altitude and satellite-based collection [3,4]. UAVs are significantly less expensive to deploy than their satellite- or aircraft-based remote sensing counterparts, and low-altitude flights enable centimeter-scale sampling while limiting the need for complicated atmospheric corrections [5]. However, the significant increase in the data volume generated by these systems presents a new challenge, namely, how to efficiently extract water quality parameters of interest from intricate pixel spectra.
Significant research efforts have focused on the development of techniques and algorithms to retrieve water quality parameters from UAV-captured hyperspectral images (HSI). Onboard computers installed alongside hyperspectral imagers enable the rapid evaluation of spectral indices from HSI band ratios [6]. These band ratios and polynomial combinations of bands have been used to successfully invert optically active water quality parameters such as turbidity directly from UAV-acquired imagery [7,8]. Supervised machine learning techniques such as tree-based models, support vector machines, and neural networks have also been used to estimate a wide range of parameters, such as colored dissolved organic matter, chlorophyll A, blue–green algae, and suspended sediment concentrations [9,10]. The calibration and evaluation of these data-driven models require a significant volume of coincident in situ data. This can be streamlined by coordinating UAV flights with reference data collection using autonomous robotic boats [11,12]. New proximal sensors for water quality parameters are continually being developed [13,14]. However, this approach relies on the prior knowledge of expected sources in order to select appropriate reference instruments for model validation. The presence of unanticipated contaminants cannot be directly identified in this sensing paradigm.
Extending the capabilities of UAV-based hyperspectral imaging to enable water quality monitoring in real-time scenarios where contaminant sources may not be known in advance requires two additional capabilities: dimensionality reduction techniques to permit the visual comparison of HSI and endmember extraction techniques which can identify spectral signatures corresponding to unique sources within the imaging scene. In remote sensing where reference data are typically sparse, many approaches have been explored. For example, principal component analysis (PCA) and t-distributed stochastic neighbor embedding (tSNE) are dimensionality reduction methods commonly used to reduce HSI to two or three dimensions for visualization [15,16]. Similarly for endmember extraction, there are a variety of established approaches including geometric methods like vertex component analysis, statistical methods like k-nearest neighbors and non-negative matrix factorization (NMF), and deep learning methods based on autoencoder architectures [17,18,19,20,21,22,23]. Methods based on linear mappings like PCA and NMF are often too restrictive for HSI, where the assumption of linear mixing is easily broken. However, the increased complexity of nonlinear methods like tSNE and autoencoders often lead to significant increases in computation time. An ideal approach should enable both visualization and nonlinear identification of relevant spectral endmembers.
The self-organizing map (SOM) developed by Teuvo Kohonen is an unsupervised machine learning method which nonlinearly maps high-dimensional data to the nodes of a two-dimensional grid [24]. By preserving the topological relationship between nodes during training, the SOM ensures that similar spectral signatures are mapped together such that related HSI pixels naturally cluster together. This presents an attractive compromise by enabling the simultaneous visualization of HSI data and endmember extraction via the weight vector associated with each SOM node [25,26,27]. When reference data are available, the SOM can be utilized to provide semisupervised labeling of HSI spectra [28]. Furthermore, Danielsen et al. demonstrated that the dimensionality reduction offered by the SOM can even be used for onboard data compression of HSIs acquired by a CubeSat [29]. Despite these clear capabilities, the SOM relies on a heuristic training algorithm with hyperparameters that can be challenging to tune and offers no direct probabilistic interpretation. To address these limitations, Bishop et al. developed generative topographic mapping (GTM), a probabilistic latent-variable model inspired by the SOM [30]. GTM has been utilized in a variety of domains, including drug design and chemical data visualization, but has yet to see adoption for the analysis of hyperspectral imagery [31,32,33].
In this paper, we explore the application of GTM to UAV-acquired HSI for the characterization water quality. Using data collected at a pond in Montague, North Texas, we first train a GTM using water-only pixels to produce a low-dimensional representation of the collected HSI. We use this mapping to explore the highly detailed small-scale variability within the pond and discuss how this can be used to guide reference data collection. Next, we demonstrate how GTM can be utilized to identify relevant spectral endmembers from a combined dataset including ground pixels, nearshore filamentous blue–green algae, water, and a simulated contaminant plume using a rhodamine tracer dye. Once identified, these endmembers can be used to rapidly map the abundance of spectral features within the pond. We demonstrate this capability by using GTM to map the abundance of filamentous blue–green algae near the shore, as well as the dispersion of a rhodamine dye plume.

2. Materials and Methods

In this study, we explore the use of GTM as a tool for dimensionality reduction and unsupervised endmember extraction of hyperspectral imagery. To this end, a dataset of HSI was collected at a pond in Montague, North Texas, on 23 November 2020 using a UAV-mounted hyperspectral imager configured as described in [11,12]. The pond spans an area <0.1 km 2 and has a maximum depth of 3 m. Across multiple flights, many data cubes like the example shown in panel a of Figure 1 were collected. Additionally, rhodamine dye was released and subsequently imaged to simulate a localized contaminant source. Using the HSI shown in Figure 1, a set of exemplar spectra were visually identified corresponding to nearshore filamentous blue–green algae, water, rhodamine dye, and the ground (a combination of soil and dry grass). The location of these samples is shown in panel b and their corresponding reflectance spectra are visualized in panel c. In the remainder of this section, we first provide a detailed overview of the GTM algorithm. Next, we describe the UAV platform used for HSI collection and the various steps employed to process captured HSI. Finally, we describe the two case studies presented in this paper for the utilization of GTM as an HSI visualization tool and as an endmember extraction technique.

2.1. Generative Topographic Mapping

GTM is a probabilistic latent-variable model inspired by the SOM for visualizing and clustering high-dimensional data. Like the SOM, GTM assumes vectors x in the d-dimensional data space (here representing reflectance spectra) are constrained to a low-dimensional embedded manifold. The SOM describes this manifold using a regular grid of nodes, each having an associated weight vector defining the mapping from the manifold to the data space. The position of each data record x in the manifold is assigned to the position of the node whose weight vector has the minimum Euclidean distance to x . An iterative training procedure updates the weight vectors of each node to fit the manifold to the data such that nodes near each other in the SOM grid correspond to similar records in the data space.
GTM mimics the grid of the SOM by assuming data are generated from latent variables ξ that are constrained to the K-many nodes of a regular grid. This assumption corresponds to establishing a prior distribution on the latent space of the form
p ( ξ ) = 1 K k K δ ( ξ ξ k )
where δ ( · ) is the Dirac delta function.
Points ξ in this latent space are mapped to the embedded manifold by a nonlinear function ψ parameterized by weights W. However, since real data are rarely noise-free, this embedded manifold will not be perfectly thin. To account for this, the points ξ are described in the data space by a radially symmetric Gaussian distribution, N ( ψ ( ξ ) , β 1 ) , with mean ψ ( ξ ) and precision β . For a dataset x n n = 1 N consisting of N-many records, this choice yields a log-likelihood function of the form
L ( W , β ) = n N ln 1 K k K p ( x n ξ k , W , β )
which can be maximized to obtain optimal W and β 1 .
The nonlinear mapping ψ is typically chosen to be given by the sum of M-many radial basis functions (RBF) evenly distributed in the latent space such that ψ ( ξ ) = W ϕ ( ξ ) with centers μ m and width σ so that
ϕ m ( ξ ) = exp ξ μ m 2 2 σ 2 .
The width σ is taken to be the distance between neighboring RBF centers multiplied by a scale factor, s. Together, the number of RBFs, M, and s are hyperparameters for the model which governs the smoothness of the resulting manifold in the data space. A visual representation of the GTM is shown in Figure 2. Additionally, sparsity of W can be enforced by introducing an additional hyperparameter α corresponding to a prior distribution over the weights given by
p ( W α ) = α 2 π M D / 2 exp α 2 W F 2 .
The occurrence of the sum within the natural logarithm in Equation (2) prevents an analytic solution for the maximum of L . However, the form of ψ enables an Expectation–Maximization routine which optimizes L by iteratively increasing a lower bound given by the expected complete-data log-likelihood [34]. During the expectation step of the fitting process, the posterior probability p ( ξ k x n , W , β ) , i.e., the responsibility of the kth node for explaining the nth record, is computed as
R k n = p ( ξ k x n , W , β ) = p ( x n ξ k , W , β ) k K p ( x n ξ k , W , β ) .
Together, these responsibilities form a matrix with entries R k n which are kept fixed during the maximization step. The maximization step is then performed by updating the weights W and variance β 1 according to
W new = Φ T G Φ + α β I 1 Φ T R X
1 β new = 1 N D n N k K R k n ψ k x n 2
where Φ k m = ϕ m ( ξ k ) , X is the data matrix formed by concatenating the records x n , and G is a diagonal matrix with G k k = n N R k n . This process is repeated until the log-likelihood converges to a predetermined tolerance level.
Slices R [ : , n ] of the matrix R define the responsibility of each latent node ξ k for the nth data record x n . Therefore, the final responsibility matrix after GTM training can be used to represent each record in the latent space by the following mean:
ξ ^ n = k K R k n ξ k .
A freely available implementation of the GTM algorithm was developed for this study and is accessible at [35]. The code is written in the Julia programming language and complies with the Machine Learning in Julia (MLJ) common interface [36,37].

2.2. UAV-Based Hyperspectral Imaging

A Freefly Alta-X autonomous quadcopter was used as the UAV platform in this study. This UAV was equipped with a Resonon Pika XC2 visible+near-infrared (VNIR) hyperspectral imager to capture HSI with 462 wavelengths per pixel ranging from 391 to 1011 nm. This imager is in a pushbroom configuration so that HSI are captured one scan-line at a time resulting in data cubes consisting of 1000 scan-lines with 1600 pixels each. Additionally, the camera includes an embedded GPS/INS unit to enable georectification of collected HSI. An upward-facing Ocean Optics UV-Vis NIR spectrometer with a cosine corrector was also included to provide measurements of the incident solar irradiance spectrum. The configuration of the UAV with the attached hyperspectral imager is shown in Figure 3. Data collection and processing were controlled by an attached Intel NUC small-form-factor computer. A detailed description of the UAV platform can be found in [11].
To account for the variability of incident light, raw HSI are converted to units of reflectance using the downwelling irradiance spectrum simultaneously captured with each HSI. With the hyperspectral imager oriented to nadir, the reflectance is given by
ρ ( λ ) = π L ( λ ) / E d ( λ )
where L is the spectral radiance, E d is the downwelling irradiance, and a factor of π steradians results from assuming Lambertian (diffuse) upwelling radiance [38]. HSI collection was performed near solar noon to maximize the amount of incident sunlight illuminating the water. For the site in North Texas, this corresponded to an average solar zenith angle of 54.9 for 23 November 2020. During flights, the hyperspectral imager was oriented to nadir, resulting in HSI with negligible sunglint effects.
After conversion to reflectance, each HSI must also be georectified to assign geographic coordinates to each pixel. The UAV flights were performed at an approximate altitude of 50 m above the water so that the imaging surface can be considered to be flat. Consequently, HSI were rapidly georectified to a 10 cm resolution using position and orientation data from the embedded GPS/INS, as outlined in refs. [39,40,41]. An example hyperspectral data cube is illustrated in Figure 1a where the log10-reflectance values are plotted along the z-axis for each pixel in the scene, and a psuedocolor image at the top of the data cube illustrates how the water would appear to the human eye from the perspective of the UAV.
As a final processing step before training each GTM, we limit the wavelengths of each HSI to λ 900 nm, as wavelengths above 900 nm showed significant noise. Additionally, each spectrum was rescaled to a peak value of 1.0 to account for incident light variability between HSI.

2.3. GTM Case Studies

To explore the ability of the GTM to segment HSI pixels and aid in source identification, we first consider a dataset of water-only spectra consisting of >36,000 records identified from collected HSI by a normalized difference water index (NDWI) greater than 0.25 , where the NDWI is defined as
NDWI = ρ ( 550 ) ρ ( 860 ) ρ ( 550 ) + ρ ( 860 )
where λ = 550 and λ = 860 nm were chosen to represent the green and NIR bands as described in ref. [42]. We then apply the trained GTM to all water-only pixels to visualize the distribution learned by the GTM and examine spectra associated with a subset of nodes in the latent space in order to assess the small-scale spatial variability within the pond.
Next, we consider a dataset of combined HSI pixels including both water and land. To simulate the dispersion of a potential contaminant source, a rhodamine tracer dye was released into the western portion of the pond and two additional UAV flights were used to collect HSI capturing the evolution of the resulting plume. From these HSI, a collection of >145,000 pixels were sampled for model training. To demonstrate the ability to extract relevant spectral endmembers from the trained GTM, we utilize the exemplar spectra corresponding to open water, ground, nearshore algae, and rhodamine dye visually identified in Figure 1. The responsibility matrix obtained by the trained GTM provides a probability distribution over the latent space nodes for every data record. We identify relevant endmembers for each reference spectrum via the set of nodes with nonzero responsibility. We then utilize the mapping ψ to obtain the spectral representation of these nodes. Since the GTM is unsupervised, these extracted endmembers do not correspond to individual records but rather spectral features learned by the GTM.
Values for the hyperparameters m, s, and α are determined by training multiple GTM models and selecting values which minimize the Bayesian Information Criterion (BIC) given by
BIC = P ln ( N ) 2 L
where P is the total number of model parameters, N is the number of records in the dataset, and L is the log-likelihood defined in Equation (2).
The normalized spectral similarity score (NS3) introduced by Nidamanuri and Zbell combines the root mean square (RMS) difference together with the spectral angle to provide a spectral distance function [43]. For two spectra ρ 1 ( λ ) and ρ 2 ( λ ) , it is defined by
NS 3 ( ρ 1 , ρ 2 ) = R M S ( ρ 1 , ρ 2 ) 2 + ( 1 cos θ ) 2
where
RMS ( ρ 1 , ρ 2 ) = 1 N 1 λ ρ 1 ( λ ) ρ 2 ( λ ) 2
cos θ = ρ 1 , ρ 2 ρ 1 ρ 2 .
We use the NS3 together with extracted GTM endmembers to map the abundance of filamentous blue–green algae near the shore, as well as the evolution of the rhodamine dye plume. To do this, the GTM node with maximum responsibility is selected from the endmembers identified for the algae and rhodamine reference spectra. The distribution of NS3 values between each endmember and all HSI pixels is then used to select a suitable cutoff threshold. The area subsumed by HSI pixels with an NS3 below this threshold is then used to infer the spatial extent of algae and rhodamine in the pond.

3. Results

3.1. Water-Only Pixel Segmentation

A GTM with K = 32 × 32 nodes was trained on the dataset of water-only pixels described in Section 2.3 in order to explore the distribution of reflectance spectra captured by the HSI. The resulting GTM is visualized in Figure 4. First, we note that the GTM has utilized most of the latent space as illustrated Figure 4b where the position corresponding to the mean responsibility of each data point ξ ^ n has been plotted. For this water-only GTM, the spectra appear to be largely clustered toward the left edge, bottom, and right edge of the latent space. Spectral signatures corresponding to GTM nodes from the four corners and center of the latent space are shown in Figure 4c, illustrating the spectral variability represented across the latent space.
To visualize the distribution of HSI learned by the GTM, we can associate a color with each dimension of the latent space. In Figure 4, we used the red channel to represent ξ 1 and the blue channel to represent ξ 2 . Applying the trained GTM to compute mean node responsibilities for all water pixels in collected HSI allows us to illustrate the spatial distribution of the spectra on a map, as shown in Figure 4c. Here, we observe a clear distinction between water near the shore and water in the middle of the pond. Additionally, the eastern alcove of the pond is significantly more blue than the rest of the water. We note that flow through this region is restricted by the small entrance near the cyan square shown in panel a. However, despite its similar depth and surface characteristics, the GTM distinguished this region from the rest of the water.
The close proximity of highly dissimilar GTM classes illustrated by sharp color gradients in the map reflects the small-scale spatial variability typical of inland water bodies like this pond. For example, water near the cyan diamond in panel a of Figure 4 includes both blue and red pixels in close proximity. As shown in Figure 4b, these colors correspond to the top-left and bottom-right corners of the GTM latent space. As the GTM model maps neighboring nodes to similar spectra, this region includes maximally dissimilar pixels. Furthermore, this point is away from the shore where the depth is approximately uniform. The precision parameter β learned by the GTM enables the model to account for spectral variability so that effects from localized disturbances such as surface ripples are constrained to small displacements in the latent space. With these considerations in mind, contrast in this region of the pond therefore reflects real variability of water composition, surface sediments, and vegetation.

3.2. Endmember Extraction

A second GTM was trained on the combined dataset of HSI with pixels covering shore, water, and the rhodamine tracer dye release. The resulting latent space distribution is shown in Figure 5, where each sampleis plotted at the position of the mean node responsibility and colored by the normalized difference vegetation index (NDVI) computed from the original reflectance spectrum. The NDVI is a spectral index sensitive to variations in vegetation health with negative values corresponding to water, small positive values corresponding to sparse vegetation, and large values near 1 corresponding to dense vegetation [44]. From this, we see that the GTM can clearly separate spectra by vegetation content with high values concentrated to the left of the latent space and negative, water-based pixels concentrated to the right. Additionally, the position of exemplar spectra for algae, rhodamine, water, and ground points are included as color-filled circles, further illustrating the separation of distinct sources obtained by the GTM. We note that a portion of the nearshore algae was observed to float at the surface of the water near the shore. Consequently, the position of the exemplar spectrum in Figure 5 used for algae occurs at a non-negative NDVI. The accumulation of records with negative NDVI values in close proximity (in the latent space) to the selected algae spectrum reflects the presence of algae within the water.
As outlined in Section 2.1, there are three hyperparameters that need to be chosen to fit a GTM model: the number of RBF centers M = m2 with m the number along each axis, the scale factor s which controls the RBF overlap, and the regularization parameter α . To choose appropriate values, we performed a grid search for m values between 2 and 20, s values between 0.1 and 3.0 , and α values between 0.001 and 10.0 . The best values were determined as those which minimized the BIC, with m = 14 , s = 1.0 , and α = 0.1 , respectively. Heatmaps comparing the BIC for different hyperparameter values are provided in Figure 6, and a table of the top 25 models is given in Appendix A. Since the matrix of RBF activations Φ need only be computed once at the GTM initialization step, the number of latent nodes given by K = k 2 can be chosen to be large enough to ensure a smooth mapping. We found that a value of k = 32 provided a sufficient number of GTM nodes without significantly impacting training time.
Examining the latent space distribution learned by the GTM provides an unsupervised method to extract endmember spectra corresponding to unique sources within the imaging scene. To demonstrate this clearly, we consider the exemplar spectra visually identified in Figure 1. Using the trained GTM, we can compute spectral signatures for each node in the latent space via the nonlinear mapping ψ . The responsibilities R k n therefore correspond to the contributions of the k t h node ξ k to the nth sample spectrum in the dataset. Endmembers are identified by nodes with nonzero responsibility values for each of the selected exemplar spectra and are plotted in Figure 7. We note that these extracted endmembers are able to accurately capture the shape of each exemplar spectrum including thin reflectance peaks. The GTM has also managed to smoothly interpolate through noisy wavelengths, as shown by the water and algae endmembers for λ > 800 nm.

3.3. Abundance Mapping with NS3

Spectral endmembers extracted from the GTM can be used to map the abundance of water constituents in HSI by identifying pixels with a NS3 value below a given threshold. In Figure 8, we demonstrate this by using the extracted algae endmember spectrum to map the abundance of filamentous blue–green algae near the western shore of the pond. The distribution of NS3 values was multimodal with a majority of pixels above 0.4 belonging to dissimilar spectra. Consequently, a cutoff threshold of 0.4275 was chosen to distinguish algae values corresponding to the 20th-quantile of the NS3 distribution. The estimated area subsumed by algae is shown in Figure 8 and calculated to be 587.3 m2. Similarly, we are able to track the evolution of the rhodamine dye release into the western portion of the pond by computing the NS3 for HSI collected across multiple flights. Here, the same cutoff threshold was used corresponding to the 33rd-quantile of the NS3 distribution. In the left panel of Figure 9, we see that the dye plume initially encompasses an area of roughly 330.7 m2. A second flight performed 15 min later reveals the dye plume to have increased to an area of 1495.6 m2. The increase in NS3 values between flights reflects the dilution of dye as it diffused into the water.

4. Discussion

The application of UAV-based hyperspectral imaging (HSI) for water quality assessment is gaining significant traction due to its ability to efficiently capture detailed spectral data at high spatial resolutions. Most studies using UAVs have employed supervised techniques that map the captured spectra to specific water quality parameters of interest. For example, Lu et al. evaluated a variety of machine learning methods for the extraction of chlorophyll-a and suspended sediment concentrations from HSI using data from UAV flights over 33 sampling locations [10]. However, this approach relies on the collection of paired in situ data, which can be challenging to obtain in sufficient quantities to facilitate model training. Additionally, supervised methods require prior knowledge of expected sources in order to select and calibrate suitable reference instruments. Using models purpose-built for specific water quality parameters like chlorophyll-a or turbidity discards potential information contained in HSI, which can reveal unanticipated sources. Therefore, unsupervised methods which aid in the visualization of HSI and enable the identification of spectral endmembers within the imaging scene are needed to complement these supervised approaches.
Recently, some researchers have begun to explore endmember extraction using UAV-based imaging where the increased spatial resolution provided by UAV platforms is hypothesised to yield more pure pixels than their satellite counterparts. For instance, Alvarez et al. used multispectral UAV imagery to extract endmembers which were then applied to unmix remote sensing data for plant abundance estimation across a broad region of France [45]. Similarly, Gu et al. explored UAV-to-satellite hyperspectral unmixing by applying endmembers extracted from UAV-based HSI to satellite observations [46]. These studies underscore the potential of combining endmember extraction techniques with UAV-based HSI, which has yet to see widespread adoption for water quality analysis.
In this study, we explored the GTM as a unsupervised approach to simultaneously enable visualization of HSI and perform endmember extraction of spectral signatures. First, we showed how the representation of data in the latent space of the GTM given by the mean responsibility can be used to visualize the small-scale structures within inland water bodies as evidenced in Figure 4. In particular, we note that the sharp gradients in colors found when visualizing the spatial distribution of GTM nodes across the pond reflects significant spectral variability at the submeter scale. This has important consequences for water quality assessment where the location of in situ data collection will have a strong impact on the ability of any model to predict water quality parameters from HSI. Visualizing the distribution of spectra from collected HSI is therefore highly relevant to guide in situ data collection, as shifting the sampling location by as little as a meter can lead to significant differences in measured values.
Based on these observations, one clear application of the GTM for real-time water quality assessment is for intelligently guided reference data collection. In our previous work, we showed that coordinating UAV-based hyperspectral imaging with in situ data collection by an autonomous boat can dramatically improve the inversion of water quality parameters from HSI pixels [12]. However, the spatial distributions of parameters such as chlorophyll-a, blue–green algae, and temperature are often highly dissimilar posing a challenge for optimal route planning. Since GTM estimates the full distribution of reflectance spectra and not a single water quality parameter, one could construct a prize collecting travelling salesman problem (PC-TSP) which seeks to find the optimal route maximizing the area explored in the GTM latent space [47]. Similar approaches have been used to guide data collection with autonomous vehicles to optimize data quality subject to resource constraints [48].
The second application of the GTM presented in this study is for the unsupervised extraction of endmembers corresponding to unique sources observed in the HSI. Here, the GTM is an attractive choice, as it does not rely on the assumptions of linear mixing and the presence of pure pixels which are easily broken by the effects of multiple scattering in realistic scenarios [49]. Additionally, because the GTM is a probabilistic model, the values of model hyperparameters can be selected objectively using information criteria like the BIC. This is a clear advantage over similar methods like the SOM, on which the GTM is based. Endmembers identified for exemplar spectra corresponding to rhodamine, algae, ground, and water pixels demonstrate that the method can successfully identify spectral signatures corresponding to diverse sources. If a set of labeled spectra for known sources are available, their representation in the GTM latent space could be used to perform a semisupervised classification, similar to the SOM approach developed by Riese et al. [28]. For example, in clear waters where variation in sediments and vegetation at the pond floor will contribute to observed spectral variability, the distribution learned by the GTM can be paired with in situ data to classify the surface type.
Once endmembers are identified, estimating their abundance using the NS3 provides a quick method to map the distribution of sources in water. We note that the spatial distribution of algae mapped in Figure 8 realistically reflects clustering of algae near the shore. Additionally, the ability for the UAV to quickly survey the same area in rapid succession is highly advantageous for tracking the diffusion of point sources, as demonstrated in Figure 9 for the rhodmaine dye release.
The primary limitation of the GTM is that the number of nodes increases exponentially with the dimension of the latent space. However, when constrained to two dimensions, in order to enable visualization of the resulting map, the number of latent nodes has a negligible impact compared with the size of the dataset on which it is trained. Additionally, the GTM considers individual HSI pixels and does not exploit spatial structure like other methods, such as convolutional autoencoders and non-negative matrix factorization using superpixels [19,23]. Extensions to GTM have been proposed to enable batch training for large datasets, as well as manifold-aligned noise models which replace the fixed precision parameter β with a full covariance matrix [50].
Finally, we note that the representation obtained by GTM can be used for nonlinear feature extraction to improve supervised models. Traditionally, PCA is used as a common preprocessing technique to reduce high-dimensional HSI by keeping the first r-many principal components. For example, Uddin et al. report improved classification of HSI by using PCA to extract features for a support vector machine [51]. GTM can similarly be used to provide a sparse representation of the input data via the latent node responsibilities R k n obtained for each record.

5. Conclusions

In this study, we present GTM as a useful unsupervised method for the visualization of UAV-based hyperspectral imagery and associated extraction of spectral endmembers. Using data collected at a North Texas pond, we demonstrate how the latent space of the GTM can be used to visualize the distribution of observed reflectance spectra revealing the small-scale spatial variability of water composition. Spectral signatures extracted from GTM nodes are used to successfully map the abundance of algae near the shore and to track the evolution of a rhodamine tracer dye plume. These examples illustrate the power of combining unsupervised learning with UAV-based hyperspectral imaging for the characterization of water composition. Future work will further develop the GTM as a tool to guide in situ data collection and enable contaminant localization for real-time applications.

Author Contributions

Methodology, J.W. and D.J.L.; conceptualization, D.J.L.; software, J.W.; field deployment and preparation, J.W., A.A., L.O.H.W., S.T., A.F., P.M.H.D., M.I., M.L., D.S., G.B. and D.J.L.; validation, J.W.; formal analysis, J.W.; investigation J.W.; resources, D.J.L.; data curation, J.W., A.A., L.O.H.W. and D.J.L.; writing—original draft preparation, J.W.; writing—review and editing, J.W. and D.J.L.; visualization, J.W.; supervision, D.J.L.; project administration, D.J.L.; funding acquisition, D.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the following grants: the Texas National Security Network Excellence Fund award for Environmental Sensing Security Sentinels; the SOFWERX award for Machine Learning for Robotic Teams and NSF Award OAC-2115094; support from the University of Texas at Dallas Office of Sponsored Programs, Dean of Natural Sciences and Mathematics, and Chair of the Physics Department is gratefully acknowledged; TRECIS CC* Cyberteam (NSF #2019135); NSF OAC-2115094 Award; and EPA P3 grant number 84057001-0.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are made openly available via the Open Storage Network at https://ncsa.osn.xsede.org/ees230012-bucket01/RobotTeam/unsupervised/gtm-data/, accessed on 19 February 2024. The open-source implementation of the GTM used in this study is available at https://github.com/john-waczak/GenerativeTopographicMapping.jl, accessed on 19 February 2024.

Acknowledgments

Don MacLaughlin, Scotty MacLaughlin, and the city of Plano, TX, are gratefully acknowledged for allowing us to deploy the autonomous robot team on their property. Christopher Simmons is gratefully acknowledged for his computational support. We thank Antonio Mannino for his advice with regard to selecting the robotic boat’s sensing suite. Annette Rogers is gratefully acknowledged for supporting the arrangement of insurance coverage. Steven Lyles is gratefully acknowledged for supporting the arrangement of a secure place for the robot team. The authors acknowledge the OIT-Cyberinfrastructure Research Computing group at the University of Texas at Dallas and the TRECIS CC* Cyberteam (NSF #2019135) for providing HPC resources that contributed to this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
GTMGenerative Topographic Mapping
SOMSelf Organizing Map
HSIHyperspectral Image
PCAPrincipal Component Analysis
tSNEt-Distributed Stochastic Neighbor Embedding
MLJMachine Learning in Julia
VNIRVisible + Near-Infrared
NDWINormalized Difference Water Index
NS3Normalized Spectral Similarity Score

Appendix A. Hyperparameter Search Results

Table A1. The top 25 models from the hyperparameter search. A variety of GTM were trained to explore the the impact of varying m, α , and s. The Bayesian Information Criterion (BIC) and Akaike Information Criterion (AIC) are given in the final two columns which can be used for hyperparameter selection.
Table A1. The top 25 models from the hyperparameter search. A variety of GTM were trained to explore the the impact of varying m, α , and s. The Bayesian Information Criterion (BIC) and Akaike Information Criterion (AIC) are given in the final two columns which can be used for hyperparameter selection.
m α skBICAIC
140.11.032 1 . 918 × 10 8 1 . 926 × 10 8
130.011.032 1.917 × 10 8 1.923 × 10 8
160.011.532 1.917 × 10 8 1.926 × 10 8
1410.01.032 1.917 × 10 8 1.924 × 10 8
160.0011.532 1.917 × 10 8 1.926 × 10 8
131.01.032 1.917 × 10 8 1.923 × 10 8
1310.01.032 1.917 × 10 8 1.923 × 10 8
140.0011.532 1.916 × 10 8 1.924 × 10 8
130.11.032 1.916 × 10 8 1.923 × 10 8
140.011.032 1.916 × 10 8 1.924 × 10 8
150.011.532 1.916 × 10 8 1.925 × 10 8
140.011.532 1.916 × 10 8 1.923 × 10 8
151.01.032 1.916 × 10 8 1.924 × 10 8
180.011.532 1.916 × 10 8 1.928 × 10 8
120.011.032 1.916 × 10 8 1.921 × 10 8
150.010.532 1.915 × 10 8 1.924 × 10 8
171.01.032 1.915 × 10 8 1.926 × 10 8
160.11.032 1.915 × 10 8 1.925 × 10 8
180.0011.532 1.915 × 10 8 1.928 × 10 8
130.0011.032 1.915 × 10 8 1.922 × 10 8
121.01.032 1.915 × 10 8 1.921 × 10 8
170.0011.532 1.915 × 10 8 1.926 × 10 8
150.0011.532 1.915 × 10 8 1.923 × 10 8
1510.01.032 1.915 × 10 8 1.923 × 10 8
120.11.532 1.915 × 10 8 1.92 × 10 8

References

  1. Koponen, S.; Pulliainen, J.; Kallio, K.; Hallikainen, M. Lake water quality classification with airborne hyperspectral spectrometer and simulated MERIS data. Remote. Sens. Environ. 2002, 79, 51–59. [Google Scholar] [CrossRef]
  2. Ritchie, J.C.; Zimba, P.V.; Everitt, J.H. Remote sensing techniques to assess water quality. Photogramm. Eng. Remote. Sens. 2003, 69, 695–704. [Google Scholar] [CrossRef]
  3. Adão, T.; Hruška, J.; Pádua, L.; Bessa, J.; Peres, E.; Morais, R.; Sousa, J.J. Hyperspectral imaging: A review on UAV-based sensors, data processing and applications for agriculture and forestry. Remote. Sens. 2017, 9, 1110. [Google Scholar] [CrossRef]
  4. Arroyo-Mora, J.P.; Kalacska, M.; Inamdar, D.; Soffer, R.; Lucanus, O.; Gorman, J.; Naprstek, T.; Schaaf, E.S.; Ifimov, G.; Elmer, K.; et al. Implementation of a UAV–hyperspectral pushbroom imager for ecological monitoring. Drones 2019, 3, 12. [Google Scholar] [CrossRef]
  5. Banerjee, B.P.; Raval, S.; Cullen, P. UAV-hyperspectral imaging of spectrally complex environments. Int. J. Remote. Sens. 2020, 41, 4136–4159. [Google Scholar] [CrossRef]
  6. Horstrand, P.; Guerra, R.; Rodríguez, A.; Díaz, M.; López, S.; López, J.F. A UAV platform based on a hyperspectral sensor for image capturing and on-board processing. IEEE Access 2019, 7, 66919–66938. [Google Scholar] [CrossRef]
  7. Vogt, M.C.; Vogt, M.E. Near-remote sensing of water turbidity using small unmanned aircraft systems. Environ. Pract. 2016, 18, 18–31. [Google Scholar] [CrossRef]
  8. Zhang, D.; Zeng, S.; He, W. Selection and quantification of best water quality indicators using UAV-mounted hyperspectral data: A case focusing on a local river network in Suzhou City, China. Sustainability 2022, 14, 16226. [Google Scholar] [CrossRef]
  9. Keller, S.; Maier, P.M.; Riese, F.M.; Norra, S.; Holbach, A.; Börsig, N.; Wilhelms, A.; Moldaenke, C.; Zaake, A.; Hinz, S. Hyperspectral data and machine learning for estimating CDOM, chlorophyll a, diatoms, green algae and turbidity. Int. J. Environ. Res. Public Health 2018, 15, 1881. [Google Scholar] [CrossRef]
  10. Lu, Q.; Si, W.; Wei, L.; Li, Z.; Xia, Z.; Ye, S.; Xia, Y. Retrieval of water quality from UAV-borne hyperspectral imagery: A comparative study of machine learning algorithms. Remote. Sens. 2021, 13, 3928. [Google Scholar] [CrossRef]
  11. Lary, D.J.; Schaefer, D.; Waczak, J.; Aker, A.; Barbosa, A.; Wijeratne, L.O.; Talebi, S.; Fernando, B.; Sadler, J.; Lary, T.; et al. Autonomous learning of new environments with a robotic team employing hyper-spectral remote sensing, comprehensive in-situ sensing and machine learning. Sensors 2021, 21, 2240. [Google Scholar] [CrossRef]
  12. Waczak, J.; Aker, A.; Wijeratne, L.O.; Talebi, S.; Fernando, B.; Hathurusinghe, P.; Iqbal, M.; Schaefer, D.; Lary, D.J. Characterizing Water Composition with an Autonomous Robotic Team Employing Comprehensive In-Situ Sensing, Hyperspectral Imaging, Machine Learning, and Conformal Prediction. Remote. Sens. 2024, 16, 996. [Google Scholar] [CrossRef]
  13. Parra, L.; Ahmad, A.; Sendra, S.; Lloret, J.; Lorenz, P. Combination of Machine Learning and RGB Sensors to Quantify and Classify Water Turbidity. Chemosensors 2024, 12, 34. [Google Scholar] [CrossRef]
  14. Chirchi, V.; Chirchi, E.; Khushi, E.C.; Bairavi, S.M.; Indu, K.S. Optical Sensor for Water Bacteria Detection using Machine Learning. In Proceedings of the 2024 11th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 28 February–1 March 2024; pp. 603–608. [Google Scholar] [CrossRef]
  15. Tyo, J.S.; Konsolakis, A.; Diersen, D.I.; Olsen, R.C. Principal-components-based display strategy for spectral imagery. IEEE Trans. Geosci. Remote. Sens. 2003, 41, 708–718. [Google Scholar] [CrossRef]
  16. Zhang, B.; Yu, X. Hyperspectral image visualization using t-distributed stochastic neighbor embedding. In Proceedings of the MIPPR 2015: Remote Sensing Image Processing, Geographic Information Systems, and Other Applications, Enshi, China, 31 October–1 November 2015; Volume 9815, pp. 14–21. [Google Scholar]
  17. Heylen, R.; Parente, M.; Gader, P. A review of nonlinear hyperspectral unmixing methods. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2014, 7, 1844–1868. [Google Scholar] [CrossRef]
  18. Nascimento, J.M.; Dias, J.M. Vertex component analysis: A fast algorithm to unmix hyperspectral data. IEEE Trans. Geosci. Remote. Sens. 2005, 43, 898–910. [Google Scholar] [CrossRef]
  19. Feng, X.R.; Li, H.; Wang, R.; Du, Q.; Jia, X.; Plaza, A.J. Hyperspectral Unmixing Based on Nonnegative Matrix Factorization: A Comprehensive Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2022, 15, 4414–4436. [Google Scholar] [CrossRef]
  20. Cariou, C.; Chehdi, K. Unsupervised nearest neighbors clustering with application to hyperspectral images. IEEE J. Sel. Top. Signal Process. 2015, 9, 1105–1116. [Google Scholar] [CrossRef]
  21. Su, Y.; Li, J.; Plaza, A.; Marinoni, A.; Gamba, P.; Chakravortty, S. DAEN: Deep autoencoder networks for hyperspectral unmixing. IEEE Trans. Geosci. Remote. Sens. 2019, 57, 4309–4321. [Google Scholar] [CrossRef]
  22. Borsoi, R.A.; Imbiriba, T.; Bermudez, J.C.M. Deep generative endmember modeling: An application to unsupervised spectral unmixing. IEEE Trans. Comput. Imaging 2019, 6, 374–384. [Google Scholar] [CrossRef]
  23. Palsson, B.; Ulfarsson, M.O.; Sveinsson, J.R. Convolutional autoencoder for spectral–spatial hyperspectral unmixing. IEEE Trans. Geosci. Remote. Sens. 2020, 59, 535–549. [Google Scholar] [CrossRef]
  24. Kohonen, T. The self-organizing map. Proc. IEEE 1990, 78, 1464–1480. [Google Scholar] [CrossRef]
  25. Cantero, M.; Perez, R.; Martinez, P.J.; Aguilar, P.; Plaza, J.; Plaza, A. Analysis of the behavior of a neural network model in the identification and quantification of hyperspectral signatures applied to the determination of water quality. In Proceedings of the Chemical and Biological Standoff Detection II SPIE, Philadelphia, PA, USA, 27–28 October 2004; Volume 5584, pp. 174–185. [Google Scholar]
  26. Duran, O.; Petrou, M. A time-efficient method for anomaly detection in hyperspectral images. IEEE Trans. Geosci. Remote. Sens. 2007, 45, 3894–3904. [Google Scholar] [CrossRef]
  27. Ceylan, O.; Kaya, G.T. Feature Selection Using Self Organizing Map Oriented Evolutionary Approach. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 4003–4006. [Google Scholar]
  28. Riese, F.M.; Keller, S.; Hinz, S. Supervised and semi-supervised self-organizing maps for regression and classification focusing on hyperspectral data. Remote. Sens. 2019, 12, 7. [Google Scholar] [CrossRef]
  29. Danielsen, A.S.; Johansen, T.A.; Garrett, J.L. Self-organizing maps for clustering hyperspectral images on-board a cubesat. Remote. Sens. 2021, 13, 4174. [Google Scholar] [CrossRef]
  30. Bishop, C.M.; Svensén, M.; Williams, C.K. GTM: The generative topographic mapping. Neural Comput. 1998, 10, 215–234. [Google Scholar] [CrossRef]
  31. Kireeva, N.; Baskin, I.; Gaspar, H.; Horvath, D.; Marcou, G.; Varnek, A. Generative topographic mapping (GTM): Universal tool for data visualization, structure-activity modeling and dataset comparison. Mol. Inform. 2012, 31, 301–312. [Google Scholar] [CrossRef]
  32. Gaspar, H.A.; Baskin, I.I.; Marcou, G.; Horvath, D.; Varnek, A. Chemical data visualization and analysis with incremental generative topographic mapping: Big data challenge. J. Chem. Inf. Model. 2015, 55, 84–94. [Google Scholar] [CrossRef]
  33. Horvath, D.; Marcou, G.; Varnek, A. Generative topographic mapping in drug design. Drug Discov. Today Technol. 2019, 32, 99–107. [Google Scholar] [CrossRef]
  34. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B 1977, 39, 1–22. [Google Scholar] [CrossRef]
  35. Waczak, J. GenerativeTopographicMapping.jl. 2024. Available online: https://zenodo.org/records/11061258 (accessed on 24 April 2024).
  36. Bezanson, J.; Karpinski, S.; Shah, V.B.; Edelman, A. Julia: A fast dynamic language for technical computing. arXiv 2012, arXiv:1209.5145. [Google Scholar]
  37. Blaom, A.D.; Kiraly, F.; Lienart, T.; Simillides, Y.; Arenas, D.; Vollmer, S.J. MLJ: A Julia package for composable machine learning. arXiv 2020, arXiv:2007.12285. [Google Scholar] [CrossRef]
  38. Ruddick, K.G.; Voss, K.; Banks, A.C.; Boss, E.; Castagna, A.; Frouin, R.; Hieronymi, M.; Jamet, C.; Johnson, B.C.; Kuusk, J.; et al. A review of protocols for fiducial reference measurements of downwelling irradiance for the validation of satellite remote sensing data over water. Remote. Sens. 2019, 11, 1742. [Google Scholar] [CrossRef]
  39. Muller, R.; Lehner, M.; Muller, R.; Reinartz, P.; Schroeder, M.; Vollmer, B. A program for direct georeferencing of airborne and spaceborne line scanner images. Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci. 2002, 34, 148–153. [Google Scholar]
  40. Bäumker, M.; Heimes, F. New calibration and computing method for direct georeferencing of image and scanner data using the position and angular data of an hybrid inertial navigation system. In Proceedings of the OEEPE Workshop, Integrated Sensor Orientation, Hannover, Germany, 17–18 September 2001; pp. 1–17. [Google Scholar]
  41. Mostafa, M.M.; Schwarz, K.P. A multi-sensor system for airborne image capture and georeferencing. Photogramm. Eng. Remote. Sens. 2000, 66, 1417–1424. [Google Scholar]
  42. McFeeters, S.K. The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. Int. J. Remote. Sens. 1996, 17, 1425–1432. [Google Scholar] [CrossRef]
  43. Nidamanuri, R.R.; Zbell, B. Normalized Spectral Similarity Score (NS3) as an Efficient Spectral Library Searching Method for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2011, 4, 226–240. [Google Scholar] [CrossRef]
  44. Thenkabail, P.S.; Lyon, J.G.; Huete, A. Hyperspectral Indices and Image Classifications for Agriculture and Vegetation; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  45. Alvarez-Vanhard, E.; Houet, T.; Mony, C.; Lecoq, L.; Corpetti, T. Can UAVs fill the gap between in situ surveys and satellites for habitat mapping? Remote. Sens. Environ. 2020, 243, 111780. [Google Scholar] [CrossRef]
  46. Gu, Y.; Huang, Y.; Liu, T. Intrinsic Decomposition Embedded Spectral Unmixing for Satellite Hyperspectral Images with Endmembers From UAV Platform. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5523012. [Google Scholar] [CrossRef]
  47. Balas, E. The prize collecting traveling salesman problem and its applications. In The Traveling Salesman Problem and Its Variations; Springer: Berlin/Heidelberg, Germany, 2007; pp. 663–695. [Google Scholar]
  48. Suryan, V.; Tokekar, P. Learning a spatial field in minimum time with a team of robots. IEEE Trans. Robot. 2020, 36, 1562–1576. [Google Scholar] [CrossRef]
  49. Han, T.; Goodenough, D.G. Investigation of nonlinearity in hyperspectral remotely sensed imagery—A nonlinear time series analysis approach. In Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, Barcelona, Spain, 23–28 July 2007; pp. 1556–1560. [Google Scholar] [CrossRef]
  50. Bishop, C.M.; Svensén, M.; Williams, C.K. Developments of the generative topographic mapping. Neurocomputing 1998, 21, 203–224. [Google Scholar] [CrossRef]
  51. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. PCA-based feature reduction for hyperspectral remote sensing image classification. IETE Tech. Rev. 2021, 38, 377–396. [Google Scholar] [CrossRef]
Figure 1. (a) Sample hyperspectral data cube. Spectra are plotted using their geographic position with the log10-reflectance colored along the z axis and a pseudocolor image on top. (b) Points taken from a sample hyperspectral data cube corresponding to algae, rhodamine dye, water, and ground (dirt and dry grass). (c) Reflectance spectra for the exemplar points scaled so the peak value of each spectrum is 1.0 . (d) The location of the pond in Montague, Texas, where data were collected for this study.
Figure 1. (a) Sample hyperspectral data cube. Spectra are plotted using their geographic position with the log10-reflectance colored along the z axis and a pseudocolor image on top. (b) Points taken from a sample hyperspectral data cube corresponding to algae, rhodamine dye, water, and ground (dirt and dry grass). (c) Reflectance spectra for the exemplar points scaled so the peak value of each spectrum is 1.0 . (d) The location of the pond in Montague, Texas, where data were collected for this study.
Remotesensing 16 02430 g001
Figure 2. Illustration of the GTM algorithm. On the left is a regular grid of K many points (green) in the latent space represented by their coordinates ξ 1 and ξ 2 . In red are the M-many radial basis functions which define the mapping ψ from the latent space to the data space. Points in the latent space are mapped nonlinearly to the data space yielding an embedded manifold in R d , here shown in three dimensions.
Figure 2. Illustration of the GTM algorithm. On the left is a regular grid of K many points (green) in the latent space represented by their coordinates ξ 1 and ξ 2 . In red are the M-many radial basis functions which define the mapping ψ from the latent space to the data space. Points in the latent space are mapped nonlinearly to the data space yielding an embedded manifold in R d , here shown in three dimensions.
Remotesensing 16 02430 g002
Figure 3. The UAV platform: (a) The Resonon Pika XC2 hyperspectral imager. (b) The Freefly Alta-X with the attached hyperspectral imager, processing computer, and downwelling irradiance spectrometer.
Figure 3. The UAV platform: (a) The Resonon Pika XC2 hyperspectral imager. (b) The Freefly Alta-X with the attached hyperspectral imager, processing computer, and downwelling irradiance spectrometer.
Remotesensing 16 02430 g003
Figure 4. Visualization of the GTM trained solely on water pixels (no land and no rhodamine plume). Each data point is colored by computing the mean latent space position, ξ ^ n according to Equation (8) with red corresponding to the ξ 1 coordinate and blue corresponding to the ξ 2 coordinate. (a) Spatial distribution learned by the GTM. (b) The distribution of data in the latent space. Nodes corresponding to the four corners and center of the latent space are identified with cyan markers. (c) reflectance spectra corresponding to the selected GTM nodes computed via the nonlinear mapping ψ ( ξ k ).
Figure 4. Visualization of the GTM trained solely on water pixels (no land and no rhodamine plume). Each data point is colored by computing the mean latent space position, ξ ^ n according to Equation (8) with red corresponding to the ξ 1 coordinate and blue corresponding to the ξ 2 coordinate. (a) Spatial distribution learned by the GTM. (b) The distribution of data in the latent space. Nodes corresponding to the four corners and center of the latent space are identified with cyan markers. (c) reflectance spectra corresponding to the selected GTM nodes computed via the nonlinear mapping ψ ( ξ k ).
Remotesensing 16 02430 g004
Figure 5. GTM latent space visualization: Each sample spectrum from the dataset of combined HSI covering shore, water, and the rhodamine dye release are plotted in the GTM latent space at the location of the mean node responsibility, ξ ^ n . Points are colored according to the NDVI computed from the original reflectance spectra. The locations of exemplar spectra for algae, rhodamine, water, and ground points in the latent space are included as color-filled circles. Spectra from the shore and open water are clearly separated into different regions of the latent space.
Figure 5. GTM latent space visualization: Each sample spectrum from the dataset of combined HSI covering shore, water, and the rhodamine dye release are plotted in the GTM latent space at the location of the mean node responsibility, ξ ^ n . Points are colored according to the NDVI computed from the original reflectance spectra. The locations of exemplar spectra for algae, rhodamine, water, and ground points in the latent space are included as color-filled circles. Spectra from the shore and open water are clearly separated into different regions of the latent space.
Remotesensing 16 02430 g005
Figure 6. Results of the hyperparameter search: (left) Variation in BIC with m and s for fixed α = 0.1 . (right) Variation in BIC with with m and α for fixed s = 1.0 . The white star in each plot indicates the parameters with the lowest BIC across the entire parameter search.
Figure 6. Results of the hyperparameter search: (left) Variation in BIC with m and s for fixed α = 0.1 . (right) Variation in BIC with with m and α for fixed s = 1.0 . The white star in each plot indicates the parameters with the lowest BIC across the entire parameter search.
Remotesensing 16 02430 g006
Figure 7. Spectral signatures ψ ( ξ k ) corresponding to GTM nodes with nonzero responsibility for the rhodamine dye plume (top left), nearshore algae (top right), ground (bottom left), and open water (bottom right). A pseudocolor image is inset into each plot with the location of the exemplar spectrum marked with a white circle.
Figure 7. Spectral signatures ψ ( ξ k ) corresponding to GTM nodes with nonzero responsibility for the rhodamine dye plume (top left), nearshore algae (top right), ground (bottom left), and open water (bottom right). A pseudocolor image is inset into each plot with the location of the exemplar spectrum marked with a white circle.
Remotesensing 16 02430 g007aRemotesensing 16 02430 g007b
Figure 8. Using the spectral endmember assigned to algae from the trained GTM together with the NS3, we are able to estimate the algal abundance near the shore. (Left) NS3 values of pixels below a threshold of 0.4275 corresponding to a total area of 587.3 m2. (Right) A picture of the pond near the shore showing the algae.
Figure 8. Using the spectral endmember assigned to algae from the trained GTM together with the NS3, we are able to estimate the algal abundance near the shore. (Left) NS3 values of pixels below a threshold of 0.4275 corresponding to a total area of 587.3 m2. (Right) A picture of the pond near the shore showing the algae.
Remotesensing 16 02430 g008
Figure 9. Using the spectral endmember assigned to Rhodamine from the trained GTM together with the NS3, we are able to map the evolution of the dye plume across two UAV flights. (top) The initial dye plume corresponding to a total area of 330.7 m2. (bottom) The same dye plume imaged approximately 15 min later extends to a total area of 1495.6 m2.
Figure 9. Using the spectral endmember assigned to Rhodamine from the trained GTM together with the NS3, we are able to map the evolution of the dye plume across two UAV flights. (top) The initial dye plume corresponding to a total area of 330.7 m2. (bottom) The same dye plume imaged approximately 15 min later extends to a total area of 1495.6 m2.
Remotesensing 16 02430 g009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Waczak, J.; Aker, A.; Wijeratne, L.O.H.; Talebi, S.; Fernando, A.; Dewage, P.M.H.; Iqbal, M.; Lary, M.; Schaefer, D.; Balagopal, G.; et al. Unsupervised Characterization of Water Composition with UAV-Based Hyperspectral Imaging and Generative Topographic Mapping. Remote Sens. 2024, 16, 2430. https://doi.org/10.3390/rs16132430

AMA Style

Waczak J, Aker A, Wijeratne LOH, Talebi S, Fernando A, Dewage PMH, Iqbal M, Lary M, Schaefer D, Balagopal G, et al. Unsupervised Characterization of Water Composition with UAV-Based Hyperspectral Imaging and Generative Topographic Mapping. Remote Sensing. 2024; 16(13):2430. https://doi.org/10.3390/rs16132430

Chicago/Turabian Style

Waczak, John, Adam Aker, Lakitha O. H. Wijeratne, Shawhin Talebi, Ashen Fernando, Prabuddha M. H. Dewage, Mazhar Iqbal, Matthew Lary, David Schaefer, Gokul Balagopal, and et al. 2024. "Unsupervised Characterization of Water Composition with UAV-Based Hyperspectral Imaging and Generative Topographic Mapping" Remote Sensing 16, no. 13: 2430. https://doi.org/10.3390/rs16132430

APA Style

Waczak, J., Aker, A., Wijeratne, L. O. H., Talebi, S., Fernando, A., Dewage, P. M. H., Iqbal, M., Lary, M., Schaefer, D., Balagopal, G., & Lary, D. J. (2024). Unsupervised Characterization of Water Composition with UAV-Based Hyperspectral Imaging and Generative Topographic Mapping. Remote Sensing, 16(13), 2430. https://doi.org/10.3390/rs16132430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop