Next Article in Journal
GENICS: A Framework for Generating Attack Scenarios for Cybersecurity Exercises on Industrial Control Systems
Next Article in Special Issue
Old Landslide Detection Using Optical Remote Sensing Images Based on Improved YOLOv8
Previous Article in Journal
Integrating Root Morphology Based on Whole-Pullout Test of Model Roots: A Case Study
Previous Article in Special Issue
Point Cloud Deep Learning Network Based on Local Domain Multi-Level Feature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images

Department of Computer Systems Technology, North Carolina A&T State University, Greensboro, NC 27410, USA
Appl. Sci. 2024, 14(2), 767; https://doi.org/10.3390/app14020767
Submission received: 26 October 2023 / Revised: 3 January 2024 / Accepted: 10 January 2024 / Published: 16 January 2024
(This article belongs to the Special Issue Novel Approaches for Remote Sensing Image Processing)

Abstract

:
The spectrums of one type of object under different conditions have the same features (up, down, protruding, concave) at the same spectral positions, which can be used as primary parameters to evaluate the difference among remotely sensed pixels. The wavelet-feature correlation ratio Markov clustering algorithm (WFCRMCA) for remotely sensed data is proposed based on an accurate description of abrupt spectral features and an optimized Markov clustering in the wavelet feather space. The peak points can be captured and identified by applying a wavelet transform to spectral data. The correlation ratio between two samples is a statistical calculation of the matched peak point positions on the wavelet feature within an adjustable spectrum domain or a range of wavelet scales. The evenly sampled data can be used to create class centers, depending on the correlation ratio threshold at each Markov step, accelerating the clustering speed by avoiding the computation of Euclidean distance for traditional clustering algorithms, such as K-means and ISODATA. Markov clustering applies several strategies, such as a simulated annealing method and gradually shrinking the clustering size, to control the clustering convergence. It can quickly obtain the best class centers at each clustering temperature. The experimental results of the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and Thermal Mapping (TM) data have verified its acceptable clustering accuracy and high convergence velocity.

1. Introduction

Identifying suspected targets from remotely sensed data is paramount in everyday life and research. Researchers have extensively investigated numerous clustering algorithms, including cutting-edge technologies, for remotely sensed images, such as Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and Thermal Mapping (TM) data. However, several limitations exist in these algorithms. One widely used clustering algorithm is K-means clustering, which, unfortunately, cannot automatically determine the number of classes [1,2]. Moreover, it exhibits slow convergence due to its reliance on minimal spatial distance [3,4].
The approaches ISODATA [5,6] and ISMC [7,8] can determine the class number through self-iteration. Nevertheless, the challenge lies in determining their parameters, particularly in adjusting distance parameters with changing dimensions. On the other hand, orthogonal projection classification suffers from projection fluctuation issues under the restriction of the number of bands [9,10]. Cui introduced a feature extraction method that computes vectorized pixel values from a localized window, enhancing Bag-of-Words (BoW) performance. However, this approach may lead to a reduction in classification accuracy [11,12]. Peng et al. proposed a graph-based structural deep spectral–spatial clustering network to sufficiently explore the structure information among pixels. They designed a self-expression-embedded multi-graph auto-encoder to explore high-order structure associations among pixels, thereby capturing robust spectral–spatial features and global clustering structure [13].
Furthermore, Firat et al. developed a hybrid 3D residual spatial–spectral convolution network to extract deep spatio-spectral features using 3D CNN and ResNet architecture [14]. Acharyya combined wavelet theory and neuro-fuzzy techniques for segmentation purposes [15,16]. However, their feature extraction approach solely considers the absolute values of wavelet coefficients, neglecting the specific spectral patterns, and the computational requirements take time and effort.
A wavelet-feature correlation ratio Markov clustering algorithm (WFCRMCA) is proposed to differentiate the pixels according to the spectrum similarity among pixels. Of course, the spectrums of one object under different conditions are different. Still, they have the same features (up, down, protruding, concave, see Figure 1) at the same spectral positions, which are the main parameters used to evaluate the difference among remotely sensed pixels [17,18]. Therefore, these characteristic positions can denote class features. Fortunately, band-pass wavelet filters can decompose data at different scales to detect these characteristics.
WFCRMCA can statistically control clustering accuracy by adjusting parameters such as Tstart, Tend, and Tstep. A new concept, correlation ratio (CR), is proposed to reflect the similarity between two wavelet-transformed samples. With an accurate description of the abrupt spectral features, wavelet correlation ratios can differentiate pixels along spectral dimensions. Expanding spectral bands of multi-spectral images increases the number of characteristic points to enrich the features of classes. WFCRMCA forms the clustering space and initial class centers with evenly sampled pixels. Without the initial parameter problem of the K-means algorithm, WFCRMCA can quickly reach the best class centers at each clustering temperature and obtain optimal class centers on the whole scope at high speed by gradually decreasing the clustering scale and temperature. Several theorems are provided and proved to strengthen the WFCRMCA in the Section 2. In the Section 3, WFCRMCA receives favorable results for clustering Landsat TM images and AVIRIS hyperspectral images.

2. Methods

Although the spectral curves of the same objects under different conditions are somewhat different, they have the same feature points (upward, downward, maximum, and minimum) at the same spectral positions (Figure 1). The WFCRMCA could detect abrupt signals through band-pass wavelet transform, such as crossing zero and extreme points. But crossing the zero point cannot be ensured to be a pulse signal, and perhaps is a smoothly changed signal, so the extreme points between adjacent zero points are much more critical. The signs in spectral vector format are classified according to the priority of importance from low to high: downward, upward, protruding, and concave (Figure 2).
Figure 2 is the result of four kinds of abrupt signals processed by ψ(t) (Equation (1), [19], 1st derivative Gauss function θ(t)). For some remotely sensed images affected by too many mixed pixels, the position of critical points will probably deviate or have many little fluctuations, so WFCRMCA could eliminate unimportant signals by setting a maximum threshold and only clustering the partial minutia at a high-level scale.
θ t = 1 2 π e t 2 / 2 , ψ t = d θ d t = 1 2 π t e t 2 / 2
Wavelet feature clustering algorithms only analyze minutia data by detecting and determining the positions of abrupt signals. Using a fast binary Mallet wavelet algorithm [20] in Equation (2) to extract wavelet coefficients, WFCRMCA can mark the upward-maximal points (Figure 2a and Figure 2a’) and downward-minimal points (Figure 2b and Figure 2b’) along the spectrum. WFCRMCA will overlook the weak signals if Tpeak is large enough, leading to a failure in identifying some valuable signs among hidden objects.
c j , k = n h n 2 k c j 1 , n d j , k = n g n 2 k c j 1 , n , j = 1 , 2 , , S c a l e l o g 2 b
The WFCRMCA uses rij (correlation ratio), which works like a distance but not Euclidean distance as clustering criteria, to evaluate the difference between two spectral vectors on partial minutia. Equation (3) uses Scale2-scale minutia of Scale-scale wavelet coefficients to cluster, ti,k is the kth feature of ith sampled vector, N(·) is the number of feature positions that match criteria. WFCRMCA could use binary values to mark whether the position is valuable enough to attend clustering. When S c a l e 2 = S c a l e = l o g 2 b , the bit number attending clustering comparison is b i = 1 l o g 2 b 1 / 2 i b 1 .
r i j = N t i , k = t j , k , k Ω N t i , k t j , k , k Ω + N t i , k = t j , k , k Ω i f   S c a l e 2 = S c a l e , Ω = 0 , b k = 1 S c a l e 1 2 k 1 i f   S c a l e 2 < S c a l e , Ω = b k = 1 S c a l e S c a l e 2 1 2 k , b k = 1 S c a l e 1 2 k 1

2.1. Expanding Bands Method for Multi-Spectral Images

As the band number of multi-spectral images (TM images have only seven bands) is not high enough for the wavelet transform to extract efficient feature points, WFCRMCA expurgates the bands with great noise and expands the rest with 2nd order and nonlinear correlated functions so that WFCRMCA can detect more wavelet features. The expanding multi-spectral bands’ method [9] is listed as follows.
1.
Second-order correlated bands include the auto-correlated bands ( B i 2 i = 1 b ) and the cross-correlated bands ( B i B j i , j = 1 ,   i j b ).
2.
Nonlinear correlated bands include the bands stretched out by the square root ( B i i = 1 b ) and those stretched out by the logarithmic function ( l o g B i i = 1 b ).
The bands created by (1) and (2), together with the 1st order bands, which are original ( B i i = 1 b ), assemble new remotely sensed data with (b2 + 7b)/2 bands.

2.2. Markov Chain Clustering in Wavelet Feature Space

The wavelet-feature Markov clustering algorithm, i.e., WFCRMCA, first denoises the original data to make the spectral features more accurate, then uses a band-pass wavelet filter to detect all dot vectors for sharp points, including upward-maximal and downward-minimal points. As a result, simulated annealing Markov chain decomposition in state space, formed by evenly spaced sampled data, could realize the best centers at each temperature and sub-finest centers on the whole scope.
According to the peculiarity of simulated annealing Markov clustering, each clustering center is one state, and the space is a definite Markov state chain. If two classes (or states) merge, according to CR, it has nothing to do with other states. For example, for Markov chain I = 1,2 , , n in definite state space X n , if any two states communicate, they must be in the same class. Thus, the whole state space (pixels) could be separated into a few isolated classes according to transferred communication. T, which is defined as a threshold value of CR rij, is used as an annealing temperature to control the clustering process.
Definition 1.
If  P i j 1 = r i j T > 0  for states i and j, they have one-step transferred communication denoted as  i 1 j .
Theorem 1.
Communication can be transferred. If  i m k  and  k n j  ( p i k m > 0 ,   p k j n > 0 ),  i m + n j p i j m + n > 0 .
Proof of Theorem 1.
According to Chapman–Kolmogorov equation:
p i j m + n = g ϵ I p i g m · p g j n p i k m · p k j n > 0 . i . e . ,   i m + n j .
 □
Definition 2.
If feature i in wavelet characteristic space has pij = 1, then i is an absorptive state, forming a single-dot set {i}.
Theorem 2.
After Markov clustering, all wavelet features in wavelet feature space are frequently returned states.
Proof of Theorem 2.
(i). A single-dot set is an absorptive, frequently returned state. (ii). As simulated annealing clustering causes T to be reduced gradually, the k+1th clustering iteration is supposed to create a non-single dot set. For instance, m pixels {1, 2,…, m} are absorbed into one class. Tk is the CR threshold of the kth iteration, and c1,…, i, … j, …, and cn are the created clustering centers of the kth iteration. Thus, r i j < T k .
During the k+1th iteration, T k + 1 = T k T s t e p , where Tstep is the depressed step of T at each iteration. If T k + 1 < r i j < T k , then p i j = r i j T k + 1 = r i j + T s t e p T k > 0 and p i j < T k T k + 1 = T s t e p , so i 1 j , then i and j are merged together.
If Tstep is small enough (i.e., the temperature is reduced slowly), and i, j, l are absorbed in k+1th iteration, p i j = p i l = p j l T s t e p so that it could be supposed as in Figure 3 that p i i = x ,   p i j = 1 x / m 1 , i j , i , j 1,2 , , m 1 .
As m states communicate with each other, other m − 1 states could be seen as one state j. Let pii = pjj = x, pij = 1 – x; thus
f i i = x + 1 x 2 + 1 x 2 x + + 1 x 2 x n + = x + 1 x 2 1 + x + x 2 + = 1
So state i is a frequently returned state. As m states communicate, the merged m states are frequently returned. □
Theorem 3.
The sufficient and necessary condition of closed set C is that, for arbitrary elements  i C , j C , there exists  p i j n = 0 , n 1  (referring [21,22]).
Theorem 4.
Definite states of Markov chain in wavelet feature space can be uniquely decomposed withoutoverlap into a definite number of frequently returned states, including closed sets C1, …, Cm  and single dot sets Cm+1, …, Cn, existing:
1. 
Any two states in Ch ( h ϵ 1 , n )  are communicated.
2. 
When  h g , h , g ϵ 1 , n , any state in Cg cannot communicate with any state in Ch (referring [21,22]).
Therefore, every state is frequently returned in the wavelet feature space at each temperature, and the number of isolated closed sets equals the number of classes. Then, the whole wavelet Markov chain feature state space has a decomposable expression that consists of several closed sets without overlap.

2.3. Adjustment of Clustering Centers

When two classes are merged whose correlation ratio rij is bigger than T, the numbers of each feature (including crossing zero part) are separately added up at the corresponding position ([0, b − 1], the number of wavelet coefficients is approximately b). In addition, their sample numbers are also added up separately.
Similar to the traditional clustering method, reasonable adjustment of clustering centers is based on the statistic of intra-class features. For each position, the feature that occurs most frequently is chosen as the common feature of the new class, and then b common features will be created. If several features come up at the same frequency, the feature with the highest priority (for example, downward-minimal or concave point >upward-maximal or protruding point) is chosen. Then, among all the class centers merged into one new class at this iteration, one pixel with the biggest CR with common features as a new class center is chosen. According to Equations (4) and (5), N B l , k c i is the statistic number of feature k on the lth position in class i, and t l is the feature of the lth position, which can be downward (0), upward (1), protruding (2), and concave (3). R(c1,c2) is the correlation ratio between vector c1 and c2, Z c i is the set of class centers absorbed by class i, and M c i is the common features of class i.
M c i = t 0 t 1 t b 1 | N B l , k c i > N B l , j c i , l 0 , b 1 , j k , j 0,3 t l = k , c i C
x Z c i , R x , M c i > R y , M c i , y x , y Z c i c i = x
During the clustering process, many pixels with high similarity are merged, causing the number of class centers that will attend the following iterative clustering comparison to decrease sharply. As only newly created centers follow next-cycle clustering, WFCRMCA has a high clustering speed. The computational complexity of WFCRMCA is O(n), which is impressive in handling large-scale datasets.

2.4. Wavelet-Feature Markov Clustering Algorithm

Based on the preceding theoretical analysis, the WFCRMCA uses a simulated annealing technique to gradually decrease CR threshold T through Markov chain decomposition in wavelet feature space at each temperature, obtaining the best clustering centers of the whole space. Supposed that ci is the class center of class i, Sci is the pixel set of class i, C is the set of all classes, Zci is the set of class centers absorbed by class i at the current temperature, Nc is the class number, Ns is the number of sampled pixels (initial class centers are sampled pixels s i , i 0 ,   N s 1 ), R(c1,c2) is the CR between c1 and c2, N Z c i is the number of class centers absorbed by class i, N S c i is the pixel number in class i, Tstart is the initial value of CR T, and Tend is the lowest CR threshold. The detailed process of WFCRMCA is provided in the flow chart in Figure 4. The simulated annealing Markov chain decomposition clustering in wavelet feature space is listed as follows:
  • Input parameters:
Stepx and Stepy are the sampling distances along horizontal or vertical directions;
b, m, n are, separately, the band number, column number, and row number of original remotely sensed images;
Scale is the wavelet transform scale;
Scale2 is the number of minutia scale attending clustering (i.e., Scale Scale2~Scale minutia sections)
2.
Data preprocessing: delete bands primarily affected by noise and atmosphere, such as the 1–6th, 33rd, 107–114th,153–168th, and 222–224th bands of AVIRIS. Multi-spectral images need to expand bands.
3.
Apply band-pass Scale-scale wavelet filter (for example, Equation (6) [23,24]) to all pixels, search extreme points above noise threshold Tpeak between neighbor crossing zero points on each minutia section, and mark upward-maximal point as one and downward-minimal point as two at the corresponding position.
H = 0.0 , 0.125 , 0.375 , 0.375 , 0.125 , 0.000 , G = 0.0061 , 0.0869 , 0.5798 , 0.5798 , 0.0869 , 0.0061
4.
According to Stepx ×  Stepy sampling distance, sample the pixels and create Ns sampled pixels evenly.
5.
Apply simulated annealing Markov state decomposition clustering to Scale-Scale2~Scale scale minutia sections of sampled data.
(a)
Set initial temperature T as Tstart, the clustering signal standard Tsignal (ratio of intra-class sampled pixel number over the number of total sampled pixels) is 1.0, and each pixel is one class center (beginning with Ns class centers). In the end, according to step b-e, apply Markov chain decomposition in state space to the wavelet features of the sampled pixels by gradually depressing the signal size.
N c = N s , C = c 0 , c 1 , , c N c 1 , S c i = s i , c i C , i 0 , N c 1
(b)
Make judgments to all present class centers. If class i is a significant signal in which the number of pixels is more prominent than T s i g n a l N s , move to the next class. Otherwise, search forward one by one for another class j whose size is smaller than T s i g n a l N s , and make clustering judgments between class j and i.
(c)
According to Equation (8), if the CR between the centers of two classes (i and j) meets the condition Pij = rijT > 0, then class j is absorbed into class i. Continue this process (b) until the last class is detected.
Z c i = c i , c i C , c i , c i C , i f   N S c i < T s i g n a l N s , N S c j < T s i g n a l N s , R c i , c j > T N B l , k c i = N B l , k c i + N B l , k c j l 0 , b 1 , k 0,2 , Z c i = Z c i c j , S c i = S c i S c j , C = C c j , N c = N c 1
(d)
According to Equations (4) and (5), re-adjust the newly created centers: among all the class centers merged into one new class at this iteration, choose one pixel with the biggest CR with common features as a new class center.
(e)
Let T = TTstep decrease clustering temperature, and Tsignal = Tsignal/2 reduce clustering size. Repeat steps (a)–(d) until T is reduced to the appointed small signal threshold Tend or the set class number is reached.
6.
According to the clustering centers created by (5), each pixel is clustered into one class whose center has the maximal CR.
Figure 4. Flow chart of WFCRMCA.
Figure 4. Flow chart of WFCRMCA.
Applsci 14 00767 g004

3. Results

The WFCRMCA uses Microsoft Visual C++ language (Microsoft, Redmond, DC, USA) and basic libraries for the code of the proposed algorithm. The TM and AVIRIS data analysis demonstrate the merits and defects of the wavelet feature clustering algorithms, say, WFCRMCA. Classified pixels are shown in white.

3.1. Multi-Spectral Data

For Mississippi (Figure 5a, 512  ×  512, 8 bit [25]) TM multi-spectral images, the sixth band heavily affected by the atmosphere is crossed out. The other 6 bands are expanded to 39: original data, 1–6; second-order auto-correlated bands, 7–12; second-order cross-correlated bands, 13–27; square root function, 28–33; and logarithmic function, 34–39.
It is supposed that Tstart = 0.95 during the discussion of the parameters’ influence on Mississippi’s clustering results. If only the original six bands are processed by two-scale wavelet decomposition, only four classes are created because the features need to be stronger. As the first iteration absorbs too many classes, the intra-class adjustment costs most of the time. The experiment also shows that second-order correlation expanded bands (7th–27th) provide more class information, but nonlinear correlation developed bands (28th–39th) make the clustering results stable.
The expanding spectrum method increases data processing complexity; however, if there are only several classes, the clustering speed is low because the big class has to spend more time calculating the center. Therefore, this method will maintain the clustering at a stable speed for multi-spectral data. Table 1 shows that this method could identify the potential specific classes, leading to higher clustering accuracy.
Figure 6 is the clustering result of the parameters in Table 2. It can be seen that class 1 is plow land or meadow (Figure 6a), class 2 is beach (Figure 6b), class 3 is river channel (Figure 6c), class 6 is dyke (Figure 6f), and class 9 is the slope on the bank (Figure 6i). The clustering results maintain significant signals and efficiently embody the minor signs. If the data are divided into 18 classes by the K-means algorithm, one iteration, on average, uses 60 s, so this clustering method, according to the features on the spectral curves of remotely sensed objects, is more flexible on parameter choice and has a quicker clustering speed than the standard clustering algorithm (such as K-means).

3.2. Hyper-Spectral Data

For 224 bands Sook Lake (Figure 5b, 256  ×  256, 16 bit [26,27]) AVIRIS hyperspectral images, the WFCRMCA crosses out heavily disturbed bands, such as 1–6th, 33rd, 107–114th, 153–168th, 222–224th bands, and uses the remaining 190 bands for algorithm analysis.
In Table 3, when Tpeak = 0, the number of classes is 133. If Tpeak is increased, the number of classes nonlinearly reduces, and clustering time depresses accordingly. When Tpeak > 7, the number of classes begins to fluctuate, so the WFCRMCA usually chooses Tpeak = 5.0, which could realize a fairly accurate classification.
In Table 4, with more minutiae attending clustering, the number of clustering classes increases sharply: two of five scale components cluster eight categories; obviously, that does not separate the objects. However, four-scale components cause objects to disperse and expand the class number. Depressing Tend could effectively decrease the class number.
If high three-scale or five-scale wavelet decompositions are chosen to attend clustering and Tpeak = 5.0, 17 classes are created (the main clustering results are seen in Figure 7), the time is 21s, and the division result is favorable; here, class 1 is the basin (Figure 7a), class 4 is for the mountain peaks (Figure 7d), and class 5 is the water body of the Sook Lake (Figure 7e).

4. Discussion

For remotely sensed data with a high density of mixed pixels, choosing partial minutia wavelet features in high-level scales could reduce the clustering difficulty caused by a significant amount of minutia, and this also somewhat applies blur to achieve ideal clustering results. Multi-scale classification from fine to coarse could be realized by this method. Furthermore, as the matching speed of abrupt point positions is very high, clustering time increases a little with the increment of referenced minutia. For example, multi-spectral data typically set Scale = Scale2; hyper-spectral data could set Scale2 = Scale − 2.
WFCRMCA applies 1D wavelet transformation on satellite spectral data. Wavelet transform can represent a signal in both time and frequency domains simultaneously. It decomposes a signal into a set of wavelets that are localized in both time and frequency, allowing me to analyze the signal’s time-localized features. Wavelet transform excels at capturing localized features and adaptability to non-stationary signals. However, the Fourier transform represents a signal in the frequency domain. It decomposes a signal into a sum of sinusoidal components of different frequencies, providing information about the signal’s frequency content. It does not capture information about when these frequencies occur. Fourier transform is excellent for spectral analysis.
Ridgelet and curvelet transformation are well-known methods for high-dimensional image analysis, but wavelet transformation is better for 1D spectral feature extraction. In the ridgelet transform, ridgelets are adapted to higher-dimensional singularities, or singularities on curves in dimension two, singularities on surfaces in dimension three, and singularities on (n − 1) dimensional hypersurfaces in dimension n [28]. The curvelet transform uses ridgelet transform as a component step, and it is good at 2D image reconstruction [29]. The proposed WFCRMCA uses wavelet transform to analyze the 1D spectral data instead of 2D images.
WFCRMCA accelerates clustering speed during the clustering process. The calculation of CR only needs simply matching corresponding characteristic points without the time-consuming floating-point measure of Euclidean distance [1,2]. A great many sampled pixels with high similarity are clustered together during the clustering process, depressing the number of class centers attending clustering comparison; moreover, clustering centers of newly created classes are re-determined according to common features. So, along with the process of this algorithm, the clustering speed continues to increase.
This WFCRMCA only makes statistics of the number of each wavelet feature on every info-position as the class feature and chooses the best pixel as the clustering center but does not directly use the CR matrix to investigate the dependency degree between sampled pixels, resolving spatial complexity problems.
Gradually depressing clustering size could let both small and large signals embody efficiently, and too many noise signals are merged so that the WFCRMCA could detect the spatial position of noise signals.
The WFCRMCA approach can be applied to any spectral data to differentiate targets. It has demonstrated favorable performance for satellite multi-spectral images and super-spectral images. The spectral analysis method has potential applications in other spectrum data from photoacoustic imaging, OCT, etc. Spectrum analysis of the spectrum data from photoacoustic imaging and OCT has no successful cases. However, the proposed method provides a possibility of enhancing the classification accuracy through spectrum analysis.
Though the Markov clustering method is not parallelizable in choosing clustering centers, it can provide optimal clustering centers of the wavelet coefficients at a high convergent velocity. After clustering centers are determined, WFCRMCA can cluster the larger dataset in parallel.
In the WFCRMCA, even though most parameters are stable and can be used in most cases, several parameters, such as Tpeak and Scale2, still need to be adjusted manually to increase clustering accuracy for specific applications. It is easy for a user to try different values to derive the best parameters for their application. Future work will continue to focus on optimizing the parameters.
The proposed WFCRMCA method analyzes the spectra of one type of object under different conditions with the same features (up, down, protruding, concave) so that it can differentiate the targets. This method has the potential to integrate with other equally essential features to increase the accuracy of classification. Incorporating multiple features in clustering provides more possibilities and challenging topics for future research.

5. Conclusions

The wavelet-feature correlation ratio is used to depict the distance between two pixels by analyzing the wavelet features of spectral curves for remotely sensed data. Based on the particularity of CR clustering, a wavelet-feature Markov clustering algorithm is proposed for searching the optimal class centers. After spatial data are evenly sampled, sharp points on the band-pass wavelet coefficients, including extreme points and crossing zero points, are captured and used for clustering matching. WFCRMCA accelerates clustering speed by avoiding the time-consuming Euclidean distance calculation used for general clustering algorithms. For multi-spectral data, nonlinear correlation expanded bands provide more class information than second-order correlation developed bands. Markov clustering based on simulated annealing realizes fast clustering convergence at each temperature. WFCRMCA can enhance classification accuracy through spectrum analysis for other applications with spectrum data.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data are public data and are referenced in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Z. Residual Clustering Based Lossless Compression for Remotely Sensed Images. In Proceedings of the 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 6–8 December 2018; pp. 536–539. [Google Scholar] [CrossRef]
  2. Wang, Z. Entropy Analysis for Clustering Based Lossless Compression of Remotely Sensed Images. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 4220–4223. [Google Scholar] [CrossRef]
  3. Theodoridis, S.; Koutroumba, K. Pattern Recognition, 4th ed.; Academic Press: Cambridge, MA, USA, 2008; pp. 741–745. [Google Scholar]
  4. Ikotun, A.M.; Ezugwu, A.E.; Abualigah, L.; Abuhaija, B.; Heming, J. K-means Clustering Algorithms: A Comprehensive Review, Variants Analysis, and Advances in the Era of Big Data. Inf. Sci. 2023, 622, 178–210. [Google Scholar] [CrossRef]
  5. Soto de la Cruz, R.; Castro-Espinoza, F.A.; Soto, L. Isodata-Based Method for Clustering Surveys Responses with Mixed Data: The 2021 StackOverflow Developer Survey. Comput. Sist. 2023, 27, 173–182. [Google Scholar] [CrossRef]
  6. Arai, K. Improved ISODATA Clustering Method with Parameter Estimation based on Genetic Algorithm. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 187–193. [Google Scholar] [CrossRef]
  7. Simpson, J.J.; McIntre, T.J.; Sienko, M. An Improved Hybrid Clustering Algorithm for Natural Scenes. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1016–1032. [Google Scholar] [CrossRef]
  8. Bo, L.; Bretschneider, T. D-ISMC: A distributed unsupervised classification algorithm for optical satellite imagery. In Proceedings of the 2003 IEEE International Geoscience and Remote Sensing Symposium, Toulouse, France, 21–25 July 2003; Volume 6, pp. 3413–3419. [Google Scholar]
  9. Ren, H.; Chang, C.-I. A Generalized Orthogonal Subspace Projection Approach to Unsupervised Multi-spectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2000, 38, 2515–2528. [Google Scholar]
  10. Ifarraguerri, A.; Chang, C.-I. Unsupervised Hyperspectral Image Analysis with Projection Pursuit. IEEE Trans. Geosci. Remote Sens. 2000, 38, 2529–2538. [Google Scholar]
  11. Cui, S.; Schwarz, G.; Datcu, M. Remote sensing image classification: No features, no clustering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 5158–5170. [Google Scholar] [CrossRef]
  12. Chen, X.; Zhu, G.; Liu, M. Bag-of-Visual-Words Scene Classifier for Remote Sensing Image Based on Region Covariance. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  13. Peng, B.; Yao, Y.; Lei, J.; Fang, L.; Huang, Q. Graph-Based Structural Deep Spectral-Spatial Clustering for Hyperspectral Image. IEEE Trans. Instrum. Meas. 2023, accepted. [Google Scholar] [CrossRef]
  14. Firat, H.; Asker, M.E.; Bayindir, M.I.; Hanbay, D. 3D residual spatial–spectral convolution network for hyperspectral remote sensing image classification. Neural Comput. Appl. 2023, 35, 4479–4497. [Google Scholar] [CrossRef]
  15. Acharyya, M.; De, R.K.; Kundu, M.K. Segmentation of remotely sensed images using wavelet features and their evaluation in soft computing framework. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2900–2905. [Google Scholar] [CrossRef]
  16. Anupong, W.; Jweeg, M.J.; Alani, S.; Al-Kharsan, I.H.; Alviz-Meza, A.; Cárdenas-Escrocia, Y. Comparison of Wavelet Artificial Neural Network, Wavelet Support Vector Machine, and Adaptive Neuro-Fuzzy Inference System Methods in Estimating Total Solar Radiation in Iraq. Energies 2023, 16, 985. [Google Scholar] [CrossRef]
  17. Wang, Z.; Zhou, P. Greedy clustering algorithm and its application for the classification and compression of remotely sensed images. J. Univ. Sci. Technol. China 2003, 33, 52–59. [Google Scholar]
  18. Wang, Z.; Zhou, P. Fast clustering based on spectral wavelet features extraction and simulated annealing algorithm for multi-spectral Images. J. Image Graph. 2002, 7A, 1257–1262. [Google Scholar]
  19. Haddad, S.A.P.; Serdijn, W.A. Ultra Low-Power Biomedical Signal Processing: An Analog Wavelet Filter Approach for Pacemakers; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2009; pp. 34–50. [Google Scholar]
  20. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  21. Joseph, A. Markov Chain Monte Carlo Methods in Quantum Field Theories: A Modern Primer; Springer Nature: Berlin/Heidelberg, Germany, 2020; pp. 29–35. [Google Scholar]
  22. Bittelli, M.; Olmi, R.; Rosa, R. Random Process Analysis with R; Oxford University Press: Oxford, UK, 2022; pp. 25–31. [Google Scholar]
  23. Li, Q.; Zhao, J.; Zhao, Y.-N. Detection of Ventricular Fibrillation by Support Vector Machine Algorithm. In Proceedings of the IEEE International Asia Conference on Informatics in Control, Automation and Robotics, Bangkok, Thailand, 1–2 February 2009; pp. 287–290. [Google Scholar]
  24. Swelends, W. The Lifting Scheme: A Custom-design Construction of Biorthogonal Wavelet. Appl. Comput. Harmon. Anal. 1996, 3, 186–220. [Google Scholar]
  25. Kulkarni, A.; McCaslin, S. Knowledge Discovery from Multi-spectral Satellite Images. IEEE Geosci. Remote Sens. Lett. 2004, 1, 246–250. [Google Scholar] [CrossRef]
  26. Goodenough, D.G.; Bhogal, A.S.; Dyk, A.; Niemann, O.; Han, T.; Chen, H.; West, C.; Schmidt, C. Calibration of Forest Chemistry for Hyperspectral Analysis. In Proceedings of the IEEE 2001 International Geoscience and Remote Sensing Symposium, Sydney, NSW, Australia, 9–13 July 2001; Volume 1, pp. 52–56. [Google Scholar]
  27. Goodenough, D.G.; Dyk, A.; Niemann, K.O.; Pearlman, J.S.; Chen, H.; Han, T.; Murdoch, M.; West, C. Processing Hyperion and ALI for Forest Classification. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1321–1331. [Google Scholar] [CrossRef]
  28. Candès, E.J.; Donoho, D.L. Ridgelets: A key to higher-dimensional intermittency? Philos. Trans. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 1999, 357, 2495–2509. [Google Scholar] [CrossRef]
  29. Starck, J.L.; Candès, E.J.; Donoho, D.L. The curvelet transform for image denoising. IEEE Trans. Image Process. 2002, 11, 670–684. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Five points at different spatial positions within the same class have the same features at the exact spectral locations.
Figure 1. Five points at different spatial positions within the same class have the same features at the exact spectral locations.
Applsci 14 00767 g001
Figure 2. The wavelet band-pass filter and four kinds of abrupt signals. (ad) are the four critical signals: upward-maximal point, downward-minimal point, protruding crossing zero point, and concave crossing zero point. ψ(t) is the band-pass wavelet filter, (a’d’) are the output of the four signals through the wavelet filter.
Figure 2. The wavelet band-pass filter and four kinds of abrupt signals. (ad) are the four critical signals: upward-maximal point, downward-minimal point, protruding crossing zero point, and concave crossing zero point. ψ(t) is the band-pass wavelet filter, (a’d’) are the output of the four signals through the wavelet filter.
Applsci 14 00767 g002
Figure 3. (a) Closed set composed of five states. (b) Closed set with two states.
Figure 3. (a) Closed set composed of five states. (b) Closed set with two states.
Applsci 14 00767 g003
Figure 5. (a) Mississippi TM 4th band image after gray balance. (b) Sook Lake AVIRIS 60th band image after gray balance.
Figure 5. (a) Mississippi TM 4th band image after gray balance. (b) Sook Lake AVIRIS 60th band image after gray balance.
Applsci 14 00767 g005
Figure 6. Mississippi TM image, WFCRMCA clustering results: (ai) are the eight significant signals.
Figure 6. Mississippi TM image, WFCRMCA clustering results: (ai) are the eight significant signals.
Applsci 14 00767 g006
Figure 7. Sook Lake AVIRIS image WFCRMCA clustering result. (af) are the seven significant signals.
Figure 7. Sook Lake AVIRIS image WFCRMCA clustering result. (af) are the seven significant signals.
Applsci 14 00767 g007
Table 1. Expanded bands number comparison (3 × 3 sampling, Scale = Scale2 = 5, Tsignal = 0.1, Tstep = 0.01).
Table 1. Expanded bands number comparison (3 × 3 sampling, Scale = Scale2 = 5, Tsignal = 0.1, Tstep = 0.01).
Band Number Tcr1Tcr2Class Number
6 (Scale = 2)0.80.84
390.70.736
Table 2. Mississippi TM clustering (sampling 5 × 5, Scale = 4, Tpeak = 5, Tstart = 0.9, Tstep = 0.05).
Table 2. Mississippi TM clustering (sampling 5 × 5, Scale = 4, Tpeak = 5, Tstart = 0.9, Tstep = 0.05).
Band No.Tcr1Tcr2Scale2Class No.
390.90.4418
Table 3. Sook lake AVIRIS hyper-spectral image, WFCRMCA clustering parameter Tpeak comparison (5 × 5 sampling, band number 190, Scale = 5, Scale2 = 4, Tstart = 0.9, Tstep = 0.05).
Table 3. Sook lake AVIRIS hyper-spectral image, WFCRMCA clustering parameter Tpeak comparison (5 × 5 sampling, band number 190, Scale = 5, Scale2 = 4, Tstart = 0.9, Tstep = 0.05).
TpeakTendClass Number
00.4133
20.465
50.438
70.426
100.422
150.424
Table 4. Sook lake AVIRIS hyper-spectral image WFCRMCA clustering parameters Tend and Scale2 comparison (5 × 5 sampling, band number is 190, Scale = 5, Tpeak = 5, Tstart = 0.9, Tstep = 0.05).
Table 4. Sook lake AVIRIS hyper-spectral image WFCRMCA clustering parameters Tend and Scale2 comparison (5 × 5 sampling, band number is 190, Scale = 5, Tpeak = 5, Tstart = 0.9, Tstep = 0.05).
TendScale2Class NumberTime/s
0.42813
0.431721
0.443851
0.648552
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z. Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images. Appl. Sci. 2024, 14, 767. https://doi.org/10.3390/app14020767

AMA Style

Wang Z. Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images. Applied Sciences. 2024; 14(2):767. https://doi.org/10.3390/app14020767

Chicago/Turabian Style

Wang, Zhaohui. 2024. "Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images" Applied Sciences 14, no. 2: 767. https://doi.org/10.3390/app14020767

APA Style

Wang, Z. (2024). Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images. Applied Sciences, 14(2), 767. https://doi.org/10.3390/app14020767

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop