Next Article in Journal
Evaluating Visible–Infrared Imaging Radiometer Suite Imagery for Developing Near-Real-Time Nationwide Vegetation Cover Monitoring in Indonesia
Previous Article in Journal
ESatSR: Enhancing Super-Resolution for Satellite Remote Sensing Images with State Space Model and Spatial Context
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Multi-Order Spatial–Spectral Residual Feature Extractor for Weak Information Mining in Remote Sensing Imagery

1
Key Laboratory of 3-Dimensional Information Acquisition and Application, Ministry of Education, Capital Normal University, Beijing 100048, China
2
Engineering Research Center of Spatial Information Technology, Ministry of Education, Capital Normal University, Beijing 100048, China
3
State Key Laboratory of Efficient Utilization of Arid and Semi-Arid Arable Land in Northern China, National Hulunber Grassland Ecosystem Observation and Research Station, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100081, China
4
Aerosphere Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
5
School of Ecology, Resources and Environment, Dezhou University, Dezhou 253023, China
6
Department of Soil, Plant and Food Sciences, University of Bari Aldo Moro, 70126 Bari, Italy
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(11), 1957; https://doi.org/10.3390/rs16111957
Submission received: 11 April 2024 / Revised: 25 May 2024 / Accepted: 26 May 2024 / Published: 29 May 2024

Abstract

:
Remote sensing images (RSIs) are widely used in various fields due to their versatility, accuracy, and capacity for earth observation. Direct application of RSIs to harvest optimal results is generally difficult, especially for weak information features in the images. Thus, extracting the weak information in RSIs is reasonable to promote further applications. However, the current techniques for weak information extraction mainly focus on spectral features in hyperspectral images (HSIs), and a universal weak information extraction technology for RSI is lacking. Therefore, this study focused on mining the weak information from RSIs and proposed the deep multi-order spatial–spectral residual feature extractor (DMSRE). The DMSRE considers the global information and three-dimensional cube structures by combining low-rank representation, high-order residual quantization, and multi-granularity spectral segmentation theories. This extractor obtains spatial–spectral features from two derived sequences (deep spatial–spectral residual feature (DMSR) and deep spatial–spectral coding feature (DMSC)), and three RSI datasets (i.e., Chikusei, ZY1-02D, and Pasture datasets) were employed to validate the DMSRE method. Comparative results of the weak information extraction-based classifications (including DMSR and DMSC) and the raw image-based classifications showed the following: (i) the DMSRs can improve the classification accuracy of individual classes in fine classification applications (e.g., Asphalt class in the Chikusei dataset, from 89.12% to 95.99%); (ii) the DMSC improved the overall accuracy in rough classification applications (from 92.07% to 92.78%); and (iii) the DMSC improved the overall accuracy in RGB classification applications (from 63.25% to 63.6%), whereas DMSR improved the classification accuracy of individual classes on the RGB image (e.g., Plantain classes in the Pasture dataset, from 32.49% to 39.86%). This study demonstrates the practicality and capability of the DMSRE method to promote target recognition on RSIs and presents an alternative technique for weak information mining on RSIs, indicating the potential to extend weak information-based applications of RSIs.

1. Introduction

Remote sensing images (RSIs) can capture details of ground objects [1]; thus, they are widely used in agriculture, military, resource exploration, and change detection [2,3,4,5]. Information extraction is one of the most important step in RSI processing and plays a significant role in the application of RSIs [6]; however, information extraction from RSIs still remains challenging [7], especially in weak information. Weak information features in RSIs generally refer to information that is less apparent or easily overshadowed by surrounding features [8], and weak information is hard extract after preprocessing [9], e.g., soil salinity [10], heavy metals [11]. Achieving optimal results for weak information features can be challenging. Therefore, developing an approach for efficient mining of weak information in RSIs is necessary to facilitate and promote their application.
The feature extraction (FE) technique is an effective way to address the challenging task of information extraction in RSIs [12]. This technique helps extract suitable feature information from the raw RSIs for remote sensing [13]. The most common FE techniques for RSIs can be divided into three major categories: spectral-, spatial-, and spatial–spectral-based FE methods. Classical spectral-based FE methods for RSIs, such as principal component analysis (PCA) [14] and linear discriminant analysis (LDA) [15], are widely applied. PCA identifies the projection with the minimum data reconstruction error, while LDA aims to retain the information necessary to separate the target classes [16]. However, the above methods are unsuitable for modeling the nonlinear relationships for complex RSI data [17,18]. Thus, numerous extension methods based on PCA and LDA, such as robust PCA [19], sparse PCA [20], subspace LDA [21], regularized LDA [22], and so on [23,24,25], have been proposed. Several other techniques have also been introduced to tackle this problem and achieve highly advantageous results, such as utilizing locality-preserving projection [26], locality-preserving nonnegative matrix factorization [27], nonnegative matrix factorization [28], and Gaussian mixture models [29]. However, the spectral-based FE methods are commonly known as the dimensionality reduction method, which loses the detailed information and the spectral relationship of RSIs.
In the context of spatial-based FE methods, the gray-level co-occurrence matrices, Gabor wavelets [30], morphological profiles (MPs) [31], and Markov random fields [32] are widely applied. However, the spatial-based FE methods via neighborhoods of pixels were limited by the scale selection and the detailed content of very high-resolution images [33]. Akcay et al. [33] exploited the structural information of RSIs using morphological operators to address these problems and achieved excellent performance. Mallinis et al. [34] extracted local patterns of images via texture measurement methods and yielded superior classification results. However, the use of only spatial or spectral information may occasionally be insufficient to achieve desired results [17]. Therefore, combining the spatial and spectral information of RSIs has the potential to achieve high performance in FE.
As imaging technology continues to evolve, the spatial resolution of RSIs has significantly increased. This advancement enhances the accuracy of RSI applications by incorporating spatial information of neighboring positions into spectral-based feature extraction methods [35]. Therefore, numerous spatial-based FE methods have been introduced into spectral-based FE methods to construct spatial–spectral-based FE methods. The extended MPs were developed for spatial–spectral feature extraction of RSIs [36] and obtained a better classification result. Kang et al. [37] extracted the spatial–spectral features via edge-preserving filtering, which assumed the correlation among neighboring objects, and significantly improved the classification accuracy. In addition, the deep learning (DL) method has achieved remarkable achievements in the spatial–spectral-based FE method in RSIs [38,39]. Chen et al. [40] introduced a deep belief network to the extract spatial–spectral features of RSIs and achieved a high accuracy of classification results. Based on deep belief network, Ma et al. [41] added a regularization term to extract spectral–spatial features by introducing an energy function in a spatially updated deep auto-encoder, thereby providing encouraging results. Moreover, Kavitha et al. [42] proposed that a deep e-CNN model using a spatial–spectral feature merging strategy effectively improves the classification performance. However, the spatial–spectral-based FE method based on DL needs a large number of training samples, which increases its nonuniversality.
In the context of weak information extraction methods for RSIs, Kang [43,44] introduced a multi-granularity spectral segmentation (MGSS) technique addressing hyperspectral weak information mining and achieved excellent performance in the quantitative application of remote sensing. MGSS enables multi-granularity segmentation (MGS) and multi-granularity reconstruction (MGRS) of spectral data via multi-granularity residual (MGR). Kang [45] proposed an optimal pixel-based residual spectral reconstruction method based on the MGRS and improved estimation accuracy of aboveground biomass on grasslands. Building on the foundation of the MGR, Pang et al. [46] simulated hyperspectral data using the multispectral of sample points and achieved favorable results in grassland micronutrient estimation. Fan et al. [10] expanded the MGSS to soil salinity monitoring and found that MGSS technology can effectively extract weak spectral information from soil, compared with the traditional spectral preprocessing methods, significantly improved the correlation between spectra and soil. An increasing number of reports have addressed this technology since the introduction of MGSS. However, the MGSS method focuses on the spectral domain and disregards the spatial information of RSIs, which limits its application in remote sensing.
The current study attempts to develop a spatial–spectral-based FE technique to address the aforementioned problem. This technique can comprehensively consider the three-dimensional structure and global information of RSIs to mine the weak information in RSIs. Inspired by MGSS theory, this study proposes a novel spatial–spectral-based FE technique by combining low-rank presentation (LRR), high-order residual quantization (HORQ), and MGSS and discusses the characteristic and practicality of the extracted feature via its classification results.

2. Basic Methodology

HORQ, MGSS, and LRR are the important foundations of the proposed method. HORQ and MGSS are used to mine the weak information in theory, and LRR is used to explore the global structure of RSIs. Therefore, the four underlying components are briefly reviewed in this section.

2.1. High-Order Residual Quantization

The HORQ technique [47] was proposed to accelerate and compress neural networks. Building on the foundation of XNOR-Net [48], HORQ incorporates +1 or −1 Binary Coding (BC ± 1) [49] to achieve significant computational acceleration while maintaining network accuracy. HORQ approximates the high-dimensional vector X as X β 1 H 1 , where H 1 represents the BC ± 1 coding matrix of X , and β 1 is a positive factor, which denotes the weighting coefficient of   H 1 . The estimations for X and β 1 H 1 are as follows:
β 1 * , H 1 * = a r g m i n β 1 , H 1 ( β 1 , H 1 ) = a r g m i n β 1 , H 1 X β 1 H 1 l 2 .
Considering Formula (1), Li [47] and Rastegari [48] provided comprehensive descriptions of its analytical steps. The solution was as follows:
β 1 * = 1 N X l 1 H 1 * = s i g n ( X ) ,
where s i g n ( X ) represents the sign function of X . Furthermore, concerning the residual of X , Li defined residual as follows [47]:
R i 1 ( X ) = X i = 1 N β i H i ,
where N is a positive integer greater than 1, and the encoding feature of X at i-order is defined as follows:
C i ( X ) = β i H i .

2.2. Multi-Granularity Spectral Segmentation

Based on the HORQ method, Kang et al. [43] introduced high-order binary coding (HOBC) for hyperspectral images (HSIs), which enables high-performance coding compression and lossless recovery of HSIs. Subsequently, Kang [45] extended HOBC to introduce the MGSS technique. MGSS involves continuous quantization and decomposition of the spectral information and mines weak spectral features within specific wavelength ranges. This technique assumes the acquisition of spectral vectors as G, and its Nth order reconstructed features can be expressed as follows:
G = i = 1 N β i H i + R N ( G ) i = 1 N β i H i .
Kang [43] provided detailed derivations and explanations of the analytical derivation of Formula (6). Thus, β i and H i can be represented as below:
β i = 1 N R i 1 ( G ) l 1 H i = s i g n ( R i 1 ( G ) ) .
The MGSS technique performs spectral feature decomposition on sampled points to attain per-pixel spectral features of RSIs. However, the aforementioned methods concentrate on spectral features on a per-pixel basis, disregarding the three-dimensional structure that is inherent in RSIs.

2.3. Low-Rank Representation of RSIs

The RSI is a three-dimensional cube that combines imagery and spectral characteristics. LRR is the introduced technique to fully consider the global spatial and spectral information of raw images. LRR is widely used in image processing due to its impressive performance in capturing the global data structure [50]. Therefore, LRR is proposed for RSIs to mine their global spectral–spatial character. During the FE process, the three-dimensional RSI is transformed into a two-dimensional low-rank matrix to incorporate the global spatial–spectral information, allowing for globally optimized extracted features. Figure 1 shows the flowchart of the LRR for RSIs.
Suppose G H m × n × b is the achieved RSI, m × n is the spatial size of G , and b is its band number. Thus, G can be introduced as follows:
G = A E ,
where A H m n × b is the abundance matrix, E H p × b is the endmember matrix, and p is the number of endmembers. G is factorized into products of A and E ; thus, the low rank of G should be satisfied.
r a n k ( G ) min ( r a n k ( A ) , r a n k ( E ) ) < < min ( m n , b ) .
Therefore, if the LRR is used in the feature extraction for RSI, then it can be expressed as below:
L * = arg min r a n k E s . t .   G A E = 0 ,   G 0 ,   1 T G = 1 T ,
where G 0 ,   1 T G = 1 T indicates E as nonnegative and oneness constraints.

3. Datasets and Methods

The datasets employed in this study are first introduced. The derivation of the proposed method is then described in detail.

3.1. Datasets

This study conducted validation using three different RSI datasets to verify the effectiveness of the proposed method in the weak information mining of RSIs, and the description of the three datasets is as follows:
(1) Chikusei dataset (CK dataset) [51]: The CK dataset was achieved using the Headwall Hyperspec-VNIR-C imaging sensor over agricultural and urban areas in Chikusei, Ibaraki, Japan (Figure 2a). The image in the CK dataset comprises 2517 × 2335 pixels and contains 128 bands, its spectral range is from 363 nm to 1018 nm, and its spatial resolution is 2.5 m. The ground truth (GT) has 19 classes (including urban and rural areas) and was collected via field survey and visual inspection using high-resolution color images obtained using a Canon EOS 5D Mark II together with the hyperspectral data.
(2) ZY1-02D dataset (ZY dataset) [52]: The ZY dataset was achieved using the hyperspectral sensor on the ZY1-02D satellite in Changzhou city, China (Figure 2b). This dataset has 166 contiguous bands covering the spectral wavelength range from 0.4 µm to 2.5 µm and comprises 1999 × 2051 pixels. The spatial and spectral resolutions of this dataset are 30 m and 10 nm, respectively. The GT is only divided into five classes (i.e., Water, Bared, Vegetation, Building, and Road), which are visually inspected using a high-resolution multispectral image obtained by the ZY1-02D satellite and field survey, due to the low spatial resolution (around 30 m) of the ZY dataset.
(3) Pasture dataset (PT dataset): The PT dataset was collected in the Hulunbuir grassland using the camera of a Huawei P40 mobile phone. The image in the PT dataset was obtained via vertical photography with a 1.5 m height. The RGB image in the PT dataset comprises 8192 × 6144 pixels. The GT for all RGB images in this dataset was obtained through expert visual interpretation combined with field surveys, leading to the classification of data into 10 categories. For the demands of this study, the used RGB image in the PT dataset only contains five categories (Figure 2c).
The CK dataset was employed to assess the spatial and spectral features using the deep spatial–spectral residual feature (DMSR) and deep spatial–spectral coding feature (DMSC). Additionally, the CK dataset was used to determine the intraclass discrimination capability of DMSR and DMSC in fine classification. The ZY dataset was utilized to evaluate the effectiveness of DMSRs and DMSCs in rough classification and examine the interclass identification capacities of DMSR and DMSC. The CK dataset and ZY dataset contain typical remote sensing images; however, RGB images are also an important source of remote sensing data. Thus, this study also aims to test whether or not the proposed method can be used to analyze RGB images with reasonable performance. Therefore, the PT dataset was utilized to explore the feasibility of extending DMSRE to RGB images.

3.2. Deep Multi-Order Spatial–Spectral Residual Feature Extraction

The study aims to develop a spatial–spectral feature extraction technique for mining weak information in RSIs. First, the proposed method tackles the global information within RSIs by converting the three-dimensional structure of raw RSIs into a two-dimensional spatial–spectral information matrix prior to the extraction process. Subsequently, the extracted spatial–spectral features are reorganized into a three-dimensional structure, ensuring dimensional consistency with the raw RSIs. Two sequences of spatial–spectral features are extracted by iteratively decomposing the residuals of the raw RSIs. Figure 3 illustrates the flowchart of the proposed method.
The proposed method extracts spatial–spectral features from RSIs in two directions, at multiple depths, and in various orders (Figure 3) and thus is named deep multi-order spatial–spectral residual feature extraction (DMSRE). DMSRE can extract two sequences of spatial–spectral features with the same dimension of RSIs. The two spatial–spectral feature sequences are referred to as DMSRs and DMSCs. Each sequence of DMSR and DMSC comprises features of multiple orders, with the nth order DMSR or DMSC labeled as the n-order DMSC or DMSR. Furthermore, DMSRE is derived as follows:
Suppose L is a low-rank matrix of the achieved RSI G . If the 1st-order DMSR of L is defined as R ( L ) 1 , then L can be expressed as follows:
L = w 1 C 1 + R ( L ) 1 ,
where C 1 1 , + 1 N is the coding matrix of L and w 1 is the weight of C 1 . C 1 and w 1 can generally be calculated as follows:
w 1 = 1 m n L 1 ,
C 1 = s i g n ( L ) .
The R L 1 in Formula (11) can then be further quantized as follows:
R L 1 = w 2 C 2 + R L 2 .
Thus, Formula (10) can be rewritten:
L = w 1 C 1 + w 2 C 2 + R L 2 ,
where
w 2 = 1 m n R L 1 1 ,
C 2 = s i g n ( R ( L ) 2 ) .
Moreover, the R L 2 can be quantized again. Therefore, the L can finally be introduced as follows:
L = i N w i C i + R ( L ) i .
Formula (17) shows that the R L i   is close to 0, while the order N is sufficiently high. Thus, Formula (10) can be written as below:
L i N w i C i .
Simultaneously, the analytical solutions of the weighted and coding matrices for each depth are, respectively, presented:
w 2 = 1 m n R L 1 1 ,
C 2 = s i g n ( R ( L ) 2 ) .
Finally, the DMSC and DMSR can be, respectively, expressed:
D M S C i = i N w i c i ,
D M S R i = L i N w i c i .
Formula (21) shows that the information content of the DMSC sequence features increases with the order until it converges with that of the raw image. Conversely, the information content of DMSR (Formula (22)) decreases as the order increases until it converges to zero. Combining Formulas (21) and (22), the raw RSI can be reconstructed by the same order of DMSR and DMSC.
DMSRE can theoretically extract the feature indefinitely, with the terminated order depending on the raw RSI and its application demands. Therefore, the mean spectral angle (MSA) [53] and structural similarity index (SSIM) [54] were introduced to discuss the terminated order. MSA characterizes the spectral similarity between the DMSC and the raw RSI, whereas SSIM represents the structural similarity between the DMSC and the raw RSI. These metrics are defined as follows:
S S I M ( X , Y ) = ( 2 μ X μ Y + C 1 ) ( 2 σ X Y + C 2 ) ( μ X 2 + μ Y 2 + C 1 ) ( σ X 2 + σ Y 2 + C 2 ) ,
M S A = 1 m n i = 1 m n 180 π a r c c o s X T Y X Y ,
where X and Y denote the raw RSI and DMSC, respectively. Herein, μ X   signifies the mean value of X, μ Y represents the mean value of Y , and σ X 2 and σ Y 2 correspond to the variances of X and Y, respectively. σ X Y stands for the covariance between X and Y. Constants C 1 and C 2 are introduced to ensure numerical stability during computations. In Formula (24), m and n refer to the number of rows and columns of raw RSIs, respectively. The value of SSIM lies in the range [0, 1], where high values indicate high structural similarity between DMSCs and raw RSIs. Conversely, small values for the MSA indicate high spectral similarity.

3.3. Evaluation Metrics

In this study, F1 scores were employed to evaluation the classification results. The F1 score was evaluated based on the recall and precision, which can be expressed as follows:
R e c a l l = T P T P + F N ,
P r e c i s i o n = T P T P + F P ,
where TP represents true positives, TN true negatives, FP false positives, and FN false negatives. Moreover, the F1 score can be expressed as follows:
F 1 _ s c o r e = 2 R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n
A higher F1 score indicates a better classification result. The overall accuracy and the classification accuracy of the single class in the classification result were used to evaluate the F1 score.

4. Results

4.1. Experimental Setting

Theoretically, the proposed method can continuously extract features indefinitely based on the raw image. However, the extracted DMSC is slightly changed while the order is high. Moreover, the terminated order varies for different datasets and demands of applications. This study attempts to find the terminated order, which is highly spectrally and spatially similar to the raw CK dataset (Figure 4). Considering the CK dataset as an example, first-order to sixteenth-order DMSCs and DMSRs are extracted from the CK dataset, and the MSA and SSIM values between the DMSC and the raw CK dataset are then calculated (Figure 5).
It can be easily found that the SSIM is closer to 1 while the order is higher than the eighth order, and the MSA is smaller than 0.05 (Figure 5). In this time, the DMSC is highly similar to the raw image in the image structure and spectral domain. Although there were some orders present that were higher (SSIM) and lower (MSA) than the eight order, little changed compared with the eighth-order DMSC. Therefore, considering the computational power and time costs, this study chose the eighth order as the terminal order of the CK dataset.

4.2. Spatial and Spectral Features of DMSR and DMSC

4.2.1. Spatial Features of DMSR and DMSC

The DMSR was obtained by continuously decomposing the residual of raw RSIs, which indicates that the difference between the DMSR and raw image increases with order. Therefore, the spatial visual of the DMSR displays the image darker than the raw CK dataset in the beginning (Figure 6a), the ground objects in the DMSR gradually fragment with high order (Figure 6), and the shape of the ground object cannot be observed in the high order (Figure 6g,h).
The color images of the DMSC for the CK dataset are shown in Figure 6. The figure demonstrates a black color in the first-order DMSC (Figure 7a), which only contains the mean value of L1 paradigm from the raw CK dataset, and the color varies depending on raw RSIs. The information content of the DMSC gradually increases with the order. High-reflectivity features, such as the roof and bare ground, are emphasized in the second-order DMSC (Figure 7b). Vegetation such as farmland is highlighted in the third-order DMSC (Figure 7c). Most ground objects are displayed in the fourth-order DMSC (Figure 7d) and are visually increasingly similar to the raw image. Two detailed areas were selected and enlarged to perform a detailed analysis to further understand the DMSR and DMSC trend, as shown in Figure 8.
Detailed image 1 shows the change in the DMSC with order (Figure 8b–i). The figure reveals a black color at the first-order DMSC (Figure 8b), the high-reflectivity features (e.g., roof) are displayed in the second order (Figure 8c), the vegetation (e.g., grassland) is emphasized in the third order, the outline of most classes was recognized but without intraclass information in the fourth order (Figure 8d), and the intraclass information was refined when the order was higher than the fifth order (Figure 8f–i).
Detailed image 2 is used to demonstrate the change in the DMSR with order (Figure 8k–r). The shapes of surface features show a tendency toward gradual fragmentation. The DMSR is visually similar to the raw image in the first order (Figure 8k), the features (e.g., roof) start to be gradually fragmented in the second order (Figure 8l), the weak information has been gradually reflected when the order is higher than third order (Figure 8m); for example, the information in the red box (Figure 8k–o), which is the shadow of the cloud on the ground, is difficult to distinguish from the surrounding grass. However, the surrounding information in the third order to fifth order (Figure 8m–o) demonstrates the capability of the DMSR to mine weak information in RSIs. Overall, the results indicate that the DMSC is appropriate for rough classification, whereas the DMSR is better suited for fine classification. The DMSC and DMSR have the potential to enhance the accuracy of remote sensing classification.

4.2.2. Spectral Features of DMSR

This study extracted the spectral features of vegetation in the DMSC and DMSR as an example for analysis. The spectral curves are displayed in Figure 9.
In Figure 9, the blue curve is the spectral feature achieved from a raw RSI, while the purple and green curves represent the curves of the DMSR and DMSC, respectively. The DMSC shows a straight line in the first order (green line in Figure 9a), which only contains the average value of its L1 paradigm value. The second-order DMSC (Figure 9b) illustrates the low-reflectivity features (e.g., vegetation) in visible spectral (0.37–0.71 µm) and high-reflectivity features in near-infrared spectral. The first- and second-order DMSCs only retain minimal spectral information of raw RSIs, and the detailed spectral information of the DMSC improves as the order increases (Figure 9c–h), bringing it close to raw RSIs.
On the contrary, the first-order DMSR (purple curve in Figure 9a) maintains all spectral information but is only small in terms of spectral value because the first-order DMSR represents the raw RSIs minus its L1 paradigm value. The second-order DMSR (Figure 8b) highlights detailed information in the red edge spectral (approximately 0.7 μm wavelength). The DMSR has increased spectral information in the first and second order, and the detailed information becomes rich while the order becomes high (Figure 9b–h). Additionally, as the order increases, the spectral value of the DMSR tends toward 0, while that of the DMSC approaches the raw RSI’s.

4.3. Classification Results Using DMSR and DMSC

This study performed classification experiments on raw RSIs and their extracted DMSR and DMSC from three RSI datasets via random forest (RF) classifier to verify the performance of the DMSC and DMSR. The CK dataset was applied for fine classification, while the ZY dataset was applied for rough classification. The classification accuracy of each single class and the overall accuracy were obtained from these classification results. High values indicate superior classification results. Figure 10 and Figure 11 illustrate the classification results for the DMSC and DMSR of the CK dataset, respectively.
The classification results using the DMSR (Figure 10) show minimal visual variation. However, significantly more misclassifications are observed while the orders are higher than the fifth order (Figure 10e–h). By contrast, the classification results via DMSC (Figure 11) gradually increase the number of recognized classes as the order rises, and most classes have been recognized while the orders are greater than the fifth order (Figure 11e–h). The classification and overall accuracies were calculated and are listed in Table 1 and Table 2, respectively, based on the classification results of the DMSR and DMSC.
Table 1 provides the classification accuracy of single class and the overall accuracy of classification results; the F1 score was used to assess classification result both of overall accuracy and the single class (the same applies to the following content). The overall accuracy of the DMSR classification result displays a minor decrease as the order increases. However, several individual classes achieved higher classification accuracy than the raw image (highlighted in blue in Table 1), because the weak information related those classes had been mined in the DMSR. The above conclusion supports that the DMSR can improve the accuracy of individual class in fine classification application.
Table 2 shows the accuracy of classification results via DMSC. The overall accuracy demonstrates an increasing trend with the orders. Notably, there are many 0 values in classification results via the first- and second-order DMSC, because first- and second-order DMSCs contain little RSI information. Compared with the classification results of the raw image, the overall accuracy of the DMRC (especially orders higher than the sixth order) only has minor differences. Meanwhile, the rough classification experiment was conducted on the ZY dataset. The classification results via DMSR and DMSC of the ZY dataset are shown in Figure 12 and Figure 13, respectively.
The classification results by first- and second-order DMSR (Figure 12a,b, respectively) can identify the most classes. The number of recognized classes slowly diminishes starting from the third order (Figure 12c), whereas the changes in recognized classes be-come less pronounced after the fifth-order DMSR (Figure 12e). The classification results via first-order DMSC (Figure 13a) only identified the water class, that is, the number of the identified classes with an increasing number as the order increases, from a visual perspective. The changes become less pronounced after the fifth order (Figure 13e). The accuracy of the DMSR and DMSC for the ZY dataset is listed in Table 3 and Table 4, respectively.
The Water and Building class in the ZY dataset can achieve high accuracy in the classification via DMSR (Table 3): the accuracy of Water improved from 95.46% to 97.56%, while that of Building improved from 96.72% to 99.71%. The overall accuracy in the classification via DMSC (Table 4) achieved a superior result, and the overall accuracy improved from 92.07% to 92.78%. This phenomenon is attributed to the DMSC’s emphasis on highlighting object boundaries by disregarding local detailed information. Consequently, the DMSC tends to yield superior results for rough classification. The DMSC theoretically has potential for improvement in overall accuracy. Furthermore, this study conducts experiments on RGB images in the PT dataset to discuss the scalability of DMSRE. Figure 14 and Figure 15 display the color image of the DMSR and DMSC, respectively.
The spatial variations in DMSR (Figure 14) and DMSC (Figure 15) of the PT dataset are consistent with those in the DMSR and DMSC in the CK and ZY datasets. The amount of information for the DMSR gradually reduces as the orders increase. The first-order DMSR (Figure 14a) only demonstrates a slight decrease in brightness. The difference between the raw image and the DMSR grows larger starting from the Fifth order (Figure 14e). From the fifth to eighth order (Figure 14e–h), only the goat grass feature remains. In the first-order DMSC (Figure 15a), the image is monochromatic, and the DMSC tends to be close to the raw image as the order increases. The classification accuracies of the DMSR and DMSC are presented in Table 5 and Table 6, respectively.
From the classification result via raw image, it can be found that the classification accuracy was not higher; only “Leymus chinensis” achieved higher accuracy, and other classes had lower accuracy. This is because “Leymus chinensis” is the main dominant species in the study area, and the other classes are companion species, which means the “Leymus chinensis” class accounted for a large proportion in the PT dataset and other classes had less proportion. Therefore, “Leymus chinensis” easily achieved a higher classification result than other classes. However, it still shows difficulty in companion species classification, and the classification results via DMSR show the potential for this task. In the classification results via DMSR (Table 5), most classes attain higher accuracy than the classification result of the raw image, excluding the “Leymus chinensis” class. However, it improved the classification accuracy of companion species slightly, which also means it has potential for species classification. The overall accuracy performance in the DMSC coincides with the results observed in the two other RSI datasets (Table 6). The overall accuracy steadily improved with the increase in orders, ultimately exceeding the overall accuracy of the raw image.

5. Discussion

5.1. Research Contributions

The main contribution of this study is as follows: First, this study proposes a deep multi-order spatial–spectral feature extraction method that adaptively extracts spatial–spectral features of RSIs from two directions, which does not need training samples. The proposed method further ensures that the extracted feature does not reduce in dimension compared with the raw RSIs because it considers the comprehensive global spatial–spectral information of the RSIs to achieve optimal results. Second, the proposed method can enhance the feature recognition capability of RGB images, demonstrating the universality of the approach. This method is not only applicable to HSI but also to RGB images.
Based on the concept of weak information extraction from RSIs, MGSS [44] was recently developed to facilitate spectral multifeatured weak information extraction technology, specifically for feature extraction in the spectral dimension. The extracted features can be effectively quantified to estimate accuracy, and research has shown certain success in quantitative applications, such as the estimation of forage aboveground biomass [45] and grass crude protein [44], and the monitoring of soil salinity [10].
However, MGSS is limited to achieving good performance in the spectral dimension and cannot be applied to the image dimension. However, information mining using MGSS in RSIs is yet to be explored further. The proposed method is based on LRR, MGSS, and HORQ theories. The extracted DMSRs and DMSCs possess unique properties. The DMSR calculates residuals of the raw image step by step, focusing on weak spatial–spectral information. By contrast, the DMSC highlights interclass boundary information by progressively accumulating the product of coding matrices and magnitudes in the convergence of the raw image. Moreover, this study employed remote sensing classification application to test their properties to effectively understand the characteristics of the DMSR and DMSC. Notably, this finding does not imply that the two features are exclusive to classification applications; rather, they have the potential to have a good performance in qualitative and quantitative applications for the purpose of weak information mining.

5.2. Limitation and Potential Future Work

This study proposed the DMSRE method for the spatial–spectral feature extraction of RSIs. However, DMSRE has certain limitations. While its mathematical foundations have been established, certain gaps in the understanding of its underlying mechanisms exist. Particularly, at lower orders, DMSRE constitutes the dominant factor in the image, yet the specific information represented by each order remains uncertain. For instance, similar to the tasseled cap transformation [55], which explicitly defines the first, second, and third components as brightness, greenness, and wetness, respectively, a similar comprehensive interpretation for DMSRE is still lacking. Furthermore, the investigation into the amplitude values (denoted as “w”) in the DMSCs of RSIs requires further exploration. This issue has also been addressed in the work of Kang [44], signifying the ongoing need for additional in-depth research in this area. Moreover, to fully consider spatial–spectral characteristics, this study handles spatial features in one dimension. The extraction of multidimensional spatial–spectral features is possible, as it also handles some spatial features to some extent; thus, the discussion of its practicality and universality for other applications should be conducted in an actual RS task in the near future. For a similar problem, Truong et al. [56] extracted spatiotemporal features with deep learning for land cover classification, and achieved a high-accuracy land cover classification. Therefore, we will focus on how to use the extracted features to improve the application performance in the next work.
The DMSR and DMSC exhibit distinct characteristics and hold varying application potentials in remote sensing. This investigation not only delves into their potential in classification but also acknowledges their possible promising prospects in remote sensing estimation and inversion due to the DMSR’s weak information extraction capabilities.

6. Conclusions

This study considers the three-dimensional structural properties and global information of RSIs, and without training samples during FE process, this study proposed an adaptive DMSRE method based on LRR, MGSS, and HORQ that can extract features of RSIs in two directions (DMSRs and DMSCs). Classification experiments with DMSRs and DMSCs from three datasets (CK, ZY, and PT datasets) with different attributes showed that the DMSR can improve the classification accuracy of individual classes in fine classification applications, and the DMSC can improve the overall accuracy in rough classification applications. Moreover, the proposed method is not only applicable to HSI but extends to remote sensing RGB images, which has the potential to promote the RSI processing of weak extracted information, and then extend the application of RSIs in different fields.

Author Contributions

Methodology, X.Z.; Software, H.P.; Validation, X.Z.; Data curation, Y.C. and J.Z.; Writing—original draft, X.Z.; Writing—review & editing, A.Z., J.W., J.P., V.G., T.G.L. and X.X.; Supervision, Y.S.; Project administration, A.Z. and C.S.; Funding acquisition, A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Number 42071303), the National Key Research and Development Program of China (Grant Number 2021YFD1300505), the Joint program of Beijing Municipal Education Commission and Beijing Municipal Natural Science Foundation of China (Grant Number KZ202110028044) and the Science and Technology Program of Qinghai Province of China (Grant Number 2022-NK-136).

Data Availability Statement

Data is unavailable due to privacy.

Acknowledgments

The authors gratefully acknowledge Space Application Laboratory, Department of Advanced Interdisciplinary Studies, University of Tokyo for providing the hyperspectral data.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Shen, H.; Li, X.; Cheng, Q.; Zeng, C.; Yang, G.; Li, H.; Zhang, L. Missing Information Reconstruction of Remote Sensing Data: A Technical Review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 61–85. [Google Scholar] [CrossRef]
  2. Qin, P.; Cai, Y.; Liu, J.; Fan, P.; Sun, M. Multilayer Feature Extraction Network for Military Ship Detection from High-Resolution Optical Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 11058–11069. [Google Scholar] [CrossRef]
  3. Shirmard, H.; Farahbakhsh, E.; Müller, R.D.; Chandra, R. A Review of Machine Learning in Processing Remote Sensing Data for Mineral Exploration. Remote Sens. Environ. 2022, 268, 112750. [Google Scholar] [CrossRef]
  4. Chen, H.; Qi, Z.; Shi, Z. Remote Sensing Image Change Detection with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5607514. [Google Scholar] [CrossRef]
  5. Ozdogan, M.; Yang, Y.; Allez, G.; Cervantes, C. Remote Sensing of Irrigated Agriculture: Opportunities and Challenges. Remote Sens. 2010, 2, 2274–2304. [Google Scholar] [CrossRef]
  6. Yu, S.; De Backer, S.; Scheunders, P. Genetic Feature Selection Combined with Composite Fuzzy Nearest Neighbor Classifiers for High-Dimensional Remote Sensing Data. In Proceedings of the SMC 2000 International Conference on Systems, Man and Cybernetics. “Cybernetics Evolving to Systems, Humans, Organizations, and their Complex Interactions” (Cat. No.00CH37166), Nashville, TN, USA, 8–11 October 2000; Volume 3, pp. 1912–1916. [Google Scholar]
  7. Bhuvaneswari, K.; Dhamotharan, R.; Radhakrishnan, N. Information Extraction from Remote Sensing Image (RSI) for a Coastal Environment Along a Selected Coastline of Tamilnadu. IJCSET Board Memb. 2011, 95. [Google Scholar]
  8. Han, W.; Li, J.; Wang, S.; Wang, Y.; Yan, J.; Fan, R.; Zhang, X.; Wang, L. A Context-Scale-Aware Detector and a New Benchmark for Remote Sensing Small Weak Object Detection in Unmanned Aerial Vehicle Images. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102966. [Google Scholar] [CrossRef]
  9. Sun, Y.; Cai, W.; Shao, X. Chemometrics: An Excavator in Temperature-Dependent Near-Infrared Spectroscopy. Molecules 2022, 27, 452. [Google Scholar] [CrossRef]
  10. Fan, X.; Kang, X.; Gao, P.; Zhang, Z.; Wang, J.; Zhang, Q.; Zhang, M.; Ma, L.; Lv, X.; Zhang, L. Soil Salinity Estimation in Cotton Fields in Arid Regions Based on Multi-Granularity Spectral Segmentation (MGSS). Remote Sens. 2023, 15, 3358. [Google Scholar] [CrossRef]
  11. Wang, J.; Zhen, J.; Hu, W.; Chen, S.; Lizaga, I.; Zeraatpisheh, M.; Yang, X. Remote Sensing of Soil Degradation: Progress and Perspective. Int. Soil Water Conserv. Res. 2023, 11, 429–454. [Google Scholar] [CrossRef]
  12. Abdollahi, A.; Pradhan, B.; Shukla, N.; Chakraborty, S.; Alamri, A. Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-Of-The-Art Review. Remote Sens. 2020, 12, 1444. [Google Scholar] [CrossRef]
  13. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution from Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  14. Ebied, H.M. Feature Extraction Using PCA and Kernel-PCA for Face Recognition. In Proceedings of the 2012 8th International Conference on Informatics and Systems (INFOS), Giza, Egypt, 14–16 May 2012; p. MM-72. [Google Scholar]
  15. Aliyari Ghassabeh, Y.; Rudzicz, F.; Moghaddam, H.A. Fast Incremental LDA Feature Extraction. Pattern Recognit. 2015, 48, 1999–2012. [Google Scholar] [CrossRef]
  16. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  17. Kumar, B.; Dikshit, O.; Gupta, A.; Singh, M.K. Feature Extraction for Hyperspectral Image Classification: A Review. Int. J. Remote Sens. 2020, 41, 6248–6287. [Google Scholar] [CrossRef]
  18. Ye, Z.; Bai, L.; He, M. Review of Spatial-Spectral Feature Extraction for Hyperspectral Image. J. Image Graph. 2021, 26, 1737–1763. [Google Scholar]
  19. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust Principal Component Analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  20. Zou, H.; Hastie, T.; Tibshirani, R. Sparse Principal Component Analysis. J. Comput. Graph. Stat. 2006, 15, 265–286. [Google Scholar] [CrossRef]
  21. Yang, J.; Yang, J. Why Can LDA Be Performed in PCA Transformed Space? Pattern Recognit. 2003, 36, 563–566. [Google Scholar] [CrossRef]
  22. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of Hyperspectral Images with Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  23. Du, Q. Modified Fisher’s Linear Discriminant Analysis for Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2007, 4, 503–507. [Google Scholar] [CrossRef]
  24. Tipping, M.E.; Bishop, C.M. Probabilistic Principal Component Analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 1999, 61, 611–622. [Google Scholar] [CrossRef]
  25. Liang, Z.; Xia, S.; Zhou, Y. Normalized Discriminant Analysis for Dimensionality Reduction. Neurocomputing 2013, 110, 153–159. [Google Scholar] [CrossRef]
  26. He, X.; Niyogi, P. Locality Preserving Projections. Adv. Neural Inf. Process. Syst. 2003, 16. [Google Scholar]
  27. Cai, D.; He, X.; Wang, X.; Bao, H.; Han, J. Locality Preserving Nonnegative Matrix Factorization. In Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence, Pasadena, CA, USA, 11–17 July 2009. [Google Scholar]
  28. Lee, D.D.; Seung, H.S. Learning the Parts of Objects by Non-Negative Matrix Factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef] [PubMed]
  29. Berge, A.; Solberg, A.S. Structured Gaussian Components for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3386–3396. [Google Scholar] [CrossRef]
  30. Bhagavathy, S.; Manjunath, B.S. Modeling and Detection of Geospatial Objects Using Texture Motifs. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3706–3715. [Google Scholar] [CrossRef]
  31. Pesaresi, M.; Benediktsson, J.A. A New Approach for the Morphological Segmentation of High-Resolution Satellite Imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 309–320. [Google Scholar] [CrossRef]
  32. Melgani, F.; Serpico, S.B. A Markov Random Field Approach to Spatio-Temporal Contextual Image Classification. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2478–2487. [Google Scholar] [CrossRef]
  33. Akcay, H.G.; Aksoy, S. Automatic Detection of Geospatial Objects Using Multiple Hierarchical Segmentations. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2097–2111. [Google Scholar] [CrossRef]
  34. Mallinis, G.; Koutsias, N.; Tsakiri-Strati, M.; Karteris, M. Object-Based Classification Using Quickbird Imagery for Delineating Forest Vegetation Polygons in a Mediterranean Test Site. ISPRS J. Photogramm. Remote Sens. 2008, 63, 237–250. [Google Scholar] [CrossRef]
  35. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2012, 101, 652–675. [Google Scholar] [CrossRef]
  36. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of Hyperspectral Data from Urban Areas Based on Extended Morphological Profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  37. Kang, X.; Li, S.; Benediktsson, J.A. Feature Extraction of Hyperspectral Images with Image Fusion and Recursive Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  38. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  39. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  40. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  41. Ma, X.; Wang, H.; Geng, J. Spectral–Spatial Classification of Hyperspectral Image Based on Deep Auto-Encoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  42. Kavitha, M.; Gayathri, R.; Polat, K.; Alhudhaif, A.; Alenezi, F. Performance Evaluation of Deep E-CNN with Integrated Spatial-Spectral Features in Hyperspectral Image Classification. Measurement 2022, 191, 110760. [Google Scholar] [CrossRef]
  43. Kang, X.; Zhang, A. A Novel Method for High-Order Residual Quantization-Based Spectral Binary Coding. Spectrosc. Spectr. Anal. 2019, 39, 3013–3020. [Google Scholar]
  44. Kang, X.; Zhang, A. Hyperspectral Remote Sensing Estimation of Pasture Crude Protein Content Based on Multi-Granularity Spectral Feature. Trans. Chin. Soc. Agric. Eng 2019, 35, 161–169. [Google Scholar]
  45. Kang, X.-Y.; Zhang, A.-W.; Pang, H.-Y. Estimation of Grassland Aboveground Biomass from UAV-Mounted Hyperspectral Image by Optimized Spectral Reconstruction. Spectrosc. Spectr. Anal. 2021, 41, 250–256. [Google Scholar]
  46. Pang, H.; Zhang, A.; Yin, S.; Zhang, J.; Dong, G.; He, N.; Qin, W.; Wei, D. Estimating Carbon, Nitrogen, and Phosphorus Contents of West–East Grassland Transect in Inner Mongolia Based on Sentinel-2 and Meteorological Data. Remote Sens. 2022, 14, 242. [Google Scholar] [CrossRef]
  47. Li, Z.; Ni, B.; Zhang, W.; Yang, X.; Gao, W. Performance Guaranteed Network Acceleration via High-Order Residual Quantization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2603–2611. [Google Scholar]
  48. Rastegari, M.; Ordonez, V.; Redmon, J.; Farhadi, A. Xnor-Net: Imagenet Classification Using Binary Convolutional Neural Networks; Springer: Berlin/Heidelberg, Germany, 2016; pp. 525–542. [Google Scholar]
  49. Courbariaux, M.; Hubara, I.; Soudry, D.; El-Yaniv, R.; Bengio, Y. Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or −1. arXiv 2016, arXiv:1602.02830. [Google Scholar]
  50. Liu, G.; Lin, Z.; Yu, Y. Robust Subspace Segmentation by Low-Rank Representation. In Proceedings of the 27th international conference on machine learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 663–670. [Google Scholar]
  51. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; Technical Report SAL-2016-05-27; Space Application Laboratory, The University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  52. Zhang, X.; Zhang, A.; Portelli, R.; Zhang, X.; Guan, H. ZY-1 02D Hyperspectral Imagery Super-Resolution via Endmember Matrix Constraint Unmixing. Remote Sens. 2022, 14, 4034. [Google Scholar] [CrossRef]
  53. Cao, C.; Yu, J.; Zhou, C.; Hu, K.; Xiao, F.; Gao, X. Hyperspectral Image Denoising via Subspace-Based Nonlocal Low-Rank and Sparse Factorization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 973–988. [Google Scholar] [CrossRef]
  54. Sara, U.; Akter, M.; Uddin, M.S. Image Quality Assessment through FSIM, SSIM, MSE and PSNR—A Comparative Study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar] [CrossRef]
  55. Zhang, X.; Schaaf, C.B.; Friedl, M.A.; Strahler, A.H.; Gao, F.; Hodges, J.C.F. MODIS Tasseled Cap Transformation and Its Utility. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Toronto, ON, Canada, 24–28 June 2002; Volume 2, pp. 1063–1065. [Google Scholar]
  56. Truong, V.T.; Hirayama, S.; Phan, D.C.; Hoang, T.T.; Tadono, T.; Nasahara, K.N. JAXA’s New High-Resolution Land Use Land Cover Map for Vietnam Using a Time-Feature Convolutional Neural Network. Sci. Rep. 2024, 14, 3926. [Google Scholar] [CrossRef]
Figure 1. Flowchart of LRR for RSI.
Figure 1. Flowchart of LRR for RSI.
Remotesensing 16 01957 g001
Figure 2. Color image sample of (a) CK dataset, (b) ZY dataset, (c) PT dataset; corresponding GT of (d) CK dataset, (e) ZY dataset, (f) PT dataset; and corresponding legend of (g) CK dataset, (h) ZY dataset, and (i) PT dataset.
Figure 2. Color image sample of (a) CK dataset, (b) ZY dataset, (c) PT dataset; corresponding GT of (d) CK dataset, (e) ZY dataset, (f) PT dataset; and corresponding legend of (g) CK dataset, (h) ZY dataset, and (i) PT dataset.
Remotesensing 16 01957 g002aRemotesensing 16 01957 g002b
Figure 3. Flowchart of the proposed method. (DMSR: deep spatial–spectral residual feature; DMSC: deep spatial–spectral coding feature).
Figure 3. Flowchart of the proposed method. (DMSR: deep spatial–spectral residual feature; DMSC: deep spatial–spectral coding feature).
Remotesensing 16 01957 g003
Figure 4. Schematic of this study.
Figure 4. Schematic of this study.
Remotesensing 16 01957 g004
Figure 5. SSIM (a) and MSA (b) values between the CK dataset and its DMSC.
Figure 5. SSIM (a) and MSA (b) values between the CK dataset and its DMSC.
Remotesensing 16 01957 g005
Figure 6. Color image of (ah) first- to eighth-order DMSR of the CK dataset.
Figure 6. Color image of (ah) first- to eighth-order DMSR of the CK dataset.
Remotesensing 16 01957 g006
Figure 7. Color images of (ah) first- to eighth-order DMSC for the CK dataset.
Figure 7. Color images of (ah) first- to eighth-order DMSC for the CK dataset.
Remotesensing 16 01957 g007
Figure 8. Image of (a) detailed image 1, (bi) first-order to eighth-order of DMSC of detailed image 1, (j) detailed image 2, and (kr) first-order to eighth-order DMSR of detailed image 2.
Figure 8. Image of (a) detailed image 1, (bi) first-order to eighth-order of DMSC of detailed image 1, (j) detailed image 2, and (kr) first-order to eighth-order DMSR of detailed image 2.
Remotesensing 16 01957 g008
Figure 9. Spectral (vegetation) features of (ah) the spectral curves of the first-order to eighth-order DMSC and DMSR.
Figure 9. Spectral (vegetation) features of (ah) the spectral curves of the first-order to eighth-order DMSC and DMSR.
Remotesensing 16 01957 g009
Figure 10. Classification results of (ah) first-order to eighth-order DMSR of the CK dataset.
Figure 10. Classification results of (ah) first-order to eighth-order DMSR of the CK dataset.
Remotesensing 16 01957 g010
Figure 11. Classification results of (ah) the first-order to eighth-order DMSC of the CK dataset.
Figure 11. Classification results of (ah) the first-order to eighth-order DMSC of the CK dataset.
Remotesensing 16 01957 g011aRemotesensing 16 01957 g011b
Figure 12. Classification results of (ah) first-order to eighth-order DMSR of the ZY dataset.
Figure 12. Classification results of (ah) first-order to eighth-order DMSR of the ZY dataset.
Remotesensing 16 01957 g012aRemotesensing 16 01957 g012b
Figure 13. Classification results of (ah) first-order to eighth-order DMSC of the ZY dataset.
Figure 13. Classification results of (ah) first-order to eighth-order DMSC of the ZY dataset.
Remotesensing 16 01957 g013
Figure 14. Color image of (ah) first-order to eighth-order DMSR of the PT dataset.
Figure 14. Color image of (ah) first-order to eighth-order DMSR of the PT dataset.
Remotesensing 16 01957 g014
Figure 15. Color image of (ah) first-order to eighth-order DMSC of the PT dataset.
Figure 15. Color image of (ah) first-order to eighth-order DMSC of the PT dataset.
Remotesensing 16 01957 g015aRemotesensing 16 01957 g015b
Table 1. Classification accuracy of different-order DMSRs of the CK dataset.
Table 1. Classification accuracy of different-order DMSRs of the CK dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Water0.99390.9951 0.9916 0.9926 0.9939 0.9924 0.9754 0.9395 0.9402
School0.98990.9850 0.9931 0.9855 0.9720 0.9483 0.9257 0.8300 0.7782
Park0.78600.7146 0.8253 0.7566 0.5662 0.1176 0.5160 0.0000 0.0156
Farmland0.98800.9826 0.9891 0.9847 0.9779 0.9630 0.9604 0.9446 0.9200
Plants0.99350.9879 0.9895 0.9922 0.9893 0.9924 0.9849 0.9817 0.9834
Weeds 0.97040.9541 0.9530 0.9681 0.9626 0.9467 0.8929 0.9016 0.9045
Forest0.99830.9962 0.9962 0.9951 0.9883 0.9857 0.9800 0.9742 0.9670
Grass0.99530.9926 0.9957 0.9959 0.9948 0.9891 0.9772 0.9692 0.9611
Rice field (grown)0.99870.9975 0.9978 0.9945 0.9847 0.9789 0.9739 0.9652 0.9539
Rice field0.99510.9991 0.9973 0.9900 0.9913 0.9863 0.9842 0.9584 0.9689
Row crops0.99710.9945 0.9936 0.9964 0.9947 0.9931 0.9846 0.9742 0.9570
House0.96860.9794 0.9791 0.9687 0.9265 0.8136 0.7306 0.5956 0.5482
Manmade 0.98340.9834 0.9871 0.9680 0.9148 0.8597 0.8416 0.8203 0.8099
Manmade (dark)0.98930.9942 0.9927 0.9928 0.9926 0.9885 0.9827 0.9723 0.9712
Manmade
(Red)
1.00000.9935 0.9868 0.9987 0.9564 0.8688 0.7524 0.5589 0.2982
Manmade
(Blue)
0.99500.9928 0.9700 0.9199 0.9258 0.8371 0.0741 0.0000 0.0000
Manmade grass0.98790.9873 0.9914 0.9835 0.9756 0.9001 0.8633 0.6983 0.5340
Asphalt 0.89120.9599 0.9511 0.9456 0.9391 0.9151 0.8593 0.5614 0.4500
Ground0.93390.9153 0.8279 0.8851 0.8185 0.8302 0.0876 0.0000 0.0000
Overall accuracy0.99220.9913 0.9921 0.9902 0.9824 0.9708 0.9570 0.9324 0.9177
Table 2. Classification accuracy of different-order DMSCs of the CK dataset.
Table 2. Classification accuracy of different-order DMSCs of the CK dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Water0.99390.00000.00000.85490.85260.97110.99130.99200.9931
School0.98990.00000.00000.92900.98260.98680.99050.99410.9920
Park0.78600.00000.00000.00000.35730.60380.67980.66990.7229
Farmland0.98800.00000.00000.95030.96200.97850.97870.98250.9833
Plants0.99350.00000.80340.96340.98290.98960.98900.99410.9927
Weeds0.97040.00000.00200.87490.92500.94560.94010.95900.9584
Forest0.99830.41840.79920.92500.96880.99360.99420.99630.9968
Grass0.99530.00000.70990.98190.99260.99440.99420.99640.9945
Rice field (grown)0.99870.00000.75300.89470.95740.99300.99430.99620.9976
Rice field0.99510.00000.00000.98850.98510.99650.99610.99650.9978
Row crops0.99710.00000.31280.96170.98380.99190.99370.99280.9920
House0.96860.00000.12830.89950.96910.97480.97820.97600.9836
Manmade0.98340.00000.96560.97810.98190.98130.98580.98900.9843
Manmade (dark)0.98930.00000.42340.94500.94850.98550.99200.99150.9931
Manmade
(Red)
1.00000.00000.36440.99481.00001.00001.00000.99240.9963
Manmade
(Blue)
0.99500.00000.80340.96241.00001.00000.99500.99260.9949
Manmade grass0.98790.00000.13530.86880.98210.98770.98750.98460.9874
Asphalt0.89120.00000.00000.65620.92530.95080.92970.93840.9488
Ground0.93390.00000.00000.00000.54000.87870.70190.86640.7500
Overall accuracy0.99220.26450.58390.66840.96250.98700.98900.99110.9917
Table 3. Classification accuracy of first-order to eighth-order DMSR of the ZY dataset.
Table 3. Classification accuracy of first-order to eighth-order DMSR of the ZY dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Water0.95460.95480.95670.97280.97560.96990.93670.81740.7897
Building0.96720.9680.97690.98520.9940.9660.99710.99690.9971
Naked0.83730.83320.80050.67520.53530.30940.28810.19580.1892
Vegetation0.90060.89980.84230.81640.97920.53210.49710.44340.4377
Road0.81310.79920.76360.640.14540.53210.0890.03460.0277
Overall accuracy0.92070.91970.90160.87750.810810.731340.70960.6450.6345
Table 4. Classification accuracy of first-order to eighth-order DMSC of the ZY dataset.
Table 4. Classification accuracy of first-order to eighth-order DMSC of the ZY dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Water0.95460.82610.97580.97640.98530.97090.96830.96160.9628
Building0.967210.85590.94920.96290.96420.96830.96830.9666
Naked0.837300.77790.82950.83340.83420.93390.83210.8359
Vegetation0.900600.80810.85370.87770.88820.89750.90030.8961
Road0.813100.24230.68550.75070.80020.80610.81010.8051
Overall accuracy0.92070.50150.85140.90620.92040.92050.92320.92780.9213
Table 5. Classification accuracy of first- to eighth-order DMSR of the PT dataset.
Table 5. Classification accuracy of first- to eighth-order DMSR of the PT dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Leymus chinensis0.76840.76730.76510.76350.76170.76160.76070.76070.7594
Astragalus obliquus0.29550.31890.28420.27270.20420.19150.11240.11800.0362
Thalictrum0.10390.13850.10400.09940.09000.09800.06790.07410.0275
Pulsatilla chinensis0.28660.28910.24850.27110.22690.21180.18790.18330.1584
Plantain0.32490.39860.35080.37470.34400.34950.28960.29600.2160
Overall accuracy0.6325 0.640.62430.61280.61140.61130.61180.61150.6112
Table 6. Classification accuracy of first- to eighth-order DMSC of the PT dataset.
Table 6. Classification accuracy of first- to eighth-order DMSC of the PT dataset.
Class NameRawFirst OrderSecond OrderThird OrderFourth OrderFifth OrderSixth OrderSeventh OrderEighth Order
Leymus chinensis0.76840.7575 0.7571 0.7581 0.7599 0.7624 0.7633 0.7644 0.7661
Astragalus obliquus0.295500.0764 0.0759 0.1808 0.2158 0.2985 0.2962 0.2992
Thalictrum0.1039000.0000 0.0020 0.0254 0.0850 0.1114 0.1000
Pulsatilla chinensis0.2866000.0560 0.1434 0.2036 0.2718 0.2709 0.2743
Plantain0.3249000.0000 0.0640 0.1910 0.3143 0.3303 0.3192
Overall accuracy0.6325 0.6096 0.6097 0.6109 0.6173 0.6247 0.6321 0.6343 0.6360
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Zhang, A.; Sun, Y.; Wang, J.; Pang, H.; Peng, J.; Chen, Y.; Zhang, J.; Giannico, V.; Legesse, T.G.; et al. Deep Multi-Order Spatial–Spectral Residual Feature Extractor for Weak Information Mining in Remote Sensing Imagery. Remote Sens. 2024, 16, 1957. https://doi.org/10.3390/rs16111957

AMA Style

Zhang X, Zhang A, Sun Y, Wang J, Pang H, Peng J, Chen Y, Zhang J, Giannico V, Legesse TG, et al. Deep Multi-Order Spatial–Spectral Residual Feature Extractor for Weak Information Mining in Remote Sensing Imagery. Remote Sensing. 2024; 16(11):1957. https://doi.org/10.3390/rs16111957

Chicago/Turabian Style

Zhang, Xizhen, Aiwu Zhang, Yuan Sun, Juan Wang, Haiyang Pang, Jinbang Peng, Yunsheng Chen, Jiaxin Zhang, Vincenzo Giannico, Tsegaye Gemechu Legesse, and et al. 2024. "Deep Multi-Order Spatial–Spectral Residual Feature Extractor for Weak Information Mining in Remote Sensing Imagery" Remote Sensing 16, no. 11: 1957. https://doi.org/10.3390/rs16111957

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop