Next Article in Journal
Customized Tracking Algorithm for Robust Cattle Detection and Tracking in Occlusion Environments
Previous Article in Journal
A Millimeter-Wave Broadband Multi-Mode Substrate-Integrated Gap Waveguide Traveling-Wave Antenna with Orbit Angular Momentum
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time-Aware Dual LSTM Neural Network with Similarity Graph Learning for Remote Sensing Service Recommendation

School of Computer and Control Engineering, Yantai University, Yantai 264005, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1185; https://doi.org/10.3390/s24041185
Submission received: 6 January 2024 / Revised: 4 February 2024 / Accepted: 9 February 2024 / Published: 11 February 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
Technological progress has led to significant advancements in Earth observation and satellite systems. However, some services associated with remote sensing face issues related to timeliness and relevance, which affect the application of remote sensing resources in various fields and disciplines. The challenge now is to help end-users make precise decisions and recommendations for relevant resources that meet the demands of their specific domains from the vast array of remote sensing resources available. In this study, we propose a remote sensing resource service recommendation model that incorporates a time-aware dual LSTM neural network with similarity graph learning. We further use the stream push technology to enhance the model. We first construct interaction history behavior sequences based on users’ resource search history. Then, we establish a category similarity relationship graph structure based on the cosine similarity matrix between remote sensing resource categories. Next, we use LSTM to represent historical sequences and Graph Convolutional Networks (GCN) to represent graph structures. We construct similarity relationship sequences by combining historical sequences to explore exact similarity relationships using LSTM. We embed user IDs to model users’ unique characteristics. By implementing three modeling approaches, we can achieve precise recommendations for remote sensing services. Finally, we conduct experiments to evaluate our methods using three datasets, and the experimental results show that our method outperforms the state-of-the-art algorithms.

1. Introduction

Recently, remote sensing technologies have made remarkable progress. Constant innovations in this field have led to advancements in sensor technology, multi-modal data processing, and geographic information analysis. Comprehending the Earth’s surface and its transformations has driven research and has allowed global challenges to be tackled in areas such as Earth science, environmental science, and meteorology. Remote sensing technology is a method of observing the Earth’s surface from a distance. It allows scientists to gather extensive data with high resolutions, both spatially and temporally. This technology is widely used in the military sector and national security for tasks like border monitoring, reconnaissance, and surveillance of hazardous areas. It provides an abundance of experimental data and intelligence support.
The demand for remote sensing data has increased significantly in various fields. Researchers from different domains strongly support the high precision and extensive coverage of remote sensing data, which significantly enhance the effectiveness of research methodologies. As a result, a substantial volume of remote sensing resources has been generated. However, users find it challenging to fully comprehend and locate the resources they need within the system. Even if they are aware of the remote sensing resources available, they still struggle to find exactly what they need for a specific task.
The remote sensing service systems are designed to integrate information from various sources. Each type of remote sensing resource in the system is specialized, requiring users to have a considerable level of expertise in their respective knowledge domains. The system is open and provides resources to users free of charge. However, due to the complexity and diversity of resources in the system, users find it challenging to explore the relevant resources accurately. Even with substantial experience, users struggle to pinpoint their specific needs, leading to an impractical search process.
Most current remote sensing information service systems rely on keyword search and subscription modes to provide users with the required resources [1]. In the search mode, users can enter specific query conditions and conduct searches accordingly. The system will return relevant resources in terms of the search conditions. In the subscription mode, users can submit their requirements to the service as resource orders. This mode can proactively push recommendations to users when data that meet their requirements become available. While these two modes precisely cater to user needs and generate accurate recommendations based on specific requirements, they also present particular challenges. The user must be very familiar with the domain knowledge. This can make it challenging for users to get started and potentially decrease their satisfaction. It can be challenging for experienced users to package their requirements effectively. These modes primarily rely on a simple resource filtering approach, limiting users’ potential to explore other resources and broaden their expertise. Additionally, the modes obscure the similarity between remote sensing resources, making it difficult for users to discover resources similar to their needs.
There is a growing preference among current users for a system that generates comprehensive and timely recommendations by taking into account their historical needs. Researchers have proposed various solutions to address these issues. Analysis of user preferences has resulted in precise recommendations for remote sensing resources, thereby enhancing the user experience. In the context of remote sensing service recommendations, this refers to the application of recommendation systems in remote sensing resources. Machine learning models or deep learning models, among others, are utilized to analyze the historical behavioral patterns of users when using remote sensing resource information service systems to determine users’ behavioral preferences more accurately. Subsequently, this model was deployed as a web service on a server to achieve recommendations for remote sensing resources.
Several researchers have tackled information filtering issues in remote sensing information service systems using conventional collaborative filtering (CF) techniques [2,3,4,5,6,7]. However, traditional CF methods can only reveal superficial linear relationships and cannot capture complex nonlinear relationships due to their inherent structure. Furthermore, this approach overlooks the contextual information present in the system, failing to explore other information on recommendation performance.
As a result, some researchers have been exploring the significance of contextual information and its role in service recommendations [8,9,10,11,12,13]. By incorporating more contextual information, researchers can gain a better understanding of the impact of user interactions on recommendation outcomes. However, it has been observed that researchers often fail to consider the effect of nonlinear relationships on recommendation performance when integrating contextual information. The emergence of deep learning has enabled researchers to explore complex and nonlinear relationships [14,15,16,17,18,19]. This has led to significant improvements in recommendation performance, particularly in remote sensing information service recommendations and other similar domains.
However, while deep learning can explore intricate nonlinear relationships, it may not be optimal for data represented as graph structures. To address this, some special neural networks have been invented, such as Graph Neural Networks (GNNs) and others, which are designed to handle graph-structured data. By perceiving the positional structural information of nodes within a graph and comprehending relationships between nodes through information propagation mechanisms, GNNs can enable more sophisticated exploration. Researchers have therefore incorporated Graph Neural Networks into service recommendations to achieve better recommendation outcomes [20,21,22,23]. Although service recommendation methods have achieved good results, they still have some shortcomings. For instance, some approaches overlook the importance of initializing node embeddings in graph-structured data, which can mask critical information during aggregation through Graph Neural Networks, such as the similarity between nodes. Additionally, certain methods fail to consider the temporal dynamics of users’ historical behaviors when making service recommendations, thus neglecting the influence of time sequences on users. Finally, some methods do not take into account users’ latent factors, which can prevent the provision of a flexible vector encoding to learn the unique properties of these latent influences.
In this study, we propose using a Time-aware Dual LSTM Neural Network with Similarity Graph Learning to address specific issues related to the remote sensing service recommendation (TGDL-RSSR). The method comprises three parts, which are designed to identify users’ unique latent influences, the temporal dynamics of their historical behaviors, and the similarity and graph structure characteristics among different remote sensing information categories. First, we organize user interactions in a time-ordered sequence based on their past behavior in the remote sensing service system. A relationship graph is created based on the connections between different remote-sensing resource categories. Then, we use collaborative filtering to construct a cosine similarity matrix that measures the similarity between category relationships. Second, we utilize an embedding layer to represent the historical interaction sequences of users. We use LSTM to capture the temporal dynamics and establish relationships between users’ next search results and their historical behavior. We use a Multi-Layer Perceptron (MLP) layer based on the cosine similarity matrix to create embedded representations of remote sensing information that share similar characteristics. We utilize GCN to model the graph structure data and this embedding, allowing us to capture the similarity and adjacency relationships among categories. We use a second LSTM to model the temporal dynamics among similarities by combining the user’s historical sequence. Using an embedding layer, we identify the user’s unique ID. Finally, after obtaining three embeddings, we concatenate them and use an MLP to predict the remote sensing resource category for users.
The main contributions of this study are summarized as follows:
  • We propose a method that utilizes LSTM to capture the dynamic relationships between search items and construct a sequence of user search history behavior.
  • We present a method that utilizes CF to create a similarity matrix between remote sensing resource categories and then uses an MLP to generate embedded representations based on the matrix.
  • We utilize GCN to model both the graph structure and similarity embeddings by establishing an adjacency graph relationship among categories.
  • We utilize LSTM to capture the temporal dynamical similarity between categories by combining historical behavior sequences with GCN modeling results.

2. Related Work

Remote sensing is a scientific and technological discipline that uses sensors installed in satellites to collect the surface information of the Earth. Amidst the ongoing advancements in satellite technology and radar sensors, remote sensing has made remarkable progress, resulting in better data acquisition rates, accuracy, and resolution of Earth’s surface resources. The deployment of high-resolution, multispectral, and synthetic aperture radar satellites has significantly improved remote sensing in areas like land cover, resource monitoring, and environmental protection [24].
The use of Light Detection and Ranging (LiDAR) technology is becoming increasingly common in areas like digital terrain modeling, urban planning, and forestry. This technology provides highly accurate three-dimensional surface information [25]. Remote sensing technology plays a crucial role in monitoring climate change, assessing natural disasters, and generating useful data for environmental protection and climate research [26]. The integration of optical sensors and remote sensing technology has further expanded applications in environmental monitoring, agriculture, and urban planning. For detailed information on the remote sensing methods, refer to Table 1. Integrating remote sensing data with graphical information systems enables applications in urban planning, resource management, and geographic information science [27].
Remote sensing data have various characteristics, including diversity, multiple sources, temporal variability [28,29,30], and an increasing spatial resolution due to the advancement in sensing technologies [31]. These extensive high-resolution datasets provide more rich geographical information, allowing for a more thorough exploration of geographic features. This, in turn, facilitates comprehensive studies of the Earth system and the discovery of more intricate relationships [32].
Remote sensing data are widely applied in various fields, such as environmental monitoring and protection, urban planning and management, and water resources management. As remote sensing technology has developed rapidly, the amount of data collected has become increasingly massive, and the relationships within the data have become more complex, making it more challenging to analyze and research the data effectively. To address this challenge, remote sensing active service technology has emerged. This technology enables the analysis and exploration of vast remote sensing data, uncovering inherent patterns and subsequently generating a supportive service model [1,33].
Remote sensing technology has the potential to assist researchers in finding the right remote sensing data. However, traditional active service approaches rely on catalog searches and resource downloads, which require researchers to have a high level of proficiency in stating their needs. This approach may limit users’ perspectives and fail to suggest similar resources, making it hard to discover new remote sensing services. The timely updating of remote sensing data makes it challenging to generate accurate recommendations based on the timeliness of remote sensing resources and the users’ behavior [34]. Poor recommendations can significantly affect the user experience and the efficient utilization of remote sensing resources. For traditional active service techniques, in addition to catalog searching, other methods include surveys, interviews, archival research, and on-site investigations. These methods primarily conduct searches based on the analysis and positioning of user-defined needs. Similar to catalog searching, they face a common challenge of being unable to generate more precise recommendations for users based on their historical behavior. Detailed information is provided in Table 2.
Researchers have proposed solutions to address the limitations of traditional remote sensing active service techniques. One such solution is the FIR method, which was introduced by Lu et al. [35]. This method recommends and ranks remote sensing images based on users’ specific areas of interest by employing two special features. These features reflect correlations between users’ interests and remote-sensing images and serve as indicators of the relationship between them. Fuzzy association rule mining is then carried out based on these features to identify the relationship between the features and user interests. Additionally, two fuzzy inference strategies have been introduced to make recommendations in terms of the discovered association rules.
Hong et al. introduced a cluster-based index structure that is specifically designed for managing a large number of remote sensing images [36]. They proposed two indicators to measure the scalability and centrality between users’ area interests and remote-sensing resources.
Zhang et al. provided a method called the Spatio-Temporal Periodic Task model (STPT) [37] for recommending remote sensing data. This approach uses a probabilistic latent topic model to represent user retrieval behavior as a mixture of latent tasks. Two distributions are introduced to capture the relationships between tasks and spatial, temporal, and image features. Finally, an inference algorithm is employed to achieve the remote sensing service recommendation.
Li et al. proposed a technique to handle adaptive remote sensing recommendations based on behavioral analysis [38]. This method involves the real-time collection of user activities on the platform to discover their historical and dynamic preferences. Subsequently, it integrates this information using a decay method, ultimately achieving high-performance recommendations.
Song et al.provided a method to recommend remote-sensing resources using CF [39]. This method introduces a noise reduction technique in the similarity calculation process to determine the similarity between remote sensing data. The technique stabilizes the rating curve, which improves the recommendation performance.
Chu et al. introduced a new personalized remote-sensing images recommendation framework [40]. Their  method employs a knowledge graph to model the relationships between different entities. Using this information, this framework describes the connections between remote sensing images and users. The approach uses a Multi-attribute Fusion-based CF Network that utilizes the nonlinear computational capabilities of deep learning to calculate scores for each candidate image. This enables the recommendation of remote-sensing images that are tailored to individual users.
Li et al. provided a solution to the challenges of the rapid increase in remote sensing data [41]. In their method, they first utilized YOLOv3 object detection to extract the position distribution vectors of targets in remote sensing images as content information. Secondly, they constructed a multi-element user interest profile, which was dynamically adjusted according to the user’s active search records to enhance the recommendation performance. Finally, they achieved accurate and intelligent recommendations by matching image content, attributes, and user profile models.
Wang et al. proposed a novel Multi-modal Knowledge Graph-aware Deep Graph Attention Network (MMKDGAT) constructed based on graph convolutional networks [42]. This approach uses remote sensing resource images to create a multi-modal knowledge graph and incorporates additional information. Subsequently, a deep Graph Attention Network (GAT) is employed to mine information from this graph structure, aiming to capture information within various multi-modal nodes better and achieve more accurate recommendations.
Some of the current methods for analyzing remote sensing incorporate diversities and temporal characteristics. Some of these methods use knowledge graphs to incorporate this information. However, these approaches do not consider the temporal regularities or time dynamics of user behavior. As a result, they do not provide a comprehensive analysis of users’ long and short-term interests, which is crucial for identifying user preferences and behavior patterns, uncovering behavior patterns, understanding interest trends, and providing dynamic recommendations of remote sensing data that change over time. Although some research considers the similarity between remote sensing resources, the computation of this similarity often remains limited to linear operations like CF. This neglects deeper information about the similarity and the graph structure relationships among remote sensing resources.
Double LSTM refers to combining two LSTM models to explore sequential data further; each LSTM has distinct functionalities. Many researchers have applied double LSTMs in various fields. For example, Long et al. used double LSTMs to predict the trajectories of surrounding vehicles [43]. In this method, the driver’s historical sequential trajectories are input into the first LSTM to identify the driver’s intentions. Then, the output of the first LSTM, along with the driver’s historical geographic location sequence, is input into the second LSTM to predict future trajectories. Combining these two LSTMs allows a deeper exploration of the driver’s historical behavior, leading to more accurate predictions. Shi et al. proposed using double LSTMs to predict the Remaining Useful Life (RUL) of sensors [44]. The two LSTMs are responsible for monitoring changes in sensor data and predicting RUL, respectively. This combination allows for better representation of long- and short-term dependencies within the sensor data, resulting in more accurate predictions. Although these methods differ from ours in application domains, they combine two LSTMs to unearth higher-order information hidden in users’ historical behavior sequences.
To address the issues summarized, we propose a framework called Time-aware Dual LSTM Neural Network with Similarity Graph Learning for recommending remote sensing services (TGDL-RSSR). This framework constructs a historical behavior sequence based on the user’s interactions with remote sensing sources. It then uses LSTM to explore the user’s long- and short-term interests. Subsequently, CF is used to create a similarity matrix by considering the co-occurrence of remote-sensing resource categories. The MLP’s nonlinear capability is utilized to capture the relationships between similarities, achieving a category representation with deep-level similarities. Using obtained category representations and GCN, we simultaneously build an interaction graph structure to explore the adjacency relationships between remote sensing resource categories. Utilizing users’ historical behavior sequences, updated representations, and LSTM, we capture the potential temporal dynamics of these categories. Furthermore, we use the unique identifier ID of the user to explore their potential interests. This helps us to implement a remote sensing resource recommendation model. With this approach, the remote sensing resource service platform can analyze the temporal interests of users, similarities among remote sensing resources, and more.

3. Proposed Method

Within this section, we first formulate the problem and explain the relevant symbols used in this study. Then, we provide an overall introduction to our approach. Subsequently, we introduce each part of the model, which includes the input layer, representation layer, connection layer, multilayer perceptron layer, output layer, and optimization function.

3.1. Problem Formulation

Remote sensing resource service systems typically consist of various components, such as human–computer interaction, a resource portal, a category directory, resource services, and databases [1]. The system’s workflow is illustrated in Figure 1, and it primarily functions by filtering resources using users’ search keywords, generating relevant results, and presenting them to users. However, this keyword-based search approach is not perfect, since it requires users to encapsulate their needs accurately. Users’ professional knowledge and ability affect the accuracy of the search conditions. If users fail to provide accurate search keywords, this can result in inaccurate search conditions and, consequently, incorrect search results.
As a result of this precise query approach, the system often filters resource categories similarly to users’ search keywords, leading to overly narrow search results. This search method does not analyze users’ past behavior, so the presented results may not be representative, requiring users to sift through a large number of search results to find what they need. This process significantly diminishes the user experience. Moreover, filtering out similar remote sensing resource categories limits users’ possibilities to explore other similar categories and narrows their perspective, which may not be ideal in some cases.

3.2. Definitions

Definition 1
(User and Remote sensing resources). U and V represent the set of users and remote sensing resources, respectively. u i is used to denote a user, and  v j is used to denote a remote sensing resources, where u i U , v j V .
Definition 2
(User unique identification ID). Let I u i represent the ID of a specific user.
Definition 3
(users’ historical behavior sequence). Let { v t 1 u i , v t 2 u i , v t 3 u i , . . . , v t m u i } represent the users’ historical behavior sequence. In the sequence, t represents time, t i denotes the i-th time, and  v t i u i signifies the interaction between a user and a resource at time t i . Here, t i < t i + 1 , i < m 1 , m is the length of the sequence.
Definition 4
(Similarity matrix of remote sensing resource categories). Let A s represent the similarity matrix among remote sensing resource categories. Where A s R | V | × | V | , | V | represents the total number of remote sensing resource categories, A v i v j s represents the similarity between category v i and category v j .
Definition 5
(Adjacency graph structure of remote sensing resource categories). Let A I represent the adjacency graph structure among remote sensing resource categories, where A I R | V | × | V | , A v i v j I represents the adjacency relationship between category v i and category v j .
The meanings of important notations used in this paper are shown in Table 3.

3.3. Overall Framework

We propose a novel neural network framework to address current issues in remote sensing information service systems with time awareness and similarity graph learning, which is illustrated in Figure 2. This framework includes five modules: the input layer, representation layer, connection layer, MLP layer, and output layer. The core part of this model is the representation layer, which comprises three main modules.
The first module captures the historical behavior sequences of users when searching for remote sensing resources. It analyzes the long- and short-term interests in historical patterns for each user, which is crucial for accurately predicting their future needs.
The second module discovers the similarity and adjacency relationships among remote sensing resource categories in the system. We model the cosine similarity matrix through CF and MLP to obtain feature representations containing similarity between categories. We construct an interaction graph based on the adjacency relationships between categories and use GCN to model features and graph structures, thereby uncovering the similarity relationships among categories. We then combine the features aggregated by users’ historical behavior sequences and GCN, using LSTM to model them and extract the temporal regularities of category similarity. This is important for recommending resources similar to the user’s current needs.
The third module uses an embedding layer to model the user’s unique identification ID. This is used to uncover the user’s potential and unique factors when searching for resources, as each user has a distinctive behavior consciousness.

3.4. Input Layer

The input layer plays a crucial role in defining the model’s inputs. As the model’s core part is divided into three modules, the input layer is also categorized into three types.
The first type is the search history behavior sequence of users. This sequence is denoted as H 1 = { v t 1 U , v t 2 U , v t 3 U , , v t m U } and represents the user’s search history behavior. Here, H 1 R L × M , and L is the sequence length, and M represents the number of user search interactions. For this sequence, the elements are arranged chronologically, with each element representing the user’s historical search behavior at a specific moment. In other words, each element corresponds to the category of remote sensing resources the user searched for at that particular time.
The second type is the set of all categories to which remote sensing resources belong in the system. Denoted as H 2 = { v 1 , v 2 , v 3 , , v n } , where H 2 R | V | and n = | V | . Each category is assigned a numerical identifier for extracting similarity matrices and constructing interaction graphs.
The third type of input is the set of all user IDs with a certain amount of historical interactions in the system. Denoted as H 3 = { I u 1 , I u 2 , I u 3 , , I u k } , where H 3 R | U | and k = | U | .

3.5. Representation Layer

The representation layer is responsible for obtaining critical feature representations required by the model. This layer comprises three primary modules: (1) Long Short-Term Interest Representation of Users’ Historical Behavior Sequences, (2) Time-dynamic Representation of Similarity Relationships among Remote Sensing Resource Categories, and (3) Representation of Potential Influencing Factors for User Unique Identification ID. Each representation method will be explained in detail below.

3.5.1. Long Short-Term Interest Representation of Users’ Historical Behavior Sequences

The input for this module corresponds to the first type of input, which is denoted as H 1 in Section 3.4. In this type of input, each element in the sequence is essentially an index representation of a remote sensing resource category, with an additional characteristic of time. To process this input, we first flatten all user behavior sequences, resulting in a flattened shape of H 1 R L × M . Subsequently, we apply an embedding layer to this flattened representation, which is represented by the following formula:
H 1 = f l a t t e n ( H 1 )
e H 1 = e m b e d d i n g ( H 1 )
The f l a t t e n function is often used in deep learning to reshape H 1 into a one-dimensional vector of length L × M . Another useful tool is the e m b e d d i n g function, which maps discrete values to flat trainable vector representations. In this case, e H 1 R L × M , d represents the embedded representation of H 1 . d represents the dimension of the mapped vector after embedding.
We transform e H 1 to R L × M × d by reversing its size. This allows us to investigate the temporal patterns of users’ historical behavior, both short-term and long-term. Next, we feed e H 1 into the LSTM. Let us take a subsequence e H 1 l from e H 1 as an example, where l < L . The LSTM calculation process is described by the following formulas.
i t = σ ( W i · [ h t 1 , e H 1 , t l ] + b i )
f t = σ ( W f · [ h t 1 , e H 1 , t l ] + b f )
C t ˜ = t a n h ( W C · [ h t 1 , e H 1 , t l ] + b C )
o t = σ ( W o · [ h t 1 , e H 1 , t l ] + b o )
C t = f t C t 1 + i t C t ˜
h t = o t t a n h ( C t )
Formula (3) represents the input gate, Formulas (4) and (5) represent the forget gate, Formula (6) corresponds to the output gate, Formula (7) represents the long-term memory in the LSTM, and Formula (8) represents the short-term memory. The input layer of LSTM primarily receives a sequence of data. Each time step in the LSTM unit corresponds to an element in the sequence. Subsequently, LSTM inputs the results from the input layer to the input gate and forget gate. For the input gate, it calculates the output by constructing a weight matrix and applying an activation function, as shown in Formula (3). This gate is mainly responsible for controlling the amount of new information entering the new state. The forget gate operates similarly to the input gate, using trainable weight matrices and an activation function to compute outputs, as shown in Formulas (4) and (5). This gate is primarily responsible for controlling the information from the previous state that is forgotten. LSTM then utilizes Formula (7) to merge the results obtained from Formulas (3)–(5). Subsequently, the output gate, represented by Formula (6), controls how much information from the updated state will flow into the state at the next time step. Finally, the results obtained from the output gate are input into Formula (8) to calculate the hidden state, which represents the information flowing from each time step to the next. The specific process is illustrated in Figure 3.
The variable e H 1 l denotes the historical interaction behavior of the l-th sequence in e H 1 at time t. Additionally, h t 1 denotes the hidden state; i t denotes the output of the input gate; f t and C t ˜ represent the output of the forget gate; o t is the output of the output gate; { W i , W f , W C , W o } are the trainable parameters for each gate; { b i , b f , b C , b o } are the biases of each gate; σ represents the activation function.
The LSTM algorithm processes the results obtained from these gates through a specific formula to derive a hidden state denoted as h t . This hidden state contains both long- and short-term memory information. At this point, e H 1 , t l at time t has been transformed into h t . These operations are performed at each time point, allowing the model to extract the temporal dynamics of a user’s historical behavior sequence, thereby uncovering their long and short-term interests. Finally, through LSTM, e H 1 is transformed into e H 1 , where e H 1 R L × d . Refer to Figure 4 for a visual illustration of the specific process and Algorithm 1 for a detailed algorithmic flow.
Algorithm 1 Long short-term interest representation of users’ historical behavior sequences
Require: 
the users’ historical behavior sequence H 1
Ensure: 
History behavior embedding e H 1 with user long and short-term interests.
 1:
Obtain the flattened users’ historical behavior sequence H 1 through Formula (1)
 2:
Obtain the embedding matrix e H 1 through Formula (2)
 3:
Initialize the parameters { W i , W f , W C , W o , b i , b t , b C , b o } of the LSTM
 4:
for  l = 1 t o L  do
 5:
    for  t = 1 t o M  do
 6:
        Input e H 1 , t l
 7:
        Obtain h t according to Formulas (3)–(8)
 8:
    end for
 9:
end for
10:
Obtain an embedding matrix e H 1 with long and short-term interests of users
11:
return e H 1

3.5.2. Time-Dynamic Representation of Similarity Relationships among Remote Sensing Resource Categories

This module requires two inputs: the categories to which remote sensing resources belong ( H 2 ) and the historical behavior sequence ( H 1 ). It should be noted that each remote sensing resource may belong to multiple categories, and the co-occurrence frequency of categories may vary among all remote sensing resources, which means that the similarity between categories may differ. To address this, we construct a matrix A t i m e s with the co-occurrence frequency of any two categories as its elements. Next, we use CF to obtain the cosine similarity between categories according to this matrix. The resulting similarity matrix is denoted as A s . The calculation process is shown in the following formula:
A v i v j s = c o s i n e _ s i m i l a r i t y ( ( A t i m e s ) T ) = ( A t i m e s ) i T · ( A t i m e s ) j T ( A t i m e s ) i T · ( A t i m e s ) j T
The notation A v i v j s refers to the similarity between categories v i and v j . The symbol ( A t i m e s ) T represents the transpose of the matrix A t i m e s . ( A t i m e s ) i T is the i-th column of the matrix and  ( A t i m e s ) j T is the j-th column. Here, i and j can represent any pair of categories.
We obtain the similarity matrix A s of categories through the process mentioned above. To further utilize the significance of this similarity, we use an MLP in combination with this matrix for initializing the embedding of remote sensing resource categories. This embedding method controls the initial distribution of feature vectors, making sure that they are not randomized but rather carry initial similarity information. The calculation process for this is shown in the following formulas.
e A s ( 1 ) = σ ( W s ( 1 ) A s + b s ( 1 ) )
e A s ( l ) = σ ( W s ( l ) ( A s ) ( l 1 ) + b s ( l ) )
e A s = σ ( W s ( L ) ( A s ) ( L 1 ) + b s ( L ) )
Formula (10) represents the input layer of the MLP, Formula (11) describes the hidden layers, and Formula (12) describes the output layer. Here, l is the layer number, and L denotes the total number of layers, where 1 < l < L . The trainable network weights in the input, hidden, and output layer are denoted by { W s ( 1 ) , W s ( l ) , W s ( L ) } , respectively. The network biases in each layer are represented by { b s ( 1 ) , b s ( l ) , b s ( L ) } . Additionally, the outputs in each layer are represented by { e A s ( 1 ) , e A s ( l ) , e A s } , where e A s represents the category feature representation with similarity information after transformation. The activation function, represented by σ , is the same across all layers. The specific operation is illustrated in Figure 5.
We complete the initialization of embeddings for remote sensing resource categories by following the steps mentioned above. To construct a category interaction graph structure G s , we utilize the co-occurrence relationship among categories. We take the categories to which remote sensing resources belong as nodes. If two categories of remote sensing resources co-occur, we consider that there is an adjacency relationship between these two categories, meaning there is an edge relationship. By analysing each type of remote sensing resource, we can obtain the edge relationships between all categories, thereby constructing the interaction graph G s . We also use GCN to propagate messages and aggregate these embeddings and graph structures. This helps us obtain category embeddings e A s with similarity relationships. The computational process is depicted in the following formulas.
e A s ( l + 1 ) = σ ( ( D 1 2 ^ A ^ D 1 2 ^ ) G s e A s ( l ) W G s ( l ) )
A G s ^ = A G s + I G s
( D G s ) i i ^ = j = 0 | V | ( A G s ) i j ^
Here, l represents the layer of GCN where 0 < l < L . e A s is equivalent to e A s ( L ) when l + 1 = L . A G s ^ represents the graph structure. I G s is an identity matrix. D G s ^ is a diagonal matrix representing the node degrees. D 1 2 ^ normalizes A G s ^ . ( D G s ) i i ^ is the degree of the i-th node, obtained by summing the i-th row of D G s ^ or the i-th column of A G s ^ . α represents the activation function, and to this end, we use the ReLU.
We obtain category embeddings e A s with similarity relationships through the above operations. In order to capture the temporal dynamics of these similarities, we use the historical behavior sequence H 1 to form a sequence H 1 s with similarity relationships. Then, we use LSTM to encode the temporal characteristics of the sequence H 1 s , which is demonstrated in the following formulas.
e H 1 s = L S T M ( H 1 s )
In Section 3.5.1, we introduce the detailed calculation process of LSTM. Therefore, in this section, we use LSTM to represent it. The embedded representation with temporal dynamics of similarity relationships captured by LSTM is denoted by e H 1 s . It is important to note that since these two LSTMs are distinct, their network parameters are also different. The algorithmic process for this part is shown in Algorithm 2.

3.5.3. Representation of Potential Influencing Factors for Unique User Identification ID

This module takes H 3 as input. We embed user ID sequences as the initialized representation, which is denoted as e H 3 R L × d . The calculation formula is shown as follows:
e H 3 = e m b e d d i n g ( H 3 )
These three modules provide us with embeddings that capture the user’s temporal interests e H 1 R L × d , time-dynamic similarity relationships e H 1 s R L × d , and latent information about the user’s unique identification e H 3 R L × d .

3.6. Concatenation Layer

This layer connects the embeddings from the three modules to create the final combined embedding representation, e f , where e f R L × 3 d . The connection operation is formulated as follows:
e f = c o n c a t ( e H 1 , e H 1 s , e H 3 ) = e H 1 | | e H 1 s | | e H 3
where c o n c a t ( ) denotes the concatenation function. This function concatenates e H 1 , e H 1 s , and  e H 3 along the last dimension.
Algorithm 2 Time-dynamic representation of similarity relationships among remote sensing resource categories
Require: 
the users’ historical behavior sequence H 1 ; the remote sensing resource category  H 2
Ensure: 
Embedding e H 1 s with Time-Dynamic Similarity Graph Structure Relationships.
 1:
Generating the co-occurrence rate matrix A t i m e s based on H 2
 2:
Obtain the embedding matrix e H 1 through Formula (2)
 3:
Initialize the similarity matrix A s
 4:
for  i = 0 t o | V |  do
 5:
    for  j = 0 t o | V |  do
 6:
        if  i = j  then
 7:
            A v i v j s = 0
 8:
        else
 9:
            A v i v j s = c o s i n e _ s i m i l a r i t y ( ( A t i m e s ) T )
10:
        end if
11:
    end for
12:
end for
13:
Initialize network parameters { W s ( 1 ) , W s ( l ) , W s ( L ) , b s ( 1 ) , b s ( l ) , b s ( L ) }
14:
Feed A s to the MLP
15:
Obtain the embedding e A s for the similarity matrix according to Formulas (10)–(12)
16:
Construct the adjacency graph structure G s based on H 2
17:
Initialize network parameters W G s ( l )
18:
e A s and G s into GCN
19:
Obtain e A s according to Formulas (13)–(15)
20:
Obtain H 1 s based on H 1 and e A s
21:
Obtain e H 1 s based on Formula (16)
22:
return e H 1 s

3.7. Multilayer Perceptron Layer

The MLP layer is used to convert e f into scores for all categories of remote sensing resources. The computation process is shown in the following formulas.
( e f ) ( 1 ) = σ ( W f ( 1 ) e f + b f ( 1 ) )
( e f ) ( l ) = σ ( W f ( l ) e f ( l 1 ) + b f ( l ) )
e f = σ ( W f ( L ) e f ( L 1 ) + b f ( L ) )
The current layer is denoted as l and the previous layer is denoted as l 1 . The last layer is denoted as L and the second-to-last is denoted as L 1 . The output of these layers is denoted as { ( e f ) ( 1 ) , ( e f ) ( l ) , e f } , where e f R L × | V | and e is the final output. The trainable network parameters in each layer are represented by { W f ( 1 ) , W f ( l ) , W f ( L ) } , with W f ( 1 ) R 3 d × h ( 1 ) , W f ( l ) R h ( l 1 ) × h ( l ) , and W f ( L ) R h ( l ) × d . The network biases in each layer are denoted by { b f ( 1 ) , b f ( l ) , b f ( L ) } , with b f ( 1 ) R h ( 1 ) × 1 , b f ( l ) R h ( l ) × 1 , and b f ( L ) R d × 1 .

3.8. Output Layer

The output layer is a crucial component that converts the final result, denoted by e f , obtained from the MLP layer into the probability of the embedding belonging to each remote sensing resource category using the Softmax layer. This layer, along with the loss in the optimization function, facilitates the model update. For instance, the calculation process of the i-th element in ( e f ) i is demonstrated in the following formulas.
( e f ) i = S o f t m a x ( ( e f ) i ) = E ( e f ) i j = 0 | V | ( e f ) i j
After applying the Softmax transformation, ( e f ) i R 1 × | V | represents a probability matrix, while e f R L × | V | represents the probability matrix corresponding to e f . Here, ( e f ) i j R 1 is the j-th element in the i-th row.

3.9. Optimization

After obtaining the probability matrix e f , we use the cross-entropy loss to optimize our model. The formula for this loss is as follows.
l o s s = C r o s s E n t r o p y L o s s ( e f , t f ) = 1 L j = 0 L i = 0 | V | ( t f ) j i l o g ( e f ) j i
where t f represents the result in the real dataset, t f R L × 1 .

3.10. Deployment of the Recommendation Algorithm

In the previous sections, we have implemented personalized recommendations for remote sensing resources. By analyzing users’ temporal interests, identifying the similarity between remote sensing resource categories, and considering users’ unique potential factors, we have not only improved the recommendation performance but also moved away from the traditional filtering-based push of remote sensing resource service systems. We have designed a remote sensing resource streaming push service based on the TGDL-RSSR algorithm. This service includes three main modules. Figure 6 describes the specific process.
The first module obtains the ID of terminal system users, their search history, remote sensing resource categories, and relationships between categories. It then processes and analyzes the data. The second module deploys our TGDL-RSSR recommendation algorithm. The third module of our system involves a streaming push service that follows the traditional B/S architecture. The system includes two parts—the application side and the service side. On the application side, the current user’s ID is obtained and the content is displayed based on the service end’s response. The user’s new behavior is then stored in the database. On the service side, users’ recent history records and the classification information of remote sensing resources are retrieved from the database. The data are then analyzed and processed based on the model’s data format requirements. The TGDL-RSSR recommendation algorithm is deployed on the service end to generate recommendations based on the retrieved and processed data and user ID information. Finally, the results are sent to the application end through the streaming push service.
As the number of users increases, single-threaded tasks may struggle to keep up with the demands of multiple users. This can result in blocking issues, causing longer wait times and decreasing the efficiency of recommendations, ultimately impacting the user experience. To overcome this problem, we have implemented multi-threading support on the server side. When multiple users send requests simultaneously, threads can intelligently allocate these requests, with each thread independently handling a subset of requests. This significantly reduces the occurrence of blocking issues and ensures a smoother experience for all users. The visual representation of the multi-threaded scenario is shown in Figure 7.

4. Experiments

In this section, we verify the effectiveness of TSDL-RSSR by comparing it with other methods. The comparison is divided into two parts: (1) experiments involving three recommender system datasets; and (2) experiments involving a remote sensing service dataset. We compare our approach with several existing methods such as CF [45] and NeuralCF [46], LSTM [47], AGCN [48], and DCF [49].

4.1. Experimental Setting

4.1.1. Experimental Environment

We use the PyTorch framework [50] to implement our proposed method. The running environment is configured with a 12th generation Intel(R) Core(TM) i7-12700H 2300 MHz CPU and 32 GB of RAM.

4.1.2. Public Recommendation Datasets

We use two publicly available recommender system datasets, MovieLens [51], Amazon-clothes [52] and Amazon-books [53], to validate the effectiveness of TGDL-RSSR. Furthermore, we validate the applicability of TGDL-RSSR by utilizing a dataset of remote sensing services that we created. Table 4 provides detailed descriptions of both datasets.
The MovieLens dataset is a popular movie rating dataset that contains four fields: user ID, movie ID, rating, and timestamp. It includes 6040 users, 3706 movies, and 1,000,208 interactions. Our approach treats the movies in the dataset as a type of remote sensing resource, with the dataset representing the interactions between users and these resources. The interactions are represented as ratings that reflect users’ evaluations of the remote sensing resources. To implement our method, we needed to process the dataset in specific ways. For the adjacency graph structure, we considered two remote sensing resources to have an adjacency relationship if the same user accessed them. Regarding the similarity matrix, we collected all remote sensing resources visited by each user, treating this situation as a co-occurrence, and aggregated all users’ situations.
The Amazon-clothes dataset is a collection of clothing ratings that includes four fields: user ID, clothing ID, rating, and timestamp. We have treated clothing as a type of remote sensing resource category and processed this dataset in a similar way to the MovieLens dataset. The Amazon-clothes dataset has 4993 users, 39 different types of clothing, and a total of 6201 interactions.
The Amazon Books dataset is a book rating dataset with multiple fields, including user ID, book ID, review ID, review text, review timestamp, and more. For our method, we will exclude fields other than user ID, book ID, rating, and review timestamp. Additionally, we will convert the review timestamp to a timestamp format. Similar to the previous two datasets, we treat books as a category of remote sensing resources, and the specific processing method is the same as before. This dataset comprises 47,400 users, 36,412 books, and 154,555 interactions.

4.1.3. Remote Sensing Service Dataset

We have created a remote sensing service dataset by collecting the historical search records, user ID information, and type of remote sensing resources used by our users. The dataset is divided into two subsets—one that stores users’ historical behaviors and the other that stores the category information of remote sensing resources. When users search for remote sensing resources, the categories to which each resource belongs are displayed. Each resource may belong to multiple categories, and any combination of these categories may co-occur multiple times, resulting in higher similarity between them. The dataset comprises 300 users and 50 categories of remote sensing resources. Figure 8 provides a detailed illustration of the specific composition of this dataset.
In Figure 8, there are two sub-datasets. The first sub-dataset has four fields: user_id, category_id, timestamp, and score. These fields represent the user ID, the index representation of the category to which the resource belongs, the rating for the remote sensing resource category, and the timestamp of the user’s historical behavior, respectively. The second dataset also has two fields: rs_id and category. These fields represent the ID of the remote-sensing resource and all categories to which the remote-sensing resource belongs.

4.1.4. Data Preprocessing

To partition the relationship dataset, we employ leave-one-out to split the dataset into training and testing sets. The training set is used for model training, while the testing set is used to predict and recommend user preferences after training the model, as well as to calculate the corresponding recommendation performance. To enhance the model’s ability to analyze user preferences during the dataset processing, we remove users with interaction history sequences involving remote sensing resources less than 10 times. This ensures that users used for model training have rich interaction histories.

4.1.5. Parameter Settings

To standardize the embedding dimensions for the user and remote sensing service categories, we designated them as 50, aligning with the dimensions employed in comparable baseline methods. The MLP hidden layer incorporates three layers, consistent with the setup of other baseline methods. In the model architecture, the GCN consists of two layers, with an embedding dimension set at 50. The GCN layers for the AGCN method are configured identically. The learning rate for the two public datasets is set at 0.001, while for the remote sensing resource dataset, it is established at 0.005. As for other baseline methods, their learning rates fall within the range of 0.001, 0.005, and 0.01.
For model training, we utilize the Adam optimizer [54] for optimization. The iteration count is set to 100 for the MovieLens dataset, 70 for the Amazon-clothes dataset, 30 for the Amazon-books dataset, and 50 for the remote sensing resource dataset.

4.1.6. Evaluation Metrics

We use three metrics to evaluate different methods, including recall@k, precision@k, and F1-score@k, represented by the following formulas.
R e c a l l @ k = 1 N u N S u ( k ) T u T u
P r e c i s i o n @ k = 1 N u N S u ( k ) T u k
F 1 @ k = 2 R e c a l l @ k P r e c i s i o n @ k R e c a l l @ k + P r e c i s i o n @ k
S u ( k ) denotes the top k categories with the highest scores in the candidate set, and T u represents the resource categories in the test set. N denotes the total number of users.

4.2. Experiments on Recommender System Datasets

The section focuses on comparing the performance of the TGDL-RSSR model with other baseline models on recommended system datasets by adjusting hyperparameters and performing ablation experiments.

4.2.1. Overall Comparison with Baseline Methods

Table 5 shows the results of the recommendation performance compared with the baseline models, from which we can draw the following conclusions.
Table 5 shows that our method outperforms other models on both the MovieLens, Amazon-clothes and Amazon-books datasets.
Traditional collaborative filtering models rely on shallow linear relationships between features and targets, which means they often overlook other important information. Consequently, their recommendation performance on the MovieLens, Amazon-clothes and Amazon-books datasets is not satisfactory. From the perspective of remote sensing resources, there is a high probability of recommending incorrect categories to users. On the other hand, the NCF model, which builds on CF, uses deep learning for optimization. By taking advantage of the powerful nonlinear capabilities of deep learning, NCF explores complex relationships between features and targets, resulting in a better recommendation performance than CF models. However, despite incorporating deep learning, NCF still lacks the ability to include other crucial information.
DCF uses deep learning to make improvements based on CF. It is similar to NCF in terms of incorporating the nonlinear computational capability of deep learning. However, DCF differs from NCF in its feature initialization method. In NCF, an embedding layer is used to initialize the IDs of users and items, whereas DCF uses MLP for user and item embedding initialization based on the rating matrix. The different initialization methods between NCF and DCF can lead to variations in the fit to the target rating matrix and consequently differences in recommendation performance.
LSTM has a unique long short-term memory function that helps it to mine the temporal interests of users effectively. It can identify hidden dynamic temporal factors within input time series information. By analyzing historical behavior sequences, the model is able to uncover users’ latent interests and provide recommendations for their next interest. Since LSTM is a form of deep learning, it possesses powerful nonlinear computational capabilities. As a result, it outperforms CF and NCF in terms of recommendation performance.
AGCN is an extension of GCN that incorporates an attention layer. It uses graph structure information to extract information for each node in the graph. This allows for efficient node propagation, making it easy to perceive adjacency relationships between nodes. Unlike GCN, which propagates only to neighbors equally, AGCN considers the impact level of each neighbor on that node, making it more practically significant. AGCN is a powerful nonlinear deep learning technique, similar to LSTM, and has demonstrated excellent performance in recommendation systems.
Our approach, TGDL-RSSR, not only considers the similarity of users’ behavior patterns but also considers the temporal dynamics of these patterns. We achieve this by incorporating the mining of similarity adjacency relationships and the temporal dynamics of relationships between item categories, using a graph structure and LSTM. Additionally, we model each user’s unique potential influencing factors using their unique identification ID to capture the users’ individual characteristics. As a result, our approach outperforms other models on the MovieLens, Amazon-clothes and Amazon-books datasets. Comparing our findings with the best results of baseline models, we can find that on the MovieLens, recall@5, precision@5, and F1-score@5 are higher by 2.76%, 2.96%, and 2.76%, respectively. Recall@10, precision@10, and F1-score@10 are higher by 4.63%, 4.47%, and 4.7%, respectively. Recall@15, precision@15, and F1-score@15 are higher by 4.08%, 3.94%, and 4.21%, respectively. Recall@20, precision@20, and F1-score@20 are higher by 13.82%, 13.68%, and 13.81%, respectively. On the Amazon-clothes, recall@5, precision@5, and F1-score@5 are higher by 10.65%, 10.41%, and 10.58%, respectively. Recall@10, precision@10, and F1-score@10 are higher by 10.08%, 10.31%, and 10.2%, respectively. Recall@15, precision@15, and F1-score@15 are higher by 6.44%, 6.56%, and 6.71%, respectively. Recall@20, precision@20, and F1-score@20 are higher by 5.88%, 6.36%, and 6.06%, respectively. Similarly, on the Amazon-books, recall@5, precision@5, and F1-score@5 are higher by 5.68%, 5.69%, and 5.76%, respectively. Recall@10, precision@10, and F1-score@10 are higher by 12.10%, 12.26%, and 12.00%, respectively. Recall@15, precision@15, and F1-score@15 are higher by 6.93%, 6.94%, and 6.96%, respectively. Recall@20, precision@20, and F1-score@20 are higher by 5.16%, 4.29%, and 4.51%, respectively.
A more intuitive representation is shown in Figure 9, Figure 10 and Figure 11.

4.2.2. Sensitivity Analysis

This section explores the impact of different learning rates on recommendation performance, as well as the impact of different embedding sizes.
The effect of learning rate. We tested different values of the learning rate, including 0.001, 0.005, 0.01, 0.05, and 0.1. The impact results on the MovieLens and Amazon-clothes datasets are shown in Figure 12 and Figure 13, respectively. We can see that the model’s recall, precision, and F1-score perform optimally when the learning rate is 0.001. Moreover, compared to other learning rates, the model’s recommendation performance remains relatively stable when the learning rate is at 0.001. However, when set at 0.1, the model’s performance on both datasets fluctuates significantly.
The effect of embedding dimension. This section aims to demonstrate how embedding dimensions can affect the recommendation performance. We experimented with embedding dimensions of 30, 40, 50, and 60 and evaluated the model’s performance on the MovieLens and Amazon-clothes datasets, as shown in Figure 14 and Figure 15, respectively. The results indicate that the model performs best when the dimension is 50, as it achieves optimal recall, precision, and f1-score values. This suggests that bigger embedding dimensions might not lead to better performance as long as the appropriate value is selected to capture the necessary information.

4.2.3. Ablation Study

We conduct ablation experiments by selectively removing three key modules from the core layer. Figure 16a demonstrates the impact of each module on recall@k on the MovieLens, while Figure 16b illustrates the impact on the Amazon-clothes. In Figure 16, TGDL-RSSR(1), TGDL-RSSR(2), TGDL-RSSR(3), and TGDL-RSSR represent models with the first module (users’ historical sequence), the second module (resource category similarity graph structure), the third module (user unique identification ID), and the original model being removed, respectively.
For the metrics of recall@5, recall@10, recall@15, and recall@20, the results presented in Figure 16a demonstrate that TGDL-RSSR outperforms TGDL-RSSR(1) by 107.3%, 45.89%, 29.18%, and 6.13% respectively. TGDL-RSSR also surpasses TGDL-RSSR(2) by 59.69%, 13.94%, 8.95%, and 22.70% respectively. Furthermore, TGDL-RSSR exceeds TGDL-RSSR(3) by 0.32%, 10.75%, 4.46%, and 17.21% respectively. In Figure 16b, TGDL-RSSR surpasses TGDL-RSSR(1) by 32.26%, 14.14%, 3.28%, and 3.75% respectively. TGDL-RSSR also outperforms TGDL-RSSR(2) by 10.96%, 4.82%, 3.01%, and 7.09% respectively. Finally, TGDL-RSSR exceeds TGDL-RSSR(3) by 10.19%, 2.83%, 0.35%, and 4.75% respectively. These results indicate the importance of leveraging LSTM to explore the temporal dynamics to improve the recommendation performance. Furthermore, the CF+GCN+LSTM pattern is effective in exploring the temporal dynamics of similarity adjacency relationships between resources in the second module. Finally, modeling user-unique identification IDs using an embedding layer is effective in uncovering unique user characteristics in the third module.
In comparing the performance of TGDL-RSSR with TGDL-RSSR(1), we find that in scenarios where users have a more extensive history, TGDL-RSSR can more accurately analyze users’ long and short-term interests, thereby achieving more precise recommendations. Similarly, in comparison with TGDL-RSSR(2), if users’ search histories cover a diverse range of remote sensing resource categories, i.e., a higher co-occurrence frequency among resource categories, TGDL-RSSR’s analysis of the relationships and similarities between remote sensing resource categories will be clearer, leading to more accurate recommendations.

4.2.4. Computational Efficiency

Within this section, we explored the training and prediction time of TGDL-RSSR on datasets of different scales. For training time, we divided the dataset into proportions of 20%, 40%, 60%, 80%, and 100% to explore the training time of the model at different ratios. Each dataset followed the same procedure. For prediction time, we used the test set to obtain the model’s prediction time. The details are shown in Figure 17.
As shown in Figure 17a, when the computational resources (dataset) size changes, the model’s training time also increases. When the data scale is large, the computational time is significantly affected due to the higher complexity of relationships in the dataset, and vice versa. Additionally, even when maintaining the computational resources at the same scale, each dataset’s complexity differs, resulting in different training times at the same proportion. Although the model takes a relatively long time to reach the convergence point at a certain scale, the prediction time is relatively short (100% data proportion). In the MovieLens dataset, the model’s prediction time is 556.312 ms; in the Amazon-clothes dataset, it is 348.76 ms; in the Amazon-books dataset, it is 751.475 ms, meaning the model’s prediction time is less than 1 s for each dataset. This indicates that TGDL-RSSR performs well in terms of prediction time, even on large-scale datasets. Therefore, when deployed in a remote sensing resource information service system, the system can provide acceptable response times to users.

4.3. Experiments on Remote Sensing Service Dataset

This section conducts experiments on a real remote sensing resource dataset we created, to validate the effectiveness of our approach in comparison to traditional methods used in remote sensing service systems.

4.3.1. Overall Comparison with Baseline Methods

We compared the recommendation performance of TGDL-RSSR with other methods using the collected actual remote sensing resource dataset. The comparative results are described in Table 6. A more intuitive representation is shown in Figure 18.
In Section 4.2.1, we have analyzed the structural differences between different models. Here, we are going to compare the performance of TGDL-RSSR with the best-performing baseline model. Although TGDL-RSSR did not achieve the highest values for recall@5, precision@5, and F1-score@5 on the remote sensing resource dataset, it still outperformed the baseline model in recall@10, precision@10, and F1-score@10 by 12.07%, 12.02%, and 12.03% respectively. Moreover, it surpassed the baseline model in recall@15, precision@15, and f1-score@15 by 6.27%, 6.25%, and 6.18%, respectively. In recall@20, precision@20, and f1-score@20, our method exceeded the baseline model by 16.34%, 16.59%, and 16.24%, respectively. This comparison demonstrates the effectiveness of our approach.

4.3.2. Usability Experiments

In this section, we explore the practicality of TGDL-RSSR by discussing the deployment process of the model, users’ historical behavior collection, and remote sensing resource category recommendations.
Deployment process of the model. First, let us discuss integrating the model into the system. In Figure 6, we describe deploying the remote sensing resource recommendation model as a web service on the server and returning its recommendations to users via stream push. Traditional remote sensing resource information service systems rely on users searching through keywords, with the server directly querying the database to display search results. For a recommendation model, it adds a recommendation model service between user-related information and the database, as illustrated in Figure 19.
Users’ historical behavior collection. Then, we demonstrate the collection of historical behaviors for some users. We take the examples of two users, u_43 and u_112 (43 and 112 are their unique identification IDs). For u_43, we have collected a three-month history of remote sensing resource searches, while for u_112, we have collected a seven-month history of such searches. Since our recommendation method requires a user’s historical behavior, we have collected the behavior of these two users. During the collection process, we used the user’s most recent search behavior as the reference and collected nine historical behaviors in reverse chronological order, summing up to a total of eight historical behaviors. The collected results are shown in Figure 20.
We present relevant image samples, ID numbers, and their corresponding categories for each dataset of remote sensing resources in Figure 20. We use an index table to remap and index each remote sensing resource category. In the user box, we have illustrated the data structure used for recommendation, which includes the user ID and their historical behavior sequence. The historical behavior sequence for User 43 is {5, 3, 13, 5, 5, 5, 31, 13}, and for User 112, it is {14, 4, 18, 18, 14, 14, 4, 18}, where the values represent the indexed mapping of remote sensing resource categories.
Recommendation of remote sensing resource categories. We showcase the recommendation results of our method based on the past behavior of two users, u_43 and u_112. Instead of recommending a single category, our method recommends a sequence of categories. The significance of similarity analysis in our method is validated by the categories in the sequence. You can find the recommendation results in Figure 21 and Figure 22. The remote sensing resources depicted in Figure 21 and Figure 22 are sourced from the literature [55,56,57,58,59,60,61,62,63,64,65,66].
For the recommendation method, we use binary stream push to return the model’s recommended results. For remote sensing resources, they are mainly composed of images. On the server side, if traditional image transmission methods are used, it may cause network congestion during the push, thereby increasing the duration of the push and affecting both the performance of the service system and the user experience. Stream push technology essentially involves processing the results that the server needs to return into binary streams using server-side languages. This approach can significantly reduce the server-side load and the size of transmitted data, thereby avoiding network congestion caused by large data volumes.
For example, let us consider u_43. Based on the historical behavior sequence, it is likely that the next area of interest for this user is remote sensing resource category 5, which is named “Classification” in the index. Figure 21 shows that our remote sensing resource recommendation method successfully suggested relevant resources in this category. The recommended resources show that the “Classification” category is often associated with other categories like “Semantic segmentation” and “Multisensor data fusion”, indicating a high degree of similarity between these categories. The “similar datasets” section in Figure 21 demonstrates that our recommendation method effectively suggests remote sensing resources based on these similar categories. This method enables users to explore other similar categories of remote sensing resources, thereby enhancing their understanding of remote sensing resources.

4.3.3. User Satisfaction Comparison

We compare TGDL-RSSR with traditional content-based retrieval methods [67] used in remote sensing resource service systems. We specifically look at the approach presented in [1]. We measure user satisfaction levels across three categories: zero, one, and two. A score of zero represents dissatisfaction, one indicates moderate satisfaction, and two signifies complete satisfaction. Then, we also collect statistics on the average usage time of users in the system. A shorter average usage time suggests that users can quickly find the resources they want when using the system. Conversely, a higher average usage time indicates that the system may not conduct an in-depth analysis of user preferences, leading users to spend more time articulating their needs and searching for resources. Finally, we conducted statistics on the system latency time for both methods. A shorter latency indicates that the method has a quicker response time, enhancing the user experience. The comparison results are presented in Table 7.
According to Table 7, our remote sensing resource recommendation method has proven to be more effective in terms of user satisfaction than traditional content-based retrieval methods. At the same time, this also indicates that by analyzing the historical behavior sequences generated by users when searching for remote sensing data, personalized analysis of users’ long and short-term interests can be conducted, thereby achieving more precise recommendations. This reduces the time users spend analyzing their needs and searching each time they use the system. By analyzing the similarity between resource categories when users search for resources, it is possible to explore resources similar to user preferences. While ensuring user usage needs are met, users can also become acquainted with similar resources. Moreover, after deploying the TGDL-RSSR method into the remote sensing resource service system, there has been a reduction in response time, allowing for more real-time effects on user requests. This ultimately enhances user satisfaction.

5. Conclusions and Future Work

Our study introduces TGDL-RSSR, a method for recommending remote sensing information services. It utilizes a dual-LSTM network with time awareness and similarity graph learning. The method consists of five layers, with the representation layer being the core layer, which is divided into three main modules. In the first module, we employ LSTM to analyze the long- and short-term interests as well as the temporal dynamics in users’ historical behavior sequences. The second module uses a CF+MLP+GCN+LSTM approach to explore the similarity graph structure adjacency relationships among remote sensing resource categories and the temporal dynamics of these relationships. In the third module, we model the unique identification ID of users using an embedding layer to uncover users’ unique potential characteristics. Finally, an MLP is employed to obtain the probability that the user’s next interest belongs to each category, thereby completing the remote sensing resources recommendations.
During the data preprocessing stage, we imposed limitations on the interaction frequency in users’ historical behavior, potentially leading to a significant decline in recommendation effectiveness when the user interaction history is sparse. The cold start phenomenon can result in an extremely sparse user interaction history. Therefore, cold start data pose a challenge, and represent an issue to be addressed in our future work. In our approach, we employ a dual LSTM. LSTM is prone to problems such as gradient vanishing and exploding during training, making it less stable when capturing temporal interests in users’ historical behavior sequences. Additionally, LSTM relies on sequential computation, which may impact the training efficiency of the model to some extent. To tackle these issues, we plan to replace LSTM with Transformer [68]. The Transformer model, besides addressing some of the drawbacks associated with LSTM, incorporates inherent positional encoding and multi-head attention. This enables it to better differentiate information from different positions in users’ historical behavior sequences and focus on various elements in the sequence, thereby achieving a more precise exploration of user preferences.

Author Contributions

Conceptualization, J.Z. and W.M.; methodology, J.Z. and W.M.; software, J.Z.; validation, E.Z.; formal analysis, W.M.; investigation, X.X.; resources, J.Z. and E.Z.; data curation, J.Z. and E.Z.; writing—original draft preparation, J.Z.; writing—review and editing, W.M.; visualization, X.X.; supervision, W.M.; project administration, J.Z. and W.M.; funding acquisition, W.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Nature Science Foundation of China (No. 61602399), Shandong Provincial Nature Science Foundation, China (ZR2020MF100), and Youth Innovation Science and Technology Support Program of Shandong Provincial under Grant 2021KJ080.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhu, L.; Wu, F.; Fu, K.; Hu, Y.; Wang, Y.; Tian, X.; Huang, K. An Active Service Recommendation Model for Multi-Source Remote Sensing Information Using Fusion of Attention and Multi-Perspective. Remote Sens. 2023, 15, 2564. [Google Scholar] [CrossRef]
  2. Jiang, Y.; Liu, J.; Tang, M.; Liu, X. An effective web service recommendation method based on personalized collaborative filtering. In Proceedings of the 2011 IEEE International Conference on Web Services, Washington, DC, USA, 4–9 July 2011; pp. 211–218. [Google Scholar]
  3. Sun, H.; Zheng, Z.; Chen, J.; Lyu, M.R. Personalized web service recommendation via normal recovery collaborative filtering. IEEE Trans. Serv. Comput. 2012, 6, 573–579. [Google Scholar] [CrossRef]
  4. Yao, L.; Sheng, Q.Z.; Ngu, A.H.; Yu, J.; Segev, A. Unified collaborative and content-based web service recommendation. IEEE Trans. Serv. Comput. 2014, 8, 453–466. [Google Scholar] [CrossRef]
  5. Liu, J.; Tang, M.; Zheng, Z.; Liu, X.; Lyu, S. Location-aware and personalized collaborative filtering for web service recommendation. IEEE Trans. Serv. Comput. 2015, 9, 686–699. [Google Scholar] [CrossRef]
  6. Zou, G.; Jiang, M.; Niu, S.; Wu, H.; Pang, S.; Gan, Y. QoS-aware Web service recommendation with reinforced collaborative filtering. In Proceedings of the Service-Oriented Computing: 16th International Conference, ICSOC 2018, Hangzhou, China, 12–15 November 2018; Proceedings 16. Springer: Berlin/Heidelberg, Germany, 2018; pp. 430–445. [Google Scholar]
  7. Peng, Q.; You, L.; Dong, N. A location-aware GIServices quality prediction model via collaborative filtering. Int. J. Digit. Earth 2018, 11, 897–912. [Google Scholar] [CrossRef]
  8. Kuang, L.; Xia, Y.; Mao, Y. Personalized services recommendation based on context-aware QoS prediction. In Proceedings of the 2012 IEEE 19th International Conference on Web Services, Honolulu, HI, USA, 24–29 June 2012; pp. 400–406. [Google Scholar]
  9. Sangers, J.; Frasincar, F.; Hogenboom, F.; Chepegin, V. Semantic web service discovery using natural language processing techniques. Expert Syst. Appl. 2013, 40, 4660–4671. [Google Scholar] [CrossRef]
  10. Xu, Y.; Yin, J.; Deng, S.; Xiong, N.N.; Huang, J. Context-aware QoS prediction for web service recommendation and selection. Expert Syst. Appl. 2016, 53, 75–86. [Google Scholar] [CrossRef]
  11. Fan, X.; Hu, Y.; Zheng, Z.; Wang, Y.; Brézillon, P.; Chen, W. CASR-TSE: Context-aware web services recommendation for modeling weighted temporal-spatial effectiveness. IEEE Trans. Serv. Comput. 2017, 14, 58–70. [Google Scholar] [CrossRef]
  12. Gao, H.; Xu, Y.; Yin, Y.; Zhang, W.; Li, R.; Wang, X. Context-aware QoS prediction with neural collaborative filtering for Internet-of-Things services. IEEE Internet Things J. 2019, 7, 4532–4542. [Google Scholar] [CrossRef]
  13. Zheng, Z.; Li, X.; Tang, M.; Xie, F.; Lyu, M.R. Web service QoS prediction via collaborative filtering: A survey. IEEE Trans. Serv. Comput. 2020, 15, 2455–2472. [Google Scholar] [CrossRef]
  14. Labbaci, H.; Medjahed, B.; Binzagr, F.; Aklouf, Y. A deep learning approach for web service interactions. In Proceedings of the International Conference on Web Intelligence, Leipzig, Germany, 23–26 August 2017; pp. 848–854. [Google Scholar]
  15. Bai, B.; Fan, Y.; Tan, W.; Zhang, J. DLTSR: A deep learning framework for recommendations of long-tail web services. IEEE Trans. Serv. Comput. 2017, 13, 73–85. [Google Scholar] [CrossRef]
  16. Xiong, R.; Wang, J.; Zhang, N.; Ma, Y. Deep hybrid collaborative filtering for web service recommendation. Expert Syst. Appl. 2018, 110, 191–205. [Google Scholar] [CrossRef]
  17. Yang, Y.; Qamar, N.; Liu, P.; Grolinger, K.; Wang, W.; Li, Z.; Liao, Z. Servenet: A deep neural network for web services classification. In Proceedings of the 2020 IEEE International Conference on Web Services (ICWS), Beijing, China, 7–11 July 2020; pp. 168–175. [Google Scholar]
  18. Smahi, M.I.; Hadjila, F.; Tibermacine, C.; Benamar, A. A deep learning approach for collaborative prediction of Web service QoS. Serv. Oriented Comput. Appl. 2021, 15, 5–20. [Google Scholar] [CrossRef]
  19. Mecheri, K.; Klai, S.; Souici-Meslati, L. Deep learning based web service recommendation methods: A survey. J. Intell. Fuzzy Syst. 2023, 44, 9879–9899. [Google Scholar] [CrossRef]
  20. Ren, X.; Zhang, W.; Bao, L.; Song, J.; Wang, S.; Cao, R.; Wang, X. DeepQSC: A GNN and attention mechanism-based framework for QoS-aware service composition. In Proceedings of the 2021 International Conference on Service Science (ICSS), Xi’an, China, 14–16 May 2021; pp. 76–83. [Google Scholar]
  21. Zhang, L.; Cao, B.; Peng, M.; Qing, Y.; Kang, G.; Liu, J.; Fletcher, K.K. Bilinear Graph Neural Network-Enhanced Web Services Classification. In Proceedings of the 2021 IEEE 23rd Int Conf on High Performance Computing & Communications; 7th Int Conf on Data Science & Systems; 19th Int Conf on Smart City; 7th Int Conf on Dependability in Sensor, Cloud & Big Data Systems & Application (HPCC/DSS/SmartCity/DependSys), Haikou, China, 20–22 December 2021; pp. 189–196. [Google Scholar]
  22. Jia, Z.; Fan, Y.; Zhang, J.; Wu, X.; Wei, C.; Yan, R. A Multi-Source Information Graph-Based Web Service Recommendation Framework for a Web Service Ecosystem. J. Web Eng. 2022, 21, 2287–2312. [Google Scholar] [CrossRef]
  23. Cao, B.; Zhang, L.; Peng, M.; Qing, Y.; Kang, G.; Liu, J. Web Service Recommendation via Combining Bilinear Graph Representation and xDeepFM Quality Prediction. IEEE Trans. Netw. Serv. Manag. 2023, 20, 1078–1092. [Google Scholar] [CrossRef]
  24. Wulder, M.A.; Loveland, T.R.; Roy, D.P.; Crawford, C.J.; Masek, J.G.; Woodcock, C.E.; Allen, R.G.; Anderson, M.C.; Belward, A.S.; Cohen, W.B.; et al. Current status of Landsat program, science, and applications. Remote Sens. Environ. 2019, 225, 127–147. [Google Scholar] [CrossRef]
  25. Carson, W.W.; Andersen, H.E.; Reutebuch, S.E.; McGaughey, R.J. LIDAR applications in forestry–an overview. In Proceedings of the ASPRS Annual Conference, Denver, CO, USA, 23–28 May 2004; pp. 1–9. [Google Scholar]
  26. Weng, Q. Advances in Environmental Remote Sensing: Sensors, Algorithms, and Applications; Remote Sensing Applications Series; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  27. Yamazaki, F. Applications of remote sensing and GIS for damage assessment. Struct. Saf. Reliab. 2001, 1, 1–12. [Google Scholar]
  28. Jensen, J.R. Remote Sensing of the Environment: An Earth Resource Perspective 2/e; Pearson Education India: Bengaluru, India, 2009. [Google Scholar]
  29. Chen, J.; Chen, J.; Liao, A.; Cao, X.; Chen, L.; Chen, X.; He, C.; Han, G.; Peng, S.; Lu, M.; et al. Global land cover mapping at 30 m resolution: A POK-based operational approach. ISPRS J. Photogramm. Remote Sens. 2015, 103, 7–27. [Google Scholar] [CrossRef]
  30. Pettorelli, N.; Laurance, W.F.; O’Brien, T.G.; Wegmann, M.; Nagendra, H.; Turner, W. Satellite remote sensing for applied ecologists: Opportunities and challenges. J. Appl. Ecol. 2014, 51, 839–848. [Google Scholar] [CrossRef]
  31. Seto, K.C.; Reenberg, A.; Boone, C.G.; Fragkias, M.; Haase, D.; Langanke, T.; Marcotullio, P.; Munroe, D.K.; Olah, B.; Simon, D. Urban land teleconnections and sustainability. Proc. Natl. Acad. Sci. USA 2012, 109, 7687–7692. [Google Scholar] [CrossRef]
  32. Fu, D.; Xiao, H.; Su, F.; Zhou, C.; Dong, J.; Zeng, Y.; Yan, K.; Li, S.; Wu, J.; Wu, W.; et al. Remote sensing cloud computing platform development and Earth science application. Yaogan Xuebao/J. Remote Sens. 2021, 25, 220–230. [Google Scholar] [CrossRef]
  33. Chen, X.; Liu, Y.; Li, F.; Li, X.; Jia, X. Remote sensing image recommendation based on spatial–temporal embedding topic model. Comput. Geosci. 2021, 157, 104935. [Google Scholar] [CrossRef]
  34. Deren, L.; Liangpei, Z.; Guisong, X. Automatic analysis and mining of remote sensing big data. Acta Geod. Cartogr. Sin. 2014, 43, 1211. [Google Scholar]
  35. Lu, E.H.C.; Hong, J.H.; Su, Z.L.T.; Chen, C.H. A fuzzy data mining approach for remote sensing image recommendation. In Proceedings of the 2013 IEEE International Conference on Granular Computing (GrC), Beijing, China, 13–15 December 2013; pp. 213–218. [Google Scholar]
  36. Hong, J.H.; Su, Z.L.T.; Lu, E.H.C. A recommendation framework for remote sensing images by spatial relation analysis. J. Syst. Softw. 2014, 90, 151–166. [Google Scholar] [CrossRef]
  37. Zhang, X.; Chen, D.; Liu, J. A space-time periodic task model for recommendation of remote sensing images. ISPRS Int. J. Geo-Inf. 2018, 7, 40. [Google Scholar] [CrossRef]
  38. Li, C.; Chang, L.; Zhang, J.; Zhou, X.; Hu, Z.; Song, Y. Adaptive recommendation technology for remote sensing information based on behavior analysis. In Proceedings of the International Conference on Artificial Intelligence, Information Processing and Cloud Computing, Sanya, China, 19–21 December 2019; pp. 1–5. [Google Scholar]
  39. Song, Y.; Chang, L.; Song, Y.; Zhou, X.; Li, C.; Shi, X. Research on remote sensing information recommendation technology based on collaborative filtering. In Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering, Xi’an, China, 20 August 2021; pp. 297–301. [Google Scholar]
  40. Chu, B.; Chen, J.; Wang, M.; Gao, F.; Guo, Q.; Li, F. Remote Sensing Image Recommendation Using Multi-attribute Embedding and Fusion Collaborative Filtering Network. In Proceedings of the Artificial Intelligence: First CAAI International Conference, CICAI 2021, Hangzhou, China, 5–6 June 2021; Proceedings, Part I 1. Springer: Berlin/Heidelberg, Germany, 2021; pp. 65–76. [Google Scholar]
  41. Li, Y.; Hou, L.; Xue, J.; Lyu, K.; Wang, Y. Remote sensing image recommendation method based on content interpretation. J. Comput. Appl. 2023. Available online: http://www.joca.cn/CN/10.11772/j.issn.1001-9081.2023030313 (accessed on 5 January 2024).
  42. Wang, F.; Zhu, X.; Cheng, X.; Zhang, Y.; Li, Y. MMKDGAT: Multi-modal Knowledge graph-aware Deep Graph Attention Network for remote sensing image recommendation. Expert Syst. Appl. 2024, 235, 121278. [Google Scholar] [CrossRef]
  43. Xin, L.; Wang, P.; Chan, C.Y.; Chen, J.; Li, S.E.; Cheng, B. Intention-aware long horizon trajectory prediction of surrounding vehicles using dual LSTM networks. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 1441–1446. [Google Scholar]
  44. Shi, Z.; Chehade, A. A dual-LSTM framework combining change point detection and remaining useful life prediction. Reliab. Eng. Syst. Saf. 2021, 205, 107257. [Google Scholar] [CrossRef]
  45. Su, X.; Khoshgoftaar, T.M. A survey of collaborative filtering techniques. Adv. Artif. Intell. 2009, 2009, 421425. [Google Scholar] [CrossRef]
  46. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.S. Neural collaborative filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
  47. Memory, L.S.T. Long short-term memory. Neural Comput. 2010, 9, 1735–1780. [Google Scholar]
  48. Feng, C.; Liu, Z.; Lin, S.; Quek, T.Q. Attention-based graph convolutional network for recommendation system. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 7560–7564. [Google Scholar]
  49. Tegene, A.T.; Liu, Q.; Muhammed, S.B.; Leka, H.L. Deep Learning Based Matrix Factorization For Collaborative Filtering. In Proceedings of the 2021 18th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 17–19 December 2021; pp. 165–170. [Google Scholar]
  50. Imambi, S.; Prakash, K.B.; Kanagachidambaresan, G. PyTorch. In Programming with TensorFlow: Solution for Edge Computing Applications; Springer: Cham, Switzerland, 2021; pp. 87–104. [Google Scholar]
  51. Harper, F.M.; Konstan, J.A. The movielens datasets: History and context. Acm Trans. Interact. Intell. Syst. (tiis) 2015, 5, 1–19. [Google Scholar] [CrossRef]
  52. Zhang, H.; Huang, W.; Liu, L.; Xu, X. Clothes collocation recommendations by compatibility learning. In Proceedings of the 2018 IEEE International Conference on Web Services (ICWS), San Francisco, CA, USA, 2–7 July 2018; pp. 179–186. [Google Scholar]
  53. Srujan, K.; Nikhil, S.; Raghav Rao, H.; Karthik, K.; Harish, B.; Keerthi Kumar, H. Classification of amazon book reviews based on sentiment analysis. In Proceedings of the Information Systems Design and Intelligent Applications: Proceedings of Fourth International Conference INDIA 2017, Da Nang, Vietnam, 15–17 June 2017; Springer: Beach Road, Singapore, 2018; pp. 401–411. [Google Scholar]
  54. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  55. Schmitt, M.; Hughes, L.; Qiu, C.; Zhu, X. SEN12MS–A Curated Dataset of Georeferenced Multi-Spectral Sentinel-1/2 Imagery for Deep Learning and Data Fusion. arXiv 2019, arXiv:1906.07789. [Google Scholar] [CrossRef]
  56. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The synthia dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3234–3243. [Google Scholar]
  57. Ekim, B.; Schmitt, M. Mapinwild: A dataset for global wilderness mapping. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 231–234. [Google Scholar]
  58. Hu, J.; Liu, R.; Hong, D.; Camero, A.; Yao, J.; Schneider, M.; Kurz, F.; Segl, K.; Zhu, X.X. MDAS: A New Multimodal Benchmark Dataset for Remote Sensing. Earth Syst. Sci. Data 2023, 15, 113–131. [Google Scholar] [CrossRef]
  59. Di Gregorio, A. Land Cover Classification System: Classification Concepts and User Manual: LCCS; Food & Agriculture Org.: Rome, Italy, 2005; Volume 2. [Google Scholar]
  60. Shermeyer, J.; Hogan, D.; Brown, J.; Van Etten, A.; Weir, N.; Pacifici, F.; Hansch, R.; Bastidas, A.; Soenen, S.; Bacastow, T.; et al. SpaceNet 6: Multi-sensor all weather mapping dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 19 June 2020; pp. 196–197. [Google Scholar]
  61. Chen, L.; Liu, F.; Zhao, Y.; Wang, W.; Yuan, X.; Zhu, J. Valid: A comprehensive virtual aerial image dataset. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 2009–2016. [Google Scholar]
  62. Robicquet, A.; Sadeghian, A.; Alahi, A.; Savarese, S. Learning social etiquette: Human trajectory understanding in crowded scenes. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VIII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 549–565. [Google Scholar]
  63. Loboda, T.; Hall, J.; Baer, A. ABoVE: Wildfire Date of Burning within Fire Scars across Alaska and Canada, 2001–2019; ORNL DAAC: Oak Ridge, TN, USA, 2017. [Google Scholar]
  64. Tom, M.; Prabha, R.; Wu, T.; Baltsavias, E.; Leal-Taixé, L.; Schindler, K. Ice monitoring in Swiss lakes from optical satellites and webcams using machine learning. Remote Sens. 2020, 12, 3555. [Google Scholar] [CrossRef]
  65. Sykas, D.; Sdraka, M.; Zografakis, D.; Papoutsis, I. A Sentinel-2 multi-year, multi-country benchmark dataset for crop classification and segmentation with deep learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3323–3339. [Google Scholar] [CrossRef]
  66. Loiseau, R.; Aubry, M.; Landrieu, L. Online Segmentation of LiDAR Sequences: Dataset and Algorithm. In Proceedings of the ECCV, 2022, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  67. Wang, M.; Song, T. Remote sensing image retrieval by scene semantic matching. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2874–2886. [Google Scholar] [CrossRef]
  68. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention is All you Need. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
Figure 1. Typical remote sensing service system.
Figure 1. Typical remote sensing service system.
Sensors 24 01185 g001
Figure 2. The overall architecture of TGDL-RSSR.
Figure 2. The overall architecture of TGDL-RSSR.
Sensors 24 01185 g002
Figure 3. LSTM processing flow.
Figure 3. LSTM processing flow.
Sensors 24 01185 g003
Figure 4. Extracting long and short-term interests of users.
Figure 4. Extracting long and short-term interests of users.
Sensors 24 01185 g004
Figure 5. Similarity embedding matrix extraction.
Figure 5. Similarity embedding matrix extraction.
Sensors 24 01185 g005
Figure 6. Recommendation algorithm deployment process.
Figure 6. Recommendation algorithm deployment process.
Sensors 24 01185 g006
Figure 7. Multi-threading support in remote sensing resource service.
Figure 7. Multi-threading support in remote sensing resource service.
Sensors 24 01185 g007
Figure 8. The remote sensing service dataset.
Figure 8. The remote sensing service dataset.
Sensors 24 01185 g008
Figure 9. Performance comparison on MovieLens.
Figure 9. Performance comparison on MovieLens.
Sensors 24 01185 g009
Figure 10. Performance comparison on Amazon-clothes.
Figure 10. Performance comparison on Amazon-clothes.
Sensors 24 01185 g010
Figure 11. Performance comparison on Amazon-books.
Figure 11. Performance comparison on Amazon-books.
Sensors 24 01185 g011
Figure 12. Parameter tuning for learning rate on MovieLens.
Figure 12. Parameter tuning for learning rate on MovieLens.
Sensors 24 01185 g012
Figure 13. Parameter tuning for learning rate on Amazon-clothes.
Figure 13. Parameter tuning for learning rate on Amazon-clothes.
Sensors 24 01185 g013
Figure 14. Parameter tuning for embedding dimension on MovieLens.
Figure 14. Parameter tuning for embedding dimension on MovieLens.
Sensors 24 01185 g014
Figure 15. Parameter tuning for embedding dimension on Amazon-clothes.
Figure 15. Parameter tuning for embedding dimension on Amazon-clothes.
Sensors 24 01185 g015
Figure 16. Ablation study on MovieLens and Amazon-clothes.
Figure 16. Ablation study on MovieLens and Amazon-clothes.
Sensors 24 01185 g016
Figure 17. Computational efficiency on MovieLens, Amazon-clothes and Amazon-books. (a) Training time on MovieLens, Amazon-clothes and Amazon-books. (b) Test time on MovieLens, Amazon-clothes and Amazon-books.
Figure 17. Computational efficiency on MovieLens, Amazon-clothes and Amazon-books. (a) Training time on MovieLens, Amazon-clothes and Amazon-books. (b) Test time on MovieLens, Amazon-clothes and Amazon-books.
Sensors 24 01185 g017
Figure 18. Performance comparison on remote sensing service dataset.
Figure 18. Performance comparison on remote sensing service dataset.
Sensors 24 01185 g018
Figure 19. Deployment process of the model.
Figure 19. Deployment process of the model.
Sensors 24 01185 g019
Figure 20. User historical behavior collection.
Figure 20. User historical behavior collection.
Sensors 24 01185 g020
Figure 21. Recommendation results for u_43.
Figure 21. Recommendation results for u_43.
Sensors 24 01185 g021
Figure 22. Recommendation results for u_112.
Figure 22. Recommendation results for u_112.
Sensors 24 01185 g022
Table 1. Remote sensing methods.
Table 1. Remote sensing methods.
MethodsSensor TypeApplication Area
Optical sensorVisible and infrared spectrumEnvironmental monitoring, Agriculture, Urban planning
Radar sensorMicrowave radiationNatural disaster monitoring, Resource management
Thermal infrared sensorInfrared spectrumSurface temperature monitoring, Vegetation health assessment
Table 2. Traditional active service approaches.
Table 2. Traditional active service approaches.
Active Service MethodsDescriptionExamples
Catalog SearchingUtilizes library catalogs, databasesLibrary catalogs, academic databases
Surveys and InterviewsEngages in communication with domain experts or practitionersProfessionals, industry practitioners
Archival ResearchRetrieves historical records, documentsArchives, historical reports
On-site InvestigationsConducts site visits, observes and recordsField inspections, survey reports
Table 3. Notations used in this paper.
Table 3. Notations used in this paper.
NotationsDescriptions
u i / v j User/remote sensing resources
I u i ID of a specific user
v t i u i Interaction between the user and remote sensing resources
A s Similarity matrix among remote sensing resource categories
A I Adjacency relationship among remote sensing resource categories
{ v t 1 u i , v t 2 u i , v t 3 u i , , v t m u i }users’ historical behavior sequence
Table 4. Description of MovieLens, Amazon-clothes and Amazon-books datasets.
Table 4. Description of MovieLens, Amazon-clothes and Amazon-books datasets.
DatasetsMovieLensAmazon-ClothesAmazon-Books
Number of users6040499347,400
Number of remote sensing resources37063936,412
Data sparsity95.53%96.82%99.99%
Table 5. Performance comparison on MovieLens, Amazon-clothes and Amazon-books.
Table 5. Performance comparison on MovieLens, Amazon-clothes and Amazon-books.
DatasetsModelsk = 5k = 10k = 15k = 20
RecallPrecisionF1RecallPrecisionF1RecallPrecisionF1RecallPrecisionF1
MovieLensCF0.07550.01510.02520.13840.01380.02520.18240.01220.02280.23270.01160.0222
NCF0.09130.01820.03050.16330.01630.02970.22020.01470.02750.25930.01290.0247
LSTM0.09780.01960.03250.18890.01890.03430.25110.01670.03140.35780.01790.0341
AGCN0.14520.02900.04840.22530.02250.04100.27180.01810.03400.31950.01600.0304
DCF0.15230.03040.05080.24610.02460.04470.30410.02030.03800.38000.01900.0362
TGDL-RSSR0.15650.03130.05220.25750.02570.04680.31650.02110.03960.43250.02160.0412
Amazon-clothes%Improv.2.76%2.96%2.76%4.63%4.47%4.70%4.08%3.94%4.21%13.82%13.68%13.81%
CF0.04940.00990.01650.10150.01020.01850.14390.00960.01800.21860.01090.0208
NCF0.08410.01680.02810.13620.01360.02480.18050.01200.02260.24230.01210.0231
LSTM0.08310.01660.02770.15920.01590.02890.25040.01670.03130.34440.01720.0328
AGCN0.13240.02210.03780.18090.01810.03290.23620.01570.02950.29870.01490.0284
DCF0.10610.02120.03540.19440.01940.03530.27480.01830.03430.34700.01730.0330
TGDL-RSSR0.14650.02440.04180.21400.02140.03890.29250.01950.03660.36740.01840.0350
Amazon-books%Improv.10.65%10.41%10.58%10.08%10.31%10.20%6.44%6.56%6.71%5.88%6.36%6.06%
CF0.07440.01490.02480.13490.01350.02450.21860.01460.02730.29300.01470.0279
NCF0.07910.01580.02630.18230.01820.03320.25860.01720.03230.38050.01900.0362
LSTM0.11300.02260.03770.23000.02300.04180.32230.02150.04030.40930.01720.0390
AGCN0.14070.02810.04690.24880.02490.04520.33870.02260.04230.42430.02120.0404
DCF0.13540.02710.04520.26110.02610.04750.36780.02450.04600.46490.02330.0443
TGDL-RSSR0.14870.02970.04960.29270.02930.05320.39330.02620.04920.46730.02340.0445
%Improv.5.68%5.69%5.76%12.10%12.26%12.00%6.93%6.94%6.96%5.16%4.29%4.51%
Table 6. Performance comparison on remote sensing service dataset.
Table 6. Performance comparison on remote sensing service dataset.
DatasetsModelsk = 5k = 10k = 15k = 20
RecallPrecisionF1RecallPrecisionF1RecallPrecisionF1RecallPrecisionF1
Remote SensingCF0.07330.01470.02440.19330.01930.03520.26000.01730.03250.32000.01600.0305
NCF0.08680.01740.02890.18370.01840.03340.27800.01850.03480.37440.01870.0357
LSTM0.10620.02120.03540.19480.01950.03540.26410.01760.03300.42450.02120.0404
AGCN0.11170.02240.03720.20010.02000.03640.29630.01980.03700.39660.01980.0378
DCF0.12930.02590.04310.23290.02330.04240.33650.02240.04210.44610.02230.0425
TGDL-RSSR0.12790.02560.04260.26100.02610.04750.35760.02380.04470.51900.02600.0494
%Improv.−1.08%−1.16%−1.16%12.07%12.02%12.03%06.27%06.25%06.18%16.34%16.59%16.24%
Table 7. Comparison of remote sensing resource retrieval methods.
Table 7. Comparison of remote sensing resource retrieval methods.
Comparison ItemsActive Service Recommendation ModelContent-Based Retrieval Method
Service modeProactivePassive
User satisfaction20
System usage time4.5 min8.3 min
System latency time0.58 s1.13 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Ma, W.; Zhang, E.; Xia, X. Time-Aware Dual LSTM Neural Network with Similarity Graph Learning for Remote Sensing Service Recommendation. Sensors 2024, 24, 1185. https://doi.org/10.3390/s24041185

AMA Style

Zhang J, Ma W, Zhang E, Xia X. Time-Aware Dual LSTM Neural Network with Similarity Graph Learning for Remote Sensing Service Recommendation. Sensors. 2024; 24(4):1185. https://doi.org/10.3390/s24041185

Chicago/Turabian Style

Zhang, Jinkai, Wenming Ma, En Zhang, and Xuchen Xia. 2024. "Time-Aware Dual LSTM Neural Network with Similarity Graph Learning for Remote Sensing Service Recommendation" Sensors 24, no. 4: 1185. https://doi.org/10.3390/s24041185

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop