Next Article in Journal
An Automatic Method for Elbow Joint Recognition, Segmentation and Reconstruction
Next Article in Special Issue
Multi-Directional Long-Term Recurrent Convolutional Network for Road Situation Recognition
Previous Article in Journal
Identification and Compensation Method of Unbalanced Error in Driving Chain for Rate-Integrating Hemispherical Resonator Gyro
Previous Article in Special Issue
Accelerated Stochastic Variance Reduction Gradient Algorithms for Robust Subspace Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SGK-Net: A Novel Navigation Scene Graph Generation Network

School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(13), 4329; https://doi.org/10.3390/s24134329
Submission received: 24 April 2024 / Revised: 26 June 2024 / Accepted: 2 July 2024 / Published: 3 July 2024

Abstract

:
Scene graphs can enhance the understanding capability of intelligent ships in navigation scenes. However, the complex entity relationships and the presence of significant noise in contextual information within navigation scenes pose challenges for navigation scene graph generation (NSGG). To address these issues, this paper proposes a novel NSGG network named SGK-Net. This network comprises three innovative modules. The Semantic-Guided Multimodal Fusion (SGMF) module utilizes prior information on relationship semantics to fuse multimodal information and construct relationship features, thereby elucidating the relationships between entities and reducing semantic ambiguity caused by complex relationships. The Graph Structure Learning-based Structure Evolution (GSLSE) module, based on graph structure learning, reduces redundancy in relationship features and optimizes the computational complexity in subsequent contextual message passing. The Key Entity Message Passing (KEMP) module takes full advantage of contextual information to refine relationship features, thereby reducing noise interference from non-key nodes. Furthermore, this paper constructs the first Ship Navigation Scene Graph Simulation dataset, named SNSG-Sim, which provides a foundational dataset for the research on ship navigation SGG. Experimental results on the SNSG-sim dataset demonstrate that our method achieves an improvement of 8.31% (R@50) in the PredCls task and 7.94% (R@50) in the SGCls task compared to the baseline method, validating the effectiveness of our method in navigation scene graph generation.

1. Introduction

The development of deep learning-based perception techniques has significantly enhanced the ability of intelligent ships to perceive objects in a scene [1,2]. However, in order to improve the safety of autonomous ship navigation, it is necessary to go beyond surface-level perception and achieve a deeper understanding of the scene. Scene graphs have emerged as an effective way to enhance the scene understanding capability of intelligent ships. However, there are some problems in the current navigation scene image data, such as complex relationships between target entities and noise in the contextual information required for relationship inference. These issues hinder the application of scene graphs in navigation scene understanding. To address these challenges, this paper conducts research on navigation scene graph generation (NSGG).
The scene graph generation (SGG) task is related to scene perception tasks while also demonstrating significant differences. Perception tasks such as object detection [3,4,5], object recognition [6,7,8], and scene segmentation [9,10] primarily focus on acquiring information regarding the category and location of objects. On the other hand, the SGG task emphasizes capturing the relationships between objects within a scene and the interactions between objects and their surrounding environment. Perception tasks primarily rely on computer vision techniques, whereas the task of SGG combines computer vision with natural language processing, making it a higher-level and more complex task [11]. SGG represents a scene using a graph structure, with object detection bounding boxes serving as nodes and the relationships between objects as edges. The generation of scene graphs provides a foundation for subsequent scene understanding tasks and intelligent decision-making. Therefore, in this study, the focus is on generating navigation scene graphs, aiming to enhance the understanding capability of intelligent ships in navigation scenes.
Since the application of scene graphs in image retrieval by Johnson et al. [12], scene graphs have gradually become a research hotspot. Scene graphs model visual information, such as images or videos, and convert them into structured representations that are computationally understandable. They possess the ability to explicitly represent objects, attributes, and relationships, providing support for advanced understanding of complex scenes [13]. Scene graphs can be likened to knowledge graphs in the field of natural language processing [14]. Knowledge graphs contain a wealth of triple information, represented as head entity, relationship, and tail entity. In contrast, scene graphs present abstract entities from natural language processing as concrete image regions. This approach allows for a more intuitive understanding of the relationships between objects in the current scene, but it also adds complexity to the construction of scene graphs.
Currently, research on SGG primarily focuses on indoor scenes [15,16,17], autonomous driving scenes [18,19,20], image retrieval [12,21,22], medical image analysis [23,24], and other fields. For intelligent ships, their navigation scenes are often complex and dynamic. Factors such as varying sea conditions, lighting conditions in navigation environments, and the presence of other dynamically changing ships pose challenges to scene understanding. Scene graphs offer an effective solution to address these challenges. For instance, scene graphs can be used to establish spatial information models of navigation environments, clarifying the positional relationships between static elements such as ships, islands, and ports within the scene. Additionally, scene graphs enable the tracking and modeling of dynamic elements such as other ships, providing insights into their dynamic behavioral attributes and facilitating safe navigation for intelligent ships. In summary, SGG plays a very positive role in understanding navigation scenes for ships, which is also the main motivation for studying the NSGG in this paper.
While the motivation and significance of studying the generation of navigation scene graphs were introduced above, we have encountered a challenging issue in practical research. Most current scene graph research is based on publicly available datasets, which makes it difficult to conduct targeted research on NSGG. Furthermore, the navigation events or object relationships reflected in current real-world navigation image data are limited, which hinders the study of complex navigation scene graph generation. Therefore, it is also not conducive to studying the generation of complex navigation scene graphs. As a result, prior to experiments on NSGG, this paper will also construct a dataset dedicated to NSGG.
In summary, this paper will focus on studying NSGG. Firstly, this paper proposes a novel NSGG network, named SGK-Net, in order to enhance the understanding capability of intelligent ships in navigation scenes. Secondly, this paper also builds the first Ship Navigation Scene Graph Simulation dataset, named SNSG-Sim, for validating the effectiveness of the proposed NSGG method. The main contributions of this paper are as follows:
  • To address the challenge of complex and diverse relationships among target entities in the current SGG process, making it difficult to obtain the key relationships that are contextually relevant, we propose the Semantic-Guided Multimodal Fusion (SGMF) module. This module leverages prior information on relationship semantics to fuse multimodal information and construct relationship features, allowing for weighted relationships between entities and providing clarity on the key relationships among target entities in the current context.
  • To tackle the issue of redundancy in relationship features during the current SGG process, we propose the Graph Structure Learning-based Structure Evolution (GSLSE) module. This module utilizes graph structure learning to reduce redundancy in relationship features and optimize the computational complexity in subsequent contextual message passing.
  • To address the issue of unstable SGG caused by noise interference in the context information relied upon for relational reasoning, this paper proposes the Key Entity Message Passing (KEMP) module. It effectively utilizes context information to refine the relational features and reduce noise interference from non-key nodes.
  • In response to the lack of domain-specific datasets for generating navigation scene graphs, this paper introduces the first ship navigation scene graph simulation dataset, SNSG-sim. The dataset consists of 2240 frames of image data captured in different navigation scenes, encompassing 10 common navigation scene entities and 20 inter-entity relationships. This dataset serves as a foundation for research on NSGG.
The subsequent structure of this paper is as follows. Section 2 will provide an overview of related work in the field. Section 3 will introduce the proposed method for generating navigation scene graphs based on semantic-guided multimodal feature fusion and key entity message passing. Section 4 will present the SNSG-sim dataset proposed in this paper. In Section 5, the proposed NSGG method will be compared with existing methods using the SNSG-sim and Visual Genome datasets, followed by an analysis and discussion of the experimental results. Section 6 will summarize the contributions of this paper and provide an outlook on future work.

2. Related Work

2.1. Scene Graph Generation

The concept of scene graphs was introduced over a decade ago. However, it was not until 2020 that the SGG task gained increasing attention, driven by advancements in pre-training models and datasets such as Visual Genome [25] and Open Images [26]. Current research on SGG can be classified into three categories: 2D SGG, 3D SGG, and spatio-temporal SGG [13]. Among them, there are more datasets available for 2D SGG, and it is a more widely studied area. 2D SGG primarily focuses on objects and their relationships within a single image [27]. RU-Net [28] proposed a regularized unrolling network to alleviate the issue of imbalanced relationship types in training data for generative models. Ref. [29] introduced an SGG method based on causal inference to address biases in the training process of traditional SGG models. Ref. [30] presented a generation method based on region-aware attention learning, which improved the understanding of fine-grained visual regions by SGG models. SGTR [31] and ReITR [32] introduced the transformer to the SGG task, enhancing the relationship modeling capabilities of generative models.
Compared to 2D SGG, Spatio-temporal SGG primarily focuses on the task of SGG in dynamic videos, requiring the consideration of both spatial and temporal dimensions of relationships [33,34,35]. Spatio-temporal scene graphs exhibit stronger expressive capabilities, but they also entail greater challenges in generation and application. With the advancement of perception tasks such as 3D semantic segmentation [9,36,37] and 3D object detection [38,39,40], the 3D SGG task has gained attention in recent years. In contrast to the 2D SGG problem at the image level, understanding and representing the interactions between objects in three-dimensional space typically involve greater complexity [41,42,43]. In terms of scene graph applications, there have been research efforts in various domains such as indoor scenes [15,16,17], autonomous driving scenes [18,19,20], image retrieval [12,21,22], and medical image analysis [23,24]. However, there is currently a scarcity of SGG research specifically focusing on navigation scenes. Therefore, this paper aims to delve into the SGG task in this domain, as it is believed to significantly enhance the understanding of navigation scenes for intelligent ships.

2.2. Scene Graph Datasets

Scene graph datasets serve as the foundation for SGG research. Currently, scene graph datasets can be categorized into public datasets and domain-specific datasets. As SGG is a downstream task of object perception, some datasets in the public domain have evolved from object detection or segmentation datasets, such as MSCOCO [44], COCO-Stuff [45], Spatialvoc2k [46], etc. Other datasets are specifically designed for SGG tasks, including VRD [47], Spatialsense [48], Visual Genome [25], Open Images V6 [26], etc. Among them, Visual Genome and Open Images V6 have relatively large numbers of data samples and annotations for objects and relationships, making them the most commonly used public datasets in general SGG research. Diverse public datasets provide an experimental and analytical foundation for SGG research. However, domain-specific SGG research requires datasets specific to the domain. Domain-specific datasets are typically developed in conjunction with specific tasks. For example, refs. [18,19,20] investigated SGG in the context of autonomous driving and proposed datasets for validation. In this paper, we primarily focus on the SGG task in the context of ship navigation scenes. Therefore, we will construct a dataset suitable for training and evaluating the performance of NSGG methods.

3. Methodology

3.1. Motivation

The generation of scene graphs enables unmanned systems to extract semantic relationships among objects in a scene, thereby facilitating a more comprehensive understanding of the current context. In the specific domain of ship navigation, SGG also contributes to enhancing the scene perception capabilities of intelligent ships. However, the effectiveness of generating navigation scene graphs is hindered by the diversity between the semantic relationships and visual representations of objects in navigation scenes. This manifests as the existence of multiple semantic relationships between the same objects, and significant variations in the visual patterns associated with the same semantic relationship. Consequently, this increases the difficulty of extracting semantic features related to relationships.
As shown in Figure 1a, the visual relationships between the front and rear speedboats can be described by various predicates, such as chase, follow, or evict. The ambiguity of visual relationship semantics makes it difficult to extract specific semantic features. However, in Figure 1b, if the detected intermediate object in the current scene is a high-value object (entity), such as a cargo ship, the contextual information can be used to determine the relationship between the unmanned surface vessels (USVs) and the cargo ship as a protecting relationship, while the relationship between the USVs is likely to be a following relationship. This indicates that relationship inference in the process of generating navigation scene graphs heavily relies on contextual information. Therefore, it is crucial to eliminate noise interference, such as non-key nodes, and identify key objects. To address the diversity of relationships between objects in navigation scenes and the impact of noise interference in contextual information on relationship feature extraction, this paper proposes an NSGG method based on semantic-guided multimodal fusion and key entity message passing, named SGK-Net. The aim of this method is to improve the quality of relationship feature extraction and achieve efficient and high-quality NSGG.

3.2. Network Architecture

In the research of navigation scene graph generation, challenges arise from the complexity of relationships among navigation objects and the presence of noise in the context information relied upon for relationship inference. To address these challenges, this paper proposes a novel NSGG network, SGK-Net. The overall network architecture of the proposed method is illustrated in Figure 2, and it consists of three innovative modules: the Semantic-Guided Multimodal Fusion (SGMF), the Graph Structure Learning-based Structure Evolvement (GSLSE), and the Key Entity Message Passing (KEMP). Firstly, the SGMF module takes as input the multimodal features and entity information extracted by the detector, and fuses the multimodal features of paired entities to generate relationship features between the entities. This yields a dense representation of relationships. Then, to reduce the propagation of noise information and computational complexity during contextual inference, we utilize the GSLSE module to reconstruct the fully connected graph into a refined relationship graph. Finally, the KEMP module quantifies the weight of each node in the message-passing process based on the connectivity density of the nodes. It relies on the information from important nodes to refine the feature representation of each node in the scene graph, thereby enhancing the accuracy of the nodes and relationships in the scene graph.

3.3. Semantic-Guided Multimodal Fusion

Traditional methods often fail to accurately model relationship features, leading to a fragmented understanding of the associations between object multimodal features and overlooking the importance of relationship semantics hidden within the multimodal features. In the context of navigation scenes, the weight of relationship semantics varies across different modalities. For example, spatial features predominantly determine object orientation relationships, while visual information plays a more significant role in object behavioral relationships. Neglecting the prior semantic clues between objects can result in ambiguous relationship semantics, thereby affecting the generation of navigation scene graphs. To address this issue, this paper proposes the Semantic-Guided Multimodal Fusion module (SGMF), which focuses on the prior relationships between object categories. It leverages this prior information to enhance the visual and spatial feature representations between object pairs, thereby strengthening the representation of multimodal relationship features and mitigating the problem of semantic ambiguity in object relationships caused by missing information in multi-modal relationship features.
First, the image is processed using an object detector to obtain the semantic label information f s , visual information f v , and bounding box information f b of the objects in the image. In previous works, f v is commonly represented using a joint region representation. Typically, the multi-modal information contained in the bounding boxes of two objects is used to represent the relational features. The features from different modalities are concatenated to form a fused feature, where different channels focus on different patterns of relational information. However, the resulting relational features inevitably get mixed with a large amount of unrelated background information. The SGMF module is designed to extract the relevant patterns of relationship features by leveraging the prior semantic association information between paired objects. Specifically, given a subject i and an object j, the semantic feature vectors obtained from the detector are denoted as f s e ( c i ) and f s e ( c j ) , respectively. The semantic prior matrix x i j , representing the semantic correlation between the object pair, is computed as:
x i j = f s e ( c i ) f s e ( c j ) T .
The symbol ⊗ denotes matrix multiplication. Next, we compute the spatial relationship features between the object pairs based on the object detection bounding boxes obtained from the detector. In this paper, the calculation formula for the spatial relationship feature x i j s is as follows:
x i j s = σ ( x i x j w j , y i y j h j , log w i w j , log h i h j ) ,
where ( x i , y i , w i , h i ) b i , ( x j , y j , w j , h j ) b j represents the spatial prior information between the objects. As shown in Figure 2, the SGMF module then performs a channel-level fusion of the visual relationship representation x i j v and the spatial relationship feature x i j s . Therefore, the multimodal relationship feature x i j v , s , which captures both visual and spatial information, can be represented as:
x i j v , s = c a t ( x i j v , x i j s ) .
Next, a series of 2D convolutions and spatial pooling on x i j are performed to achieve channel attention, resulting in the attention vector e i j c . The obtained semantic relationship attention vector is then combined with the multi-modal relationship features through channel-wise operations to update the feature representation of the relationship. This process helps select feature patterns that have stronger semantic correlations.
e i j c = σ ( G p o o l i n g ( G c o n v n σ ( G c o n v 1 ( x i j ) ) )
x i j r = x i j v , s × e i j c ,
where c is the number of channels in the multi-modal relationship feature x i j v , s , σ represents the activation function, G c o n v denotes the 2D convolution operation, G p o o l i n g represents the 2D pooling operation. Finally, x i j r is obtained as the updated relationship feature, with × indicating the dot product.

3.4. Graph Structure Learning-Based Structure Evolvement

In previous research, to improve the prediction accuracy of relationships in scene graphs, it is common to model the relationships between each pair of objects. If N objects are detected in an image, the constructed scene graph would involve N 2 relationship features that need to be processed and predicted. For instance, MOTIFS [49] takes the features of all objects and feeds them into a Bi-LSTM to generate N 2 relationship features. MSDN [50] groups N objects into N ( N 1 ) pairs, excluding pairs with identical objects, and connects all different objects using directed edges. However, such approaches have negative impacts on learning contextual features in the following ways: (1) Noise data are propagated during the context message passing process. Some noisy nodes transmit their information in this dense graph, significantly affecting the convergence of features in other parts of the graph; (2) The computational complexity of dense graph connections is substantial, making it challenging for the entire model to converge. To address these issues, this paper draws inspiration from graph learning theory and proposes the graph structure learning-based structure evolvement module to simplify dense scene graphs and select appropriate connections as much as possible.
In the GSLSE module, the relationship feature x i j r is first fed into the GSLSE module to generate the edge-weighted skeleton graph or structured organization rule set S G = { r 1 , 2 , r 2 , 3 , , r n , n 1 } . The GSLSE module can be regarded as a binary classification task to determine the existence of associations between nodes, and its confidence reflects the degree of association between nodes in the scene graph. By replacing the edges with classification confidences, the fully connected graph is reconstructed into an edge-weighted skeleton graph. The implemented GSLSE module in this section consists of three fully connected layers, with ReLU non-linear activation functions applied between each layer. GSLSE directly takes the initial relationship feature x i j r as input and predicts the association between the objects o i and o j . GSLSE only considers the positive values of the classification results as the final confidence scores, which are then normalized using softmax. To control the high or low confidence scores, these scores are input into a gating function τ ( · ) proposed by [51], which helps align the predicted organization closer to the skeleton of the ground truth scene graph. The entire process is illustrated as follows:
r i , j = τ ( s o f t m a x ( G S L S E ( x i j r ) )
τ ( x ) = 0 x β α x α β β < x < 1 α + β 1 x 1 α + β ,
where α and β are two learnable parameters. The function τ ( · ) is used to process the result of the softmax function.
After the processing by the SGMF module in the previous section, a dense connectivity graph of the entire scene is obtained. The GSLSE module significantly reduces the connectivity between nodes in the graph. The GSLSE module’s processing not only reduces the propagation of redundant connection noise but also decreases the computational complexity during scene graph inference.

3.5. Key Entity Message Passing

Through the analysis conducted in Section 3.1, it has been observed that the improvement in the effectiveness of NSGG relies heavily on contextual information. Information is primarily conveyed through messages, wherein messages represent the essence of features and are exchanged between elements of the scene graph, including objects and relationships. For the source elements, messages are considered as their own features, while for the target elements, they serve as contextual information. Intuitively, predictions regarding objects and relationships can benefit from their contextual information. For instance, in the IMP [52] method, separate feature representations are constructed for nodes and relationships. Dual-form GRUs are utilized to transmit contextual information between nodes and edge contexts, facilitating information propagation between nodes and edges. VCTREE [53], on the other hand, employs a dynamic tree structure to place objects within visual contexts and utilizes bidirectional TreeLSTM to encode the visual contexts. However, the current utilization of contextual mechanisms suffers from a problem where the weight of entity information transmission remains consistent during the message-passing process. This uniform weight distribution leads to the overlooking of important entity information and the propagation of more noise information when understanding the contextual cues of navigation scenes. To address this issue, this section proposes the Key Entity Message Passing (KEMP) module to optimize the contextual propagation process.
Specifically, the first step involves utilizing the graph skeleton structure S G generated by the GSLSE module, which aggregates the connectivity information of all nodes in the graph. The connectivity status of the graph forms the basis for selecting important entities in the scene. The degree of each node can be calculated based on its connectivity status. We define the importance of an entity to be positively correlated with the degree of its corresponding node, thus quantifying the importance of entities. The calculation of the degree of a node o i is given by the following formula:
d o i = j N * i   ( r i , j I ( r i , j ) ) ,
where the function I ( · ) is an indicator function. It takes a condition r i , j as input and returns a value of 1 if r i , j is true and 0 if r i , j is false. Subsequently, in the message-passing process, to highlight the importance of neighboring nodes in aggregating neighborhood information for node o i , we employ a weight-based message aggregation approach. The context weights are calculated based on the degrees of neighboring nodes as follows:
Z d o i = d o i m i n ( d o ) m a x ( d o ) ,
where d o represents the set of node degrees, and Z ( · ) is the weight function for the node degrees. Next, in the message passing process, to highlight the importance of neighboring nodes in aggregating information within the neighborhood of node o i , we employ a weight-based message aggregation method. The context weights are calculated based on the out-degree of neighboring nodes, and the calculation is formalized as Z ( · ) . Subsequently, we use o i and x i j r to initialize the hidden state: E o i ( 0 ) = o i , V r i j ( 0 ) = x i j r . The formal expression for context message passing of key entities is as follows:
E o i ( l ) = g e e ( r i , j , d o j , E o j ( l 1 ) )
= j N * i ( r i , j · Z d o i · E o i ( l 1 ) + E o i ( l 1 ) )
V r i j ( l ) = g e r ( r i , j , E o i ( l 1 ) , E o j ( l 1 ) )
= λ [ g s ( r i , j · E o i ( l 1 ) ) + g o ( r i , j · E o j ( l 1 ) ) ] + V r i j ( l 1 ) ,
where E o i ( l ) represents the node features after l-th layer message passing, V r i j ( l ) represents the relationship features after l-th layer message passing, N * i denotes the neighborhood of node o i . e r signifies feature propagation between nodes, while e r denotes the utilization of node information to update edge features. g e e , g e r , g s , g o , etc., all refer to multi-layer perceptrons used to refine hidden layer representations. λ represents a learnable value.

3.6. Training Losses

The model training in this study is performed using supervised learning. The overall loss of the NSGG model, denoted as L a l l , is divided into three components: graph structure training loss L g s l s e , object detection loss L o b j , and relation predicate prediction loss L r e l :
L a l l = L o b j + L r e l + L g s l s e .
The L o b j component is primarily used to determine the object class and object region in the image. The L o b j loss is composed of classification loss and regression loss. In this study, the classification loss is calculated using cross-entropy loss, while the regression loss is calculated using smooth L1 loss. On the other hand, L r e l is mainly used to learn the relationships between objects, and this study employs cross-entropy loss for this purpose. Now, let us focus on introducing the graph structure training loss L g s l s e , employed in the GSLSE module in this study.
The traditional SGG models are typically supervised with triplets in the form of < e i , e i j , e j > , rather than the intuitive representation of scene graphs. This representation lacks structural and hierarchical information, resulting in a deficiency of explicit relationships. As a consequence, the translation of visually structured information into implicit representations leads to suboptimal performance. To address this issue, we propose the GSLSE module to generate the graph structure S G , while the ground truth structure representation of the scene graph is denoted as:
S G T = { b 12 , b 13 , , b n n 1 } ,
where b i j = 1 indicates the existence of an edge from o i to o j , and when b i j = 0 , the edge does not exist. The loss of the GSLSE module is then defined as follows:
L g s l s e = B C E ( S G , S G T ) .
The binary cross-entropy loss function, B C E ( · ) , is utilized in this context. It is worth noting that our graph structure data does not include any labels for nodes and edges; it solely contains information that constrains the model to learn structured graph representations.
By jointly training multiple tasks, we can prevent modules from converging independently in a static space, thus effectively integrating structured information into feature learning.

4. SNSG-Sim Dataset

Due to the limited availability of object relationships in existing ship navigation scene image data, and the scarcity of data suitable for scene graph generation tasks, there is currently no publicly available dataset specifically designed for ship navigation scene graph generation research. To facilitate a better understanding of ship navigation scene graph generation, this study presents the first Ship Navigation Scene Graph Simulation dataset, SNSG-sim. This dataset is constructed utilizing the Unity3D engine, taking into consideration various object relationships in different navigation scenes. It comprises a total of 2240 image data samples from diverse navigation scenes such as islands, ports, and open seas. The dataset includes 10 types of navigational entities (objects), including cargo ships, cruise ships, and USVs, as well as 20 common navigational object relationships.

4.1. Dataset Construction

To construct a simulation scene graph dataset, it is crucial to have a stable simulation platform that can accurately simulate the objects and backgrounds encountered during ship navigation. Existing scene graph generation datasets primarily rely on data collection through networks, lacking targeted domain-specific data and lacking precedents for constructing navigation scene graph datasets based on simulation environments. Therefore, in this study, the selection of the simulation environment was guided by the upstream perception task datasets, such as Vkitti [54] and S2S-sim [1], which utilized Unity3D as the simulation platform. Building upon this simulation engine, the scenes were constructed for the dataset.
In the construction of scenes, this study refers to commonly observed navigation scenes and primarily divides them into port, island, and open sea, as shown in Figure 3. Regarding the selection of target entities, in addition to common cargo ships, cruise ships, and USVs, it also includes offshore drilling platforms (ODPs) and other obstacles, as illustrated in Figure 4. In terms of entity relationship selection, apart from intuitive relationships such as “follow” and “beside”, we have also identified more complex implicit relationships, such as “defense” and “rescue”, which require contextual understanding through simulation. The specific entity and relationship annotations are presented in Figure 5. The object labels we annotated include categories and bounding boxes, while relationships are stored in the form of triplets, such as <USV, defense, cargo>. Based on simulations and collected actual samples, we eventually selected and annotated a total of 2240 image data samples, encompassing 20 types of entity relationships.

4.2. Dataset Analysis

We conducted a statistical analysis of the constructed SNSG-sim dataset. Figure 6 illustrates the distribution of target entities in the dataset. It can be observed that the annotated target entities in our dataset include foreground objects such as USVs, cargo ships, and cruise ships, as well as background elements like ports and islands, commonly found in navigation scenes. Among them, USVs have a relatively higher representation. This is because, in our dataset, USVs tend to appear in groups during tasks such as protection and defense, leading to a higher frequency of occurrence within the same image sample. Figure 7 displays the distribution of relationships in the constructed dataset. As mentioned earlier, our dataset specifically annotates 20 types of relationships between target entities, addressing the characteristics of navigation scenes. This augmentation enhances the diversity of scene graph generation in our study and improves the understanding of various navigation scenes by using ship models. Furthermore, the statistical graph reveals a similar long-tail distribution issue in both target entities and relationships within navigation scenes, as observed in public scene graph datasets [25,26]. This poses a challenge to the performance of NSGG methods.

5. Experiments

In this section, we will evaluate and analyze the proposed NSGG method through experiments on the SNSG-sim dataset. Additionally, we will also conduct experiments on the widely used public dataset, Visual Genome, to validate the generalizability of the proposed method.

5.1. Scene Graph Evaluation Tasks and Metrics

In order to better evaluate the learning and reasoning capabilities of the scene graph generation model regarding relationship predicates, this study selects two commonly used evaluation tasks in scene graph generation:
Predicate Classification (PredCls): The PredCls task involves predicting the predicates that describe the relationships between pairs of localized objects. This task independently examines the model’s performance in terms of predicate classification without considering other factors. It is primarily used to assess the model’s ability to learn predicates.
Scene Graph Classification (SGCls): The SGCls task involves predicting the object categories and predicates for each subject-object pair within a given set of localized objects. This task is primarily used to evaluate the scene graph generation model’s ability to perform contextual reasoning.
Common evaluation metrics used in SGG include Recall@K and mean Recall@K. Recall@K measures the proportion of correctly predicted relationships among the top K-predicted relationships when sorted by confidence scores. In our experiments on the SNSG-sim dataset, we set K values to 20 and 50. A higher value of this metric indicates better predictive performance of the model. The basic formula for recall R @ K is as follows:
R @ K = 1 Y y = 1 Y T y k G y ,
where Y represents the number of images in the test set. After the model’s computation, K relationship triplets will be generated for each image. G y represents the number of annotated relationship triplets in the dataset used, and T y k represents the number of predicted values that match the ground truth.
The mean Recall@K is also an important metric for evaluating SGG. The formula for calculating mean recall R @ K is as follows:
m R @ K = 1 X x = 1 X R @ K x ,
where X represents the number of relationships in the test set. The recall R @ K x for each relationship category is computed, and the mean recall m R @ K can be obtained by taking the mean of these recall values.

5.2. Experimental Settings

Dataset Settings: In the experimental section, the experiments mainly utilize the proposed SNSG-sim dataset. For the experiments, we partitioned the 2240 data samples into training and testing sets in a 4:1 ratio. Additionally, to evaluate the generalizability of our proposed method, we also conducted experiments on the publicly available Visual Genome dataset. The Visual Genome dataset consists of over 100,000 images, with the testing set containing 5000 sample images. On average, each image in the dataset is annotated with 35 object labels, 26 attribute labels, and 21 relationship labels. The dataset covers a wide range of everyday life scenes and complex social scenes. Due to the larger number of objects and attributes in the Visual Genome dataset, we followed the settings of other methods and set the value of K (Recall@K) to 50 and 100 during testing.
Implementation Details: We adopted the currently prevalent framework for scene graph generation implementation. The scene graph generation model was trained by incorporating a relationship prediction head into the Mask R-CNN algorithm framework. In our approach, the object detection component of the scene graph generation network utilized Faster R-CNN as the detector, with ResNet-101 serving as the backbone network, and the output dimension of the backbone network was set to 256. During training, we first pre-trained the Faster R-CNN object detector. Subsequently, when training the scene graph model, we froze all parameters before the ROIAlign network in Faster R-CNN. For each image, only the top 64 object candidates from the output of Faster R-CNN were retained. During the training process, a batch size of 12 was used, and the learning rate was set to 0.0015. The optimization was performed using the SGD (Stochastic Gradient Descent) algorithm, with a momentum factor of 0.81. The maximum number of iterations for the training process was set to 50,000. All experiments were conducted using one RTX 4090 GPU.

5.3. Quantitative Results and Comparison

SNSG-sim test results. The comparative experimental results on the SNSG-sim dataset are presented in Table 1. From the table, it can be observed that our proposed NSGG method exhibits superior performance in both the PreCLs and SGCls evaluation tasks. Thanks to the targeted design of our method for navigation scenes, it is the only method in the PreCLs task that achieves a recall rate of over 60% at both R@20 and R@50. Compared to the baseline method RU-Net, our method achieves improvements of 9.1% and 8.3% in R@20 and R@50, respectively, for the PreCLs task, and improvements of 7.7% and 8.0% in R@20 and R@50, respectively, for the SGCls task. Compared to the recent transformer-based work RelTR [32], our method exhibits significant improvements of 6.9% and 6.0% in R@20 and R@50, respectively, for the PreCLs task, and remarkable improvements of 15.2% and 13.9% in R@20 and R@50, respectively, for the SGCls task. These results validate the effectiveness of our method in generating navigation scene graphs. Figure 8 presents the visualization of our method for generating scene graphs on sample images from the SNSG-sim dataset. It can be observed that the generated scene graphs by our method accurately depict the objects and relationships in navigation scenes.
Visual Genome test results. The comparative experimental results on the Visual Genome dataset are presented in Table 2. Since most existing methods have been evaluated on the Visual Genome dataset, we are able to compare our method with a larger number of approaches on this dataset. It can be observed that the proposed scene graph generation method in this paper can also adapt to large-scale visual understanding tasks and generate high-quality scene graphs. Compared to current mainstream methods, our method achieves the best performance in both evaluation tasks. Specifically, the proposed method achieves an R@50 of 69.1% and an R@100 of 71.1% for the PreCLs task on the Visual Genome dataset. For the SGCls task, the method achieves an R@50 of 42.7% and an R@100 of 43.9%. Compared to the baseline model RU-Net, our method improves the R@50 by 1.4% and 1.5% for the PreCLs task and by 1.3% and 1.4% for the SGCls task. Compared to the recent work of RelTR, our method also shows significant improvements in the R@50 metric, with an increase of 4.9% for the PreCLs task and 6.1% for the SGCls task. Figure 9 showcases the visualization of our method for generating scene graphs on sample images from the Visual Genome dataset. These experimental results demonstrate the effectiveness of the proposed method in generating scene graphs on the Visual Genome dataset, thus confirming the generalizability of our method.

5.4. Ablation Studies

In order to validate the effectiveness of the innovative modules in our proposed NSGG method, we conducted ablation experiments on both the SNSG-sim dataset and the Visual Genome dataset. In the ablation experiments, the baseline method used was RU-Net, and the recall rate K values for all experiments were set to 20 and 50.
Table 3 presents the ablation experiment results of our method on the SNSG-sim dataset. It can be observed that, compared with the baseline EXP.1, the EXP.2 and EXP.3 that introduced SGMF or GSLSE modules achieve improved performance on both PreCls and SGCls evaluation tasks, and the performance gain brought by the GSLSE module is particularly significant. However, in the EXP.4, which solely introduced the KEMP module, the model performance did not show a significant improvement compared with the baseline method. But in EXP.5, combining GSLSE and KEMP further improved the performance. The main reason is that GSLSE optimizes the scene graph structure, making the message passing between key nodes in the KEMP module more smooth. In EXP.6, introducing all three modules proposed in this paper achieved the best performance, indicating that combining the use of all modules designed in this paper is the most reasonable approach.
Table 4 presents the ablation experiment results of our method on the Visual Genome dataset. From EXP.2, it can be observed that compared to the baseline method, the introduction of the SGMF module leads to a 1.3% improvement in R@50 for the PreCls task. This indicates that the SGMF module is capable of better modeling relationship features. From EXP.3, it can be seen that the introduction of the GSLSE module results in a 1.8% increase in R@50 for the SGCls task. This demonstrates that optimizing the scene graph’s connectivity through the GSLSE effectively facilitates feature learning. Similar to the ablation experiments conducted on the SNSG-sim dataset, the introduction of the KEMP module in EXP.4 does not show significant improvements. However, by analyzing EXP.4 and EXP.5 together, it can be observed that the KEMP module requires the assistance of the GSLSE module to unleash its effectiveness. This relates to the computation of key nodes, where GSLSE aids the KEMP module in calculating key nodes within the scene graph. EXP.6 further demonstrates that the combined usage of the proposed modules in this paper achieves the best performance in both the PreCls and SGCls evaluation tasks. This validates the effectiveness of the innovative modules proposed in this paper.

6. Conclusions and Future Work

To address the challenges of generating navigation scene graphs with poor performance due to the complexity of relationships between objects and the presence of noise in the contextual information relied upon for relationship inference, this paper proposes a novel NSGG network, SGK-Net. The proposed network comprises three innovative modules: SGMF, GSLSE, and KEMP. (1) The SGMF module leverages prior information about relationship semantics to fuse multimodal information and construct relationship features, reducing the semantic ambiguity of relationships between objects caused by missing multimodal relationship feature information. (2) The GSLSE module reduces redundancy in relationship features and optimizes subsequent message propagation calculations. (3) The KEMP module utilizes contextual information to refine relationship features and mitigate noise interference from non-key nodes. Additionally, (4) this paper introduces the SNSG-sim dataset, the first simulated dataset for ship navigation scene graph generation, providing a valuable data foundation for research in this area. The effectiveness and generalizability of the proposed method are validated through experiments conducted on the SNSG-sim dataset and the publicly available Visual Genome dataset. Ablation experiments demonstrate that the coordinated operation of the innovative modules in this paper enables the effective extraction of relationship features between target entities, resulting in the generation of navigation scene graphs.
This paper focuses on the NSGG task, and the proposed method provides valuable insights into improving the understanding of navigation scenes for intelligent ships. However, there are also some limitations of this research, primarily in two respects. Firstly, although the SGMF module in this paper achieves a fusion of multimodal features, there still exists a semantic gap between visual representation and relationship predicates, which hinders the effectiveness of scene graphs. In future work, we will pay more attention to addressing the semantic representation issues across different modalities. Secondly, the research in this paper is primarily based on the simulated dataset SNSG-sim, without conducting experiments in real navigation scenes. However, there exist domain differences between real-world data and simulated data. Therefore, transferring the proposed method to real navigation scenes to enhance its practicality is a task we need to accomplish in the future.

Author Contributions

Conceptualization, W.Y. and H.Q.; methodology, H.Q. and W.Y.; software, H.Q.; validation, W.Y. and H.Q.; formal analysis, W.Y.; investigation, W.Y.; resources, X.L.; data curation, H.Q.; writing—original draft preparation, W.Y.; writing—review and editing, H.Q.; visualization, W.Y.; supervision, S.X.; project administration, X.L.; funding acquisition, S.X. All authors have read and agreed to the published version of the manuscript.

Funding

The research reported in this paper was supported by the National Natural Science Foundation of China under grant No. 61991415, and the Development Project of Ship Situational Intelligent Awareness System under grant MC-201920-X01.

Institutional Review Board Statement

Written informed consent for publication of this paper was obtained from Shanghai University and all authors.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data required to reproduce these findings cannot be shared at this time as the data also form part of an ongoing study. They can be requested from the author by e-mail ([email protected]) in the future.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, W.; Wang, X.; Luo, X.; Xie, S.; Chen, J. S2S-Sim: A Benchmark Dataset for Ship Cooperative 3D Object Detection. Electronics 2024, 13, 885. [Google Scholar] [CrossRef]
  2. Bae, I.; Hong, J. Survey on the developments of unmanned marine vehicles: Intelligence and cooperation. Sensors 2023, 23, 4643. [Google Scholar] [CrossRef] [PubMed]
  3. Guo, Y.; Yu, H.; Ma, L.; Zeng, L.; Luo, X. THFE: A Triple-hierarchy Feature Enhancement method for tiny boat detection. Eng. Appl. Artif. Intell. 2023, 123, 106271. [Google Scholar] [CrossRef]
  4. Guo, Y.; Yu, H.; Xie, S.; Ma, L.; Cao, X.; Luo, X. DSCA: A Dual Semantic Correlation Alignment Method for domain adaptation object detection. Pattern Recognit. 2024, 150, 110329. [Google Scholar] [CrossRef]
  5. Berwo, M.A.; Khan, A.; Fang, Y.; Fahim, H.; Javaid, S.; Mahmood, J.; Abideen, Z.U.; MS, S. Deep learning techniques for vehicle detection and classification from images/videos: A survey. Sensors 2023, 23, 4832. [Google Scholar] [CrossRef] [PubMed]
  6. Sun, X.; Wang, P.; Yan, Z.; Xu, F.; Wang, R.; Diao, W.; Chen, J.; Li, J.; Feng, Y.; Xu, T.; et al. FAIR1M: A benchmark dataset for fine-grained object recognition in high-resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2022, 184, 116–130. [Google Scholar] [CrossRef]
  7. Tan, J.; Wang, C.; Li, B.; Li, Q.; Ouyang, W.; Yin, C.; Yan, J. Equalization loss for long-tailed object recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11662–11671. [Google Scholar]
  8. Gao, G.; Zhou, P.; Yao, L.; Liu, J.; Zhang, C.; Duan, D. A bi-prototype BDC metric network with lightweight adaptive task attention for few-shot fine-grained ship classification in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5622416. [Google Scholar] [CrossRef]
  9. Yang, W.; Sheng, S.; Luo, X.; Xie, S. Geometric relation based point clouds classification and segmentation. Concurr. Comput. Pract. Exp. 2022, 34, e6845. [Google Scholar] [CrossRef]
  10. Han, W.; Zhao, B.; Luo, J. Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios. Sensors 2023, 23, 4789. [Google Scholar] [CrossRef]
  11. Li, L.; Ji, W.; Wu, Y.; Li, M.; Qin, Y.; Wei, L.; Zimmermann, R. Panoptic scene graph generation with semantics-prototype learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 3145–3153. [Google Scholar]
  12. Johnson, J.; Krishna, R.; Stark, M.; Li, L.J.; Shamma, D.; Bernstein, M.; Fei-Fei, L. Image retrieval using scene graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3668–3678. [Google Scholar]
  13. Li, H.; Zhu, G.; Zhang, L.; Jiang, Y.; Dang, Y.; Hou, H.; Shen, P.; Zhao, X.; Shah, S.A.A.; Bennamoun, M. Scene graph generation: A comprehensive survey. Neurocomputing 2024, 566, 127052. [Google Scholar] [CrossRef]
  14. Peng, C.; Xia, F.; Naseriparsa, M.; Osborne, F. Knowledge graphs: Opportunities and challenges. Artif. Intell. Rev. 2023, 56, 13071–13102. [Google Scholar] [CrossRef]
  15. Gao, L.; Sun, J.M.; Mo, K.; Lai, Y.K.; Guibas, L.J.; Yang, J. Scenehgn: Hierarchical graph networks for 3d indoor scene generation with fine-grained geometry. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 8902–8919. [Google Scholar] [CrossRef]
  16. Wald, J.; Dhamo, H.; Navab, N.; Tombari, F. Learning 3d semantic scene graphs from 3d indoor reconstructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3961–3970. [Google Scholar]
  17. Tahara, T.; Seno, T.; Narita, G.; Ishikawa, T. Retargetable AR: Context-aware augmented reality in indoor scenes based on 3D scene graph. In Proceedings of the 2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Recife, Brazil, 9–13 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 249–255. [Google Scholar]
  18. Gu, J.; Zhao, H.; Lin, Z.; Li, S.; Cai, J.; Ling, M. Scene graph generation with external knowledge and image reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1969–1978. [Google Scholar]
  19. Malawade, A.V.; Yu, S.Y.; Hsu, B.; Muthirayan, D.; Khargonekar, P.P.; Al Faruque, M.A. Spatiotemporal scene-graph embedding for autonomous vehicle collision prediction. IEEE Internet Things J. 2022, 9, 9379–9388. [Google Scholar] [CrossRef]
  20. Yu, S.Y.; Malawade, A.V.; Muthirayan, D.; Khargonekar, P.P.; Al Faruque, M.A. Scene-graph augmented data-driven risk assessment of autonomous vehicle decisions. IEEE Trans. Intell. Transp. Syst. 2021, 23, 7941–7951. [Google Scholar] [CrossRef]
  21. Pei, J.; Zhong, K.; Yu, Z.; Wang, L.; Lakshmanna, K. Scene graph semantic inference for image and text matching. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2023, 22, 1–23. [Google Scholar] [CrossRef]
  22. Wang, Y.; Dai, P.; Jia, X.; Zeng, Z.; Li, R.; Cao, X. Hi-SIGIR: Hierachical Semantic-Guided Image-to-image Retrieval via Scene Graph. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 6400–6409. [Google Scholar]
  23. D’souza, J.; Aleema, P.; Dhanyashree, S.; Fernandes, C.; Kavitha, K.; Naik, C. Knowledge-Based Scene Graph Generation in Medical Field. In Proceedings of the 2023 IEEE International Conference on Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER), Mangalore, India, 13–14 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 232–237. [Google Scholar]
  24. Holm, F.; Ghazaei, G.; Czempiel, T.; Özsoy, E.; Saur, S.; Navab, N. Dynamic Scene Graph Representation for Surgical Video. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 81–87. [Google Scholar]
  25. Krishna, R.; Zhu, Y.; Groth, O.; Johnson, J.; Hata, K.; Kravitz, J.; Chen, S.; Kalantidis, Y.; Li, L.J.; Shamma, D.A.; et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 2017, 123, 32–73. [Google Scholar] [CrossRef]
  26. Kuznetsova, A.; Rom, H.; Alldrin, N.; Uijlings, J.; Krasin, I.; Pont-Tuset, J.; Kamali, S.; Popov, S.; Malloci, M.; Kolesnikov, A.; et al. The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale. Int. J. Comput. Vis. 2020, 128, 1956–1981. [Google Scholar] [CrossRef]
  27. Lin, X.; Ding, C.; Zhan, Y.; Li, Z.; Tao, D. Hl-net: Heterophily learning network for scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19476–19485. [Google Scholar]
  28. Lin, X.; Ding, C.; Zhang, J.; Zhan, Y.; Tao, D. Ru-net: Regularized unrolling network for scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19457–19466. [Google Scholar]
  29. Tang, K.; Niu, Y.; Huang, J.; Shi, J.; Zhang, H. Unbiased scene graph generation from biased training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3716–3725. [Google Scholar]
  30. Liu, A.A.; Tian, H.; Xu, N.; Nie, W.; Zhang, Y.; Kankanhalli, M. Toward region-aware attention learning for scene graph generation. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7655–7666. [Google Scholar] [CrossRef] [PubMed]
  31. Li, R.; Zhang, S.; He, X. Sgtr: End-to-end scene graph generation with transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19486–19496. [Google Scholar]
  32. Cong, Y.; Yang, M.Y.; Rosenhahn, B. Reltr: Relation transformer for scene graph generation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 11169–11183. [Google Scholar] [CrossRef]
  33. Xu, L.; Qu, H.; Kuen, J.; Gu, J.; Liu, J. Meta spatio-temporal debiasing for video scene graph generation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 374–390. [Google Scholar]
  34. Cong, Y.; Liao, W.; Ackermann, H.; Rosenhahn, B.; Yang, M.Y. Spatial-temporal transformer for dynamic scene graph generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16372–16382. [Google Scholar]
  35. Ji, J.; Krishna, R.; Fei-Fei, L.; Niebles, J.C. Action genome: Actions as compositions of spatio-temporal scene graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10236–10247. [Google Scholar]
  36. Zhan, L.; Li, W.; Min, W. FA-ResNet: Feature affine residual network for large-scale point cloud segmentation. Int. J. Appl. Earth Obs. Geoinf. 2023, 118, 103259. [Google Scholar] [CrossRef]
  37. Lu, Y.; Jiang, Q.; Chen, R.; Hou, Y.; Zhu, X.; Ma, Y. See more and know more: Zero-shot point cloud segmentation via multi-modal visual data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 21674–21684. [Google Scholar]
  38. Shi, S.; Jiang, L.; Deng, J.; Wang, Z.; Guo, C.; Shi, J.; Wang, X.; Li, H. PV-RCNN++: Point-voxel feature set abstraction with local vector representation for 3D object detection. Int. J. Comput. Vis. 2023, 131, 531–551. [Google Scholar] [CrossRef]
  39. Mao, J.; Shi, S.; Wang, X.; Li, H. 3D object detection for autonomous driving: A comprehensive survey. Int. J. Comput. Vis. 2023, 131, 1909–1963. [Google Scholar] [CrossRef]
  40. Wang, L.; Zhang, X.; Song, Z.; Bi, J.; Zhang, G.; Wei, H.; Tang, L.; Yang, L.; Li, J.; Jia, C.; et al. Multi-modal 3d object detection in autonomous driving: A survey and taxonomy. IEEE Trans. Intell. Veh. 2023, 8, 3781–3798. [Google Scholar] [CrossRef]
  41. Feng, M.; Hou, H.; Zhang, L.; Guo, Y.; Yu, H.; Wang, Y.; Mian, A. Exploring hierarchical spatial layout cues for 3d point cloud based scene graph prediction. IEEE Trans. Multimed. 2023. early access. [Google Scholar] [CrossRef]
  42. Koch, S.; Hermosilla, P.; Vaskevicius, N.; Colosi, M.; Ropinski, T. Sgrec3d: Self-supervised 3d scene graph learning via object-level scene reconstruction. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 3404–3414. [Google Scholar]
  43. Wu, S.C.; Wald, J.; Tateno, K.; Navab, N.; Tombari, F. Scenegraphfusion: Incremental 3d scene graph prediction from rgb-d sequences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7515–7525. [Google Scholar]
  44. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  45. Caesar, H.; Uijlings, J.; Ferrari, V. Coco-stuff: Thing and stuff classes in context. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1209–1218. [Google Scholar]
  46. Belz, A.; Muscat, A.; Anguill, P.; Sow, M.; Vincent, G.; Zinessabah, Y. Spatialvoc2k: A multilingual dataset of images with annotations and features for spatial relations between objects. In Proceedings of the 11th International Conference on Natural Language Generation, Tilburg, The Netherlands, 5–8 November 2018; pp. 140–145. [Google Scholar]
  47. Lu, C.; Krishna, R.; Bernstein, M.; Fei-Fei, L. Visual relationship detection with language priors. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 852–869. [Google Scholar]
  48. Yang, K.; Russakovsky, O.; Deng, J. Spatialsense: An adversarially crowdsourced benchmark for spatial relation recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2051–2060. [Google Scholar]
  49. Zellers, R.; Yatskar, M.; Thomson, S.; Choi, Y. Neural motifs: Scene graph parsing with global context. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5831–5840. [Google Scholar]
  50. Li, Y.; Ouyang, W.; Zhou, B.; Wang, K.; Wang, X. Scene graph generation from objects, phrases and region captions. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1261–1270. [Google Scholar]
  51. Li, R.; Zhang, S.; Wan, B.; He, X. Bipartite graph network with adaptive message passing for unbiased scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11109–11119. [Google Scholar]
  52. Xu, D.; Zhu, Y.; Choy, C.B.; Fei-Fei, L. Scene graph generation by iterative message passing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5410–5419. [Google Scholar]
  53. Tang, K.; Zhang, H.; Wu, B.; Luo, W.; Liu, W. Learning to compose dynamic tree structures for visual contexts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6619–6628. [Google Scholar]
  54. Gaidon, A.; Wang, Q.; Cabon, Y.; Vig, E. Virtual worlds as proxy for multi-object tracking analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4340–4349. [Google Scholar]
  55. Herzig, R.; Raboh, M.; Chechik, G.; Berant, J.; Globerson, A. Mapping images to scene graphs with permutation-invariant structured prediction. Adv. Neural Inf. Process. Syst. 2018, 31, 7211–7221. [Google Scholar]
  56. Lin, X.; Ding, C.; Zeng, J.; Tao, D. Gps-net: Graph property sensing network for scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3746–3753. [Google Scholar]
  57. Yang, J.; Lu, J.; Lee, S.; Batra, D.; Parikh, D. Graph r-cnn for scene graph generation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 670–685. [Google Scholar]
  58. Zhong, Y.; Shi, J.; Yang, J.; Xu, C.; Li, Y. Learning to generate scene graph from natural language supervision. In Proceedings of the CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1803–1814. [Google Scholar]
  59. Lu, Y.; Rai, H.; Chang, J.; Knyazev, B.; Yu, G.; Shekhar, S.; Taylor, G.W.; Volkovs, M. Context-aware scene graph generation with seq2seq transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 15931–15941. [Google Scholar]
Figure 1. Semantic relationships of objects in different navigation scenes.
Figure 1. Semantic relationships of objects in different navigation scenes.
Sensors 24 04329 g001
Figure 2. Network Architecture. The SGK-Net comprises three innovative modules: Semantic-Guided Multimodal Fusion (SGMF), Graph Structure Learning-based Structure Evolution (GSLSE), and Key Entity Message Passing (KEMP). These modules aim to address the existing issues in current scene graph generation methods. SGMF is specifically designed to tackle the problem of ambiguous relationships between entities. GSLSE is utilized to optimize the connectivity structure of the navigation scene graph, thereby reducing redundant connections between entity nodes. KEMP is designed to mitigate the interference caused by non-key nodes on relationship inference.
Figure 2. Network Architecture. The SGK-Net comprises three innovative modules: Semantic-Guided Multimodal Fusion (SGMF), Graph Structure Learning-based Structure Evolution (GSLSE), and Key Entity Message Passing (KEMP). These modules aim to address the existing issues in current scene graph generation methods. SGMF is specifically designed to tackle the problem of ambiguous relationships between entities. GSLSE is utilized to optimize the connectivity structure of the navigation scene graph, thereby reducing redundant connections between entity nodes. KEMP is designed to mitigate the interference caused by non-key nodes on relationship inference.
Sensors 24 04329 g002
Figure 3. Simulation of common navigation scenes.
Figure 3. Simulation of common navigation scenes.
Sensors 24 04329 g003
Figure 4. Part of objects appearing in simulated navigation scenes.
Figure 4. Part of objects appearing in simulated navigation scenes.
Sensors 24 04329 g004
Figure 5. An example of entity and relationship annotation for attack and defense entities is shown in the figure. The cargo ship is the main object, while the defense USV is depicted in red, and the attack USV is depicted in blue.
Figure 5. An example of entity and relationship annotation for attack and defense entities is shown in the figure. The cargo ship is the main object, while the defense USV is depicted in red, and the attack USV is depicted in blue.
Sensors 24 04329 g005
Figure 6. Distribution of entities (objects) in the SNSG-sim dataset.
Figure 6. Distribution of entities (objects) in the SNSG-sim dataset.
Sensors 24 04329 g006
Figure 7. Distribution of relationship in the SNSG-sim dataset.
Figure 7. Distribution of relationship in the SNSG-sim dataset.
Sensors 24 04329 g007
Figure 8. The visualization results of our method on the SNSG-sim dataset for scene graph generation.
Figure 8. The visualization results of our method on the SNSG-sim dataset for scene graph generation.
Sensors 24 04329 g008
Figure 9. The visualization results of our method on the Visual Genome dataset for scene graph generation.
Figure 9. The visualization results of our method on the Visual Genome dataset for scene graph generation.
Sensors 24 04329 g009
Table 1. Performance comparisons with state-of-the-art methods on SNSG-sim dataset.
Table 1. Performance comparisons with state-of-the-art methods on SNSG-sim dataset.
MethodBSGCls
R@20R@50BR@50
MOTIFS [49]54.455.940.541.6
IMP [52]48.649.342.344.1
VCTREE [53]59.660.729.431.7
HL-Net [27]57.259.151.952.7
RU-Net [28]52.954.049.550.3
RelTR [32]55.156.342.144.4
Ours62.062.357.358.3
Table 2. Performance comparisons with state-of-the-art methods on the Visual Genome dataset.
Table 2. Performance comparisons with state-of-the-art methods on the Visual Genome dataset.
MethodPreClsSGCls
R@50R@100R@50R@100
MOTIFS [49]65.867.135.836.5
IMP [52]59.361.334.635.7
GPI [55]65.166.936.538.8
VCTREE [53]66.468.138.138.8
GPS-Net [56]66.968.839.240.1
G-RCNN [57]54.259.129.631.6
SGGNLS [58]65.667.440.040.8
Seq2Seq-RL [59]66.468.538.339.0
RU-Net [28]67.769.641.442.3
RelTR [32]64.2-36.6-
Ours69.171.142.743.6
Table 3. Ablation experiments on SNSG-sim dataset.
Table 3. Ablation experiments on SNSG-sim dataset.
EXP.SGMFGSLSEKEMPPreClsSGCls
R@20R@50R@20R@50
1 52.854.049.550.3
2 54.255.752.253.5
3 57.858.354.655.2
4 52.355.348.350.1
5 60.661.155.756.4
662.062.357.358.3
Table 4. Ablation experiments on the Visual Genome dataset.
Table 4. Ablation experiments on the Visual Genome dataset.
EXP.SGMFGSLSEKEMPPreClsSGCls
R@20R@50R@20R@50
1 57.663.435.235.3
2 58.264.735.835.9
3 59.366.336.237.1
4 58.165.035.037.4
5 61.768.337.442.5
662.369.142.743.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, W.; Qiu, H.; Luo, X.; Xie, S. SGK-Net: A Novel Navigation Scene Graph Generation Network. Sensors 2024, 24, 4329. https://doi.org/10.3390/s24134329

AMA Style

Yang W, Qiu H, Luo X, Xie S. SGK-Net: A Novel Navigation Scene Graph Generation Network. Sensors. 2024; 24(13):4329. https://doi.org/10.3390/s24134329

Chicago/Turabian Style

Yang, Wenbin, Hao Qiu, Xiangfeng Luo, and Shaorong Xie. 2024. "SGK-Net: A Novel Navigation Scene Graph Generation Network" Sensors 24, no. 13: 4329. https://doi.org/10.3390/s24134329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop