Next Article in Journal
Free Rocking of a Rigid Block on a Flexible Structure with Non-Smooth Contact Dynamics
Previous Article in Journal
The Effect of Pre-Treatment and the Drying Method on the Nutritional and Bioactive Composition of Sea Cucumbers—A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Construction Method of Knowledge Graphs for Pirate Events

1
Department of Information Fusion, Naval Aviation University, Yantai 264001, China
2
The School of Aviation Basis, Naval Aviation University, Yantai 264001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(15), 6482; https://doi.org/10.3390/app14156482
Submission received: 1 April 2024 / Revised: 22 May 2024 / Accepted: 22 May 2024 / Published: 25 July 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
With the development of seaborne trade, international maritime crime is becoming increasingly complex. Detecting maritime threats by fusing the physical movement data from traditional physical sensors is not sufficient. Thus, soft data, including intelligence reports and news articles, need to be incorporated into the situational awareness models of maritime threats. In this regard, this study developed an automated construction method of knowledge graphs for pirate events, which lays a foundation for subsequent maritime threat reasoning and situational awareness. First, a knowledge graph ontology model for pirate events was designed. Secondly, the BERT-BiLSTM-CRF model is proposed for named-entity recognition, and an entity linking algorithm based on distant learning and context attention mechanism is proposed to remove the conceptual ambiguity. Thirdly, based on traditional distant supervision relation extraction, which is based on sentence-level attention mechanism, bag-level and group-level attention mechanism methods are additionally proposed to further enhance the performance of distant supervision relation extraction. The proposed model demonstrated high performance in named-entity recognition, entity linking, and relation extraction tasks, with an overall F1-score of over 0.94 for NER and significant improvements in entity linking and relation extraction compared to traditional methods. The constructed knowledge graphs effectively support maritime threat reasoning and situational awareness, offering a substantial contribution to the field of maritime security. Our discussion highlights the model’s strengths and potential areas for future work, while the conclusion emphasizes the practical implications and the readiness of our approach for real-world applications.

1. Introduction

Maritime shipping is the most important mode of international transportation in the world today and plays a vital role in the development of world trade and economies [1]. However, non-traditional security risks, mainly piracy and maritime terrorism, have become the main issues affecting maritime shipping security. Among them, pirate attacks have the widest impact. Once such attacks occur, they will not only cause property losses but also pose a great threat to the lives of the people on board and may even have an adverse impact on navigation and the environment [2]. Therefore, constructing a situational awareness model for pirate events using available information is particularly important in the early warning of pirate events and maritime crime cracking [3].
Much of the current research focuses on automated processing, fusing, and reasoning from different data sources in maritime surveillance to assist official decision making. Using only the physical movement data from traditional physical sensors (i.e., hard data) makes full-scale situational awareness difficult [4]. Thus, soft data need to be incorporated into maritime threat situational awareness models. Here, soft data refer to structured text data, semi-structured and unstructured news, social media data, maritime incident reports, and so on [5,6,7,8,9,10]. These soft data include structured text data, semi-structured and unstructured news, social media data, maritime incident reports, and other relevant sources. In response, this paper presents methods that can extract such knowledge and construct knowledge graphs for pirate events, which are conducive to subsequent maritime threat reasoning for maritime threat situational awareness in general.
This paper presents an in-depth study on how to construct knowledge graphs for pirate events, and its contributions can be summarized as follows:
(1)
A Pirate Event Model (PEM) is designed as an ontology model of the knowledge graphs for pirate events. Four types of entities are set: pem: Aggressor, pem: Victim, pem: Location, and pem: Date, alongside the entity attributes of these four types of entities. In addition, two major categories of 11 types of relations among entities are designed.
(2)
For entity extraction, the BERT-BiLSTM-CRF model is used for NER from the soft data of pirate events. In the entity linking process, the method of distance learning is introduced for entity linking model training to address the problem of the lack of large amount of labeled text data. In addition, the context attention mechanism is proposed to select the words that are helpful for disambiguation, so as to enhance the performance of entity linking model training.
(3)
For RE, based on the traditional sentence-level attention mechanism, a bag-level attention mechanism is further introduced, which emphasizes the sentences that are related to the bag label at the global level by calculating the correlation between sentences. As such, the information of the bag is expressed, while sentences that are not related to the bag label are suppressed. Furthermore, bags with the same labels are assembled as a bag-group, and comprehensive discriminative features are mined at the bag-group-level as high-quality features of the relations between representation entity pairs.
(4)
For the algorithms involved in instance layer filling, experiments were designed to verify the effectiveness and superiority of the proposed models, including NER, entity linking, and relation extraction models, at the same time. According to the given pirate event text, using the proposed NER, entity linking, and relation extraction models, knowledge triples can be accurately extracted, and a high-quality knowledge graph for pirate events is constructed.

2. Related Works

2.1. Knowledge Graph Ontology Model

Knowledge graph construction includes ontology layer design and instance layer filling. The ontology layer is designed through an event-based ontology model to describe conceptual entities and their relations. Subsequently, through named-entity recognition (NER) and relation extraction, entities, attributes, and relations are extracted from diverse data sources to obtain knowledge triples, thus generating the ontologized knowledge representations and completing the construction of knowledge graphs. Currently, generic knowledge graphs, such as Freebase [11], Wikidata [12], and YAGO4 [13], are widely used, with corresponding vertical industry applications in medical, academic, and financial fields [14,15,16]. However, in the field of navigation, there are few relevant studies, especially studies on knowledge graphs for pirate events. Therefore, constructing a knowledge graph ontology model for pirate events is a primary research task. Currently, the main ontology models used in academia to describe events are the ABC (purposely given an innocuous name) [17], simple event model (SEM) [18], event ontology (EO) [19], and CIDOC-CRM [20] models. Based on the SEM model, a Pirate Event Model (PEM) was designed as a knowledge graph ontology model for pirate events.

2.2. Entity Extraction and Linking

Once the ontology layer is constructed, NER needs to be utilized to extract the required entities from the unstructured text data. Currently, NER based on neural networks has become a mainstream method [21,22], among which the convolutional neural network-conditional random field (CNN-CRF) model [21] and bidirectional long short-term memory network-conditional random field (BiLSTM-CRF) model [22] are the most commonly used. With a large number of entity mentions obtained through NER, there is a need for the further identification of the real-world entity to which an entity mention refers to through entity linking. There are two main steps: candidate entity generation and candidate entity linking. There are three main methods for candidate entity generation: name dictionary-based methods [23], surface form expansion [24,25], and object library-based methods [26]. After obtaining the entity mentions and feature representations of the contextual information of entities, entity linking performs similarity calculation between the entity mentions and representations to select the target entities. There are two main methods: traditional feature methods and representation learning methods. The essence of traditional feature methods is how to manually design effective features. Honnibal et al. [27] used the Bow model to obtain vectors of entity mentions and candidate entities; subsequently, the one with the highest cosine similarity score was regarded as the candidate entity. Background knowledge, priori knowledge, and entity typing information of candidate entities are also important for entity disambiguation; thus, many researchers have taken this information into account to improve the accuracy of disambiguation [28,29]. Traditional feature methods are heuristic for both target entity and entity mention representation, such as the bag-of-words model and term frequency-inverse document frequency (TF-IDF). It is difficult to adjust these heuristic algorithms or to capture the more fine-grained semantic and structural information; thus, traditional feature methods are not dominant. The essence of representation learning methods is how to obtain a distributed representation of the context of entities and entity mentions. The representation of entities is generally complex. An entity may be represented at different granularities, and the representation may also involve the entity type information. Neural networks are often used to automatically learn entities as well as distributed representations of entity mentions [30,31,32]. Usually, there is a lot of contextual information of entity mentions. However, some words in the context are not very relevant to the entity mentions. Therefore, the continuous representation of the trained context contains noise, which affects the accuracy of entity disambiguation. Researchers have proposed combining an attention mechanism with deep neural networks to train the semantic feature vectors of context to improve entity disambiguation models [33]. According to Sun et al. [34], important clues about entity mentions and candidate entities from the surrounding context could be discovered automatically through the attention mechanism, facilitating entity disambiguation. Raiman et al. [35] proposed a strategy in which the symbolic information was explicit-ensembled into the neural network inference process and subsequently applied this strategy to the entity disambiguation task so that entity linking was performed by training a type system. The current work focuses on supervised learning, where the training text contains annotated entity mentions (i.e., entity mentions linked to a knowledge base). However, in many specific verticals, there does not exist a large amount of labeled data, only a knowledge base and collection of unlabeled texts datasets from the corresponding domains. How to conduct the training of entity linking models in this regard was a focus of this research.

2.3. Relation Extraction

After extracting the desired entities from unstructured text data through named-entity recognition and entity linking techniques, the relations between the entities can be obtained through relation extraction (RE). Traditional RE techniques usually adopt the fully supervised learning paradigm. To achieve good performance, these methods require a large pool of well-labeled training corpora [36,37]. However, labeling large corpora involves significant economic and time costs, which hinders the application of these traditional supervision methods in the real world. Mintz et al. [38] first proposed an RE algorithm based on distant supervision. Distant supervision assumes that sentences containing the same entity pairs will express the same relations. Obviously, this assumption would lead to a considerable number of mislabels if distant supervision was applied directly to each sentence. In this respect, several studies [39,40,41] have used the multiple-instance learning paradigm, which packages sentences with the same entity pairs and learnt bag representation. In recent years, as the advantages of the attention mechanism have been gradually recognized, researchers have used it in distant supervision-based RE algorithms [42,43,44]. Such methods apply a sentence-level attention mechanism to score all sentences in the bag, and their contribution to the bag representation can subsequently be evaluated. However, all of the above methods ignore the correlation between sentences, which leads to the loss of supervisory information. In addition, such methods based on a sentence-level attention mechanism are unable to extract global sentence information from the bag.

2.4. Natural Language Processing Applications in the Maritime

Natural language processing technology has been applied in the maritime field. Gan et al. [45] present a knowledge graph construction method for analyzing ship collision accidents, automating the extraction of key insights from 241 CMSA investigation reports using NLP and ontology. Liu et al. [46] introduces a BERT-MCNN deep learning model for extracting key information from Chinese maritime pollution prevention regulations, achieving high precision in knowledge graph formation, which aids PSC officers in efficient decision-making through a Neo4j-based application. Cao et al. [47] employ a Bayesian network model to analyze the impact of Accident Influential Factors (AIFs) on marine accident severity, identifying key factors and potential risks through data-driven analysis and machine learning, ultimately aiding in the prediction and prevention of maritime accidents for stakeholders. Kamal et al. [48] conduct a quantitative assessment of 418 marine accidents in the Istanbul Strait using a Tree Augmented Naive Bayes approach to identify probabilistic relationships between accident occurrence and influencing factors, revealing that smaller vessels are particularly prone to adrift accidents, and providing insights for authorities to enhance maritime safety measures.

3. Construction of Knowledge Graph Ontology Model for Pirate Events

3.1. Entities

The key step in designing a knowledge graph is to define an ontology model [11,12,13,14,15,16]. Inspired by the SEM model, a Pirate Event Model (PEM) was designed as an ontology model for pirate event-oriented knowledge graphs, which can be used to guide the progress of the proposed methods of named-entity recognition, entity linking, and relation extraction. It has four types of entities: (1) pem: Aggressor, i.e., the aggressor in the pirate event; (2) pem: Victim, i.e., victims in the pirate event; (3) pem: Location, i.e., where the pirate event occurred, such as a specific geographical name or latitude and longitude; and (4) pem: Date, i.e., the date of the pirate event. In addition, entity attribute constraints are set for the four types of entities: pem: AggressorTpye, pem: VictimTpye, pem: LocationTpye, and pem: DateTpye.

3.2. Association Relations

The associative relations of the PEM model can be divided into two categories: (1) binary relations between pem: Aggressor, pem: Victim, pem: Location, and pem: Data and (2) behavior relations between pem: Aggressor and pem: Victim. The details of the entity relations are shown in Table 1. In addition, Figure 1 visualizes the proposed ontology model.

4. Instance Layer Filling of Knowledge Graph for Pirate Events

We use pirate events, which are unstructured text reports, as model inputs, for instance, the layer filling of knowledge graph for pirate events.

4.1. Entity Extraction

This section extracts four types of entities: pem: Aggressor, pem: Victim, pem: Location, and pem: Date, from the given pirate event text. The BIO format [21,22] is used to label words in text data, i.e., B: Begin, representing the beginning of the named-entity; I: Inside, representing the inside of the named-entity; and O: Other, not representing any entity.
The BERT-BiLSTM-CRF model is used for NER. Bi-directional Encoder Representations from Transformers (BERT) is a self-supervised deep learning language model that pre-trains the text of multilayer bi-directional transformer coding structures mainly through the masking mechanism. For an input text sequence S = { w 1 , w 2 , , w N } , the BERT model outputs a coding sequence X = { x 1 , x 2 , , x N } N × d w in dimension d w . In the following steps, the BiLSTM model maps the coding sequence X = { x 1 , x 2 , , x N } to a tag score sequence P = { p 1 , p 2 , , p N } N × K , where p i K denotes the score of the K types of BIO-labeled entities corresponding to the word w i . Among them, the BIO-labeled entities are listed in Table 2.
Finally, the tag score sequence P is fed into the Conditional Random Field (CRF) layer, which is parameterized by the label transition matrix A K + 2 × K + 2 . A i , j represents the score transitioned from label i to label j . In addition, the matrix A contains two additional states: the beginning and end of the sequence.
For the input sequence X = { x 1 , x 2 , , x N } and the corresponding BIO-labeled sequence Y = { y 1 , y 2 , , y N } , y i { 1 , 2 , , K } , the evaluation scores for the CRF layer are given by Equation (1) [49].
S c o r e ( X , Y ) = i = 1 N A y i , y i + 1 + i = 1 N P i , y i
where A y i , y i + 1 denotes the score for transition from label y i to label y i + 1 , and P i , y i denotes the score for the probability that the i -th character is labeled as y i .
The model is trained to maximize the log probability of the correct tag sequence [49]:
log ( p ( Y | X ) ) = S c o r e ( X , Y ) log ( Y ˜ Y X e S c o r e ( X , Y ˜ ) )
where Y X denotes all possible BIO-labeling sequences.
Finally, Viterbi decoding [21,22] is used to find the highest scoring label sequence from the input sequences as a way to obtain the entity extraction results.

4.2. Entity Linking

Entity linking is the process of mapping entity mentions to real-world entities. Here, ambiguity can be considered a problem for the four types of entity mentions extracted from the event text, due to the different ways of expression. Therefore, entity linking based on context attention mechanism and distance learning is proposed in this paper, such that entity disambiguation can be achieved, and the quality of the knowledge graphs can be improved.
For entity mention m = { w h , , w k } in the event text S = { w 1 , w 2 , , w N } , E + is defined as a positive candidate entity set, which contains the correct entity e corresponding to the entity mention m ; E is the negative candidate entity set, which contains only the incorrect entities. For the construction of the candidate entity set E + , matching heuristics [40] are used to select entities from the knowledge base Wikipedia by measuring the overlap between entity mentions and the entity names. Specifically, all entities that contain entity mention m are listed, and these selected entities are subsequently sorted according to the order in the knowledge base. Finally, the top | E + | ranked entities are selected as the positive candidate entity set E + . This method is not particularly accurate in selecting the optimal entity, but the provided candidate entity set usually contains the correct entity e corresponding to the entity mention m . For the construction of the negative candidate entity set E , | E | candidate objects from the rest of the knowledge base are selected as the negative candidate entity set E .
For the entities e in the candidate entity set E + and E , the word2vec method [50] is used to map them into vectors. Secondly, the BERT model (mentioned in Section 4.1) is then used to map event text S = { w 1 , w 2 , , w N } into vector X = { x 1 , x 2 , , x N } , which includes mapping vector X m = { x h , , x t } corresponding to entity mention m . The association scores of each word in the event text S = { w 1 , w 2 , , w N } are calculated as a measure of its association with the entity e , as shown in Equation (3) [33]:
u ( w ) = max e E + , E x e T A e w x w
where A e w is the trainable parameterized diagonal matrix. The word set corresponding to the top R ranked words is selected based on the ranking of the association scores from highest to lowest [33]:
S ¯ = { w S | u ( w ) top R ( u ( v ) ) }
According to Equation (4), the attention weights are calculated as follows [33]:
β ( w ) = { exp [ u ( w ) ] v S ¯ exp [ u ( w ) ] w S ¯ 0 o t h e r w i s e
The representation vector X m a t t of entity mention m is calculated as follows:
X m a t t = w = h t β ( w ) x w
Finally, x e and X m a t t are fed into a feedforward neural network with two fully connected layers:
g ( e , m , S ) = F F N ( [ x e , X m a t t ] )
The feedforward neural network contains 128 hidden units and uses ReLU nonlinear activation.
During training, it is expected that, through the training model, at least one entity in E t r a i n + scores higher than any entity in E t r a i n . Max-margin loss is used to achieve the following:
L = ( m , S ) D t r a i n l ( m , S ) l ( m , S ) = [ max e E   g ( e , m , S ) + δ max e E +   g ( e , m , S ) ] +
where D t r a i n is the training set, and δ is the margin factor.
Based on Equation (8), the Adam optimization algorithm is used to minimize the model parameters. After the model training, for the test set sample ( m i , S i ) D t e s t , entity linking is achieved according to Equation (9):
e * = arg max e E t e s t +   g ( e , m i , S i )
In summary, the entity linking based on context attention mechanism and distance learning can be summarized as Algorithm 1.
Algorithm 1: Entity Linking Based on Context Attention Mechanism and Distance Learning.
Require:  Event   Text   S = { w 1 , w 2 , , w N } ;   Entity   Mention   m = { w h , , w k } ;   Positive   and   Negative   Candidate   entity   sets   E + ,   E
1. The entity e in E + ,   E   is   mapped   to   vector   x e using word2vec. Event text S   is   mapped   to   vector   X = { x 1 , x 2 , , x N } , and entity mention m   is   mapped   to   vector   X m = { x h , , x t } using the BERT model;
2. The association scores of each word in the event text S with the entity e are calculated according to Equation (3);
3. The word set S ¯ corresponding to the top R ranked words is selected based on the ranking of the association scores from highest to lowest;
4. Based on the word set S ¯ and Equation (5), the attention weights are calculated;
5. Calculate   the   representation   vector   X m a t t of entity mention m according to Equation (6);
6. Feed   x e ,   X m a t t into the feedforward neural network and output according to Equation (7);
7. The training loss function is calculated according to Equation (8), and the Adam optimization algorithm is used to minimize the model parameters;
8. After   model   training ,   for   the   test   set   sample   ( m i , S i ) D t e s t , entity linking is achieved according to Equation (9).

4.3. Relation Extraction

For the relations of the PEM model, the corresponding entities can be extracted from the text through NER and entity linking, which in turn can be used to directly build the binary relations between entity pem: Aggressor, pem: Victim and pem: Location, and pem: Date. Therefore, this section focuses on extracting the behavioral relations between entity pem: Aggressor and pem: Victim.
The text sequence of the containing entity pair of pem: Aggressor and pem: Victim C = { w 1 , w 2 , , w s } is transformed into a pre-trained coding sequence X = { x 1 , x 2 , , x s } by the BERT model. For word vector x i d w , their relative distance to the target entity in the sentence is calculated, i.e., d 1 , d 2 . Its positional embedding vector is subsequently obtained by consulting the positional embedding table [51] x i d 1 , x i d 2 d p , and it is spliced with the word vector x i as a word embedding representation q i d ( d = d w + 2 d p ) . Repeating the above steps, the text sequence C = { w 1 , w 2 , , w s } is transformed into the embedding representation C = { q 1 , q 2 , , q s } s × d . Subsequently, a piecewise convolutional neural network (PCNN) [52] is used to extract text sequence features. This consists of a convolutional layer and piecewise max pooling layer. The output c i , j from the convolutional layer is calculated as follows:
c i , j = w i q j : j + l 1
where w i l d denotes the weight matrix corresponding to the i -th convolutional kernel, l denotes the size of the convolutional kernel, and q j : j + l 1 denotes the splicing of word embedding representation from q j to q j + l 1 .
For the text sequence containing entity pair of pem: Aggressor and pem: Victim, the corresponding c i , j is divided into three segments { c i 1 , c i 2 , c i 3 } by the entity pairs. Max pooling is applied to each segment separately to produce three different output features p i = { p i 1 , p i 2 , p i 3 } . Afterwards, they are concatenated into a feature vector p 1 : n 3 n , where n represents the number of convolutional kernels. Finally, the hyperbolic tangent function s = tan   h ( p 1 : n ) , which is the activation function, is applied to the feature vector p 1 : n . The dimensions of the feature vector s are fixed and are no longer related to the length of the text sequence.
A set of text sequences containing the same entity pair is encapsulated into one bag. Suppose the bag i contains m i text sequences, which can be represented as { s 1 ( i ) , s 2 ( i ) , , s m i ( i ) } . The representation vector b i 3 n of the bag i is obtained by calculating the weighted sum of m i text sequences:
b i = j = 1 m i α j ( i ) s j ( i )
where α j ( i ) is the attention weight assigned to the j -th text sequence in the bag i , calculated as follows:
α j ( i ) = exp ( j = 1 , , m i j j s j ( i ) ( s j ( i ) ) T ) k = 1 m i exp ( j = 1 , , m i j j s k ( i ) ( s j ( i ) ) T )
Equation (12) shows that if the j -th text sequence in the bag i has a high similarity to the other text sequences in the bag, it will be assigned a larger attention weight when calculating the representation vector b i of the bag i . Finally, the representation vector { b 1 , b 2 , , b M } of all bags is calculated.
Bags containing the same relation label are packaged into a bag-group G = { b 1 , b 2 , , b m } m × 3 n . G a v g is first calculated to express the overall discriminative features of this bag-group.
G a v g = 1 3 n i = 1 3 n G i T G
where G i denotes the i -th row of the matrix G . Interactions are designed on this basis to calculate the similarity sim i between G a v g and b i :
sim i = G a v g b i T G a v g 2 b i T 2
Similarity sim i can be considered as the similarity of the overall discriminative features that the representation vector b i is able to represent the bag-group G . On this basis, we calculate the attention weight β i assigned to the representation vector b i .
β i = exp ( sim i ) j = 1 m exp ( sim j )
Then, we calculate the representation vector g 3 n of the bag-group G .
g = i = 1 m β i b i
Finally, a fully connected layer is added to the end of the above model to implement model training, using the negative log-likelihood as the training objective function.
J ( θ ) = i = 1 | g | log p ( r i | g i ; θ ) = i = 1 | g | log ( softmax ( tanh ( g i ) r i T + b ) )
where r 3 n denotes the label of the bag-group, | g | denotes the number of bag-groups, θ denotes the model parameters, and b denotes the bias parameters. Based on Equation (17), the Adam optimization algorithm is used to minimize the model parameters.
After the model training, relation prediction is performed using Equation (18):
k * = arg   max k ( softmax ( s r k T + b ) )
The proposed RE method can be summarized as Algorithm 2.
Algorithm 2: Proposed RE Method.
Require: Text sequence containing entity pairs of pem: Aggressor and pem: Victim   C = { w 1 , w 2 , , w s }
1. The   text   sequence   C = { w 1 , w 2 , , w s }   is   transformed   into   the   embedding   representation   C = { q 1 , q 2 , , q s } ;
2. PCNN is used to extract text sequence features s ;
3. m i text sequences containing the same entity pair are encapsulated in bag i { s 1 ( i ) , s 2 ( i ) , , s m i ( i ) } ;
4. According   to   Equations   ( 11 )   and   ( 12 ) ,   representation   vector   b i of the bag i   is   obtained   by   calculating   the   weighted   sum   of   m i   text   sequences ,   and   finally ,   the   representation   vectors   { b 1 , b 2 , , b M } of all bags are calculated;
5. Bags   containing   the   same   relation   label   are   packaged   into   a  bag-group  G = { b 1 , b 2 , , b m } ;
6. The   overall   discriminative   features   G a v g of the bag-group are calculated according to Equation (13);
7. Calculate   the   similarity   between   G a v g   and   b i according to Equation (14);
8. The   attention   weight   β i   assigned   to   the   representation   vector   b i is calculated according to Equation (15);
9. The representation vector g of the bag-group G is calculated according to Equation (16);
10. The training loss function is calculated according to Equation (17), and the Adam optimization algorithm is used to minimize the model parameters.
11. After the model training, relation prediction is performed using Equation (18)

5. Experiment and Analysis

5.1. Data Sources

The source of the soft data in this study was Worldwide Threats to Shipping (WWTTS), which contains a summary of piracy incidents and hostile acts against global commercial shipping since 1994. The data can be obtained in https://msi.nga.mil/Piracy, URL (accessed on 21 December 2023). The data are in English text and annotated automatically based on DeepKE (v2.2.7). Automatic annotation significantly reduces the need for manual labeling, thereby decreasing the time and cost. It also diminishes biases caused by human factors, improving the consistency and objectivity of data annotation. Additionally, training the proposed model with data that has been automatically annotated can lessen the dependency on large model of DeepKE, thereby promoting the development of the proposed method to a lightweight model and better adapt to domain-specific (pirate incident) entity and relation extraction tasks. In this study, 7580 events were extracted from the WWTTS records of the last 8 years (2015–2023) for experimental validation.

5.2. Performance Validation of NER

For the performance validation of BERT-BiLSTM-CRF-based NER, the event text was labeled automatically based on DeepKE using the BIO format according to Table 2, and a NER corpus was built. It was subsequently divided into training and test sets in a ratio of 8:2. The results of the experiments are shown in Table 3 and Figure 2.
From the results in Table 3 and Figure 2, it is clear that the BERT-BiLSTM-CRF model scores somewhat differently across entities. Compared to the entity pem: Location and pem: Data, the F1-score of entity pem: Aggressor and pem: Victim was lower. However, the F1-score of all types of entities reached more than 0.9, and the average recognition rate reached more than 0.94. Overall, the BERT-BiLSTM-CRF model performed well in NER.
Furthermore, several methods were used for comparison: (i) Hidden Markov Model (HMM); (ii) Machine Learning CRF model; (iii) Deep Learning BiLSTM model; and (iv) Deep Learning BiLSTM-CRF model. The above four named-entity algorithms were tested on the same dataset, and the test results were compared with the BERT-BiLSTM-CRF model. The experimental results of each model are shown in Table 4 and Figure 3.
As can be seen from Table 4 and Figure 3, the deep learning models outperformed the traditional machine learning models HMM and CRF in all respects. BiLSTM-CRF was obtained by adding sequence labeling CRF to the deep learning model, and it outperformed the BiLSTM model in all respects, suggesting that the CRF improves the performance of the model after taking into account the global label information of the sequences. After adding the BERT model to the deep learning BiLSTM-CRF model, F1-score was improved by 4.05%, which indicates that the proposed BERT-BiLSTM-CRF model has an obvious advantage in the NER of pirate events compared to other models. Thereby, the quality of the knowledge graph construction for pirate events can be improved.

5.3. Performance Validation of Entity Linking

Firstly, all the events were divided into training and test sets in the ratio of 8:2. For each entity mention in the event text of the training set, the matching heuristics mentioned in Section 4.2 were used to build the training candidate set: E t r a i n + and E t r a i n , where | E t r a i n + | = | E t r a i n | = 20 . For each entity mention in the event text of the test set, matching heuristics were also used to build the test candidate set E t e s t + , where | E t e s t + | = 20 . The difference is that E t e s t + includes entity mentions to correctly corresponding entities.
Table 5 and Figure 4 show the results of the test candidate set. The proposed model showed some variability in entity linking results for different entity types. Specifically, for the entity pem: Data and pem: Location, Precision, Recall, and F1-score of entity linking reached values above 0.93, especially for entity pem: Data, whose Precision, Recall, and F1-score of entity linking were all close to 1. This is due to the fact that these two types of entities are more standardized in their text representations, and there is rarely any ambiguity in their representations. For entity pem: Aggressor and pem: Victim, their Precision, Recall, and F1-score of entity linking were above 0.86, which are lower compared to those of pem: Data and pem: Location. This is due to the fact that these two types of entities have some diversity in textual representations, making the entity linking results subject to some error. However, overall, the average Precision, Recall, and F1-score of the proposed model reached more than 0.92. In particular, the proposed model achieved good test results by conducting entity linking training on the basis of unlabeled texts, which verifies the excellent performance of the proposed model of entity linking in specific vertical domains, consequently helping to improve quality of knowledge graph construction for pirate events.

5.4. Performance Validation of Relation Extraction

After NER and entity linking, the sentences containing entity pair pem: Aggressor and pem: Victim were selected. Next, the relation type was manually discriminated for a particular entity pair, and all sentences containing that entity pair were subsequently packaged into a bag; meanwhile, the above relation type is set as the relation type label of that bag, which serves as the basis for the experimental dataset. On this basis, bags containing the same relation label are packaged into a bag-group.
The performance of the proposed model in RE was evaluated, and the experimental results are shown in Figure 5 and Table 6. Little variation can be seen in the distribution of precision–recall curves for different behavior relations. In addition, their corresponding AUC values are more similar, and all of them reach values over 0.7, which proves that the proposed model is effective in categorizing all the behavior relations under study.
Subsequently, to verify the superior performance of the proposed model, it was compared with six existing state-of-the-art methods: self-attention mechanism and convolution and self-selective attention (SelfCON + SATT) [44], multicast [53], DSRE variational autoencoder (DSREVAE) [54], false-negative adversarial network (FAN) [55], contrastive instance learning (CIL) [56], and passage relation extraction (PARE) [57]. The results of the experiments are shown in Figure 6 and Table 7.
As shown in Figure 6, by observing the distribution of precision–recall curves, it can be seen that the vast majority of recalls of the proposed model are better than those of other models. The AUC values can quantitatively characterize the RE performance of different models. Table 7 shows the AUC values corresponding to the precision–recall curves of the proposed model and competitors. The AUC of the proposed model achieves the highest value and is 1.6–9.8% higher than those of other models, which further proves the superiority of the proposed model in RE, consequently helping to improve quality of knowledge graph construction for pirate events.

5.5. Knowledge Graph Generation for Pirate Events

In this section, the corresponding knowledge graphs were generated based on two simple pirate event texts. The texts of the pirate events are as follows:
  • Pirate event 1: On 30 July 2019, Somali sailors boarded the general cargo ship OYA underway near position 04:10N-006:59E, 15 nm southwest of Bonny Island. The pirates kidnapped five crewmen and escaped.
  • Pirate event 2: On 2 January 2019, Somali sailors attacked a small container ship M/V MANDY, near position 05:28N-002:21E, 55 nm south of Cotonou. Six crewmen were kidnapped.
Based on the event text, the knowledge triples are obtained as shown in Table 8 using the NER, entity linking, and relation extraction model proposed in this paper. The knowledge triples are stored through the neo4j database and visualized in Figure 7.
By comparing the pirate event text with the knowledge triples shown in Table 8 as well as the pirate event knowledge graph shown in Figure 7, it can be seen that the NER, entity linking, and relation extraction model proposed in this paper are able to efficiently extract entities and accurately establish the relations between entities, which further strongly suggests that the proposed model performs well in the construction of knowledge graphs for pirate events, especially in instance layer filling. Therefore, it is able to effectively generate high-quality knowledge graphs for pirate events.

6. Conclusions

In this paper, an automated construction method of knowledge graphs for pirate events was proposed and studied. Firstly, a Pirate Event Model (PEM) was designed as a knowledge graph ontology model for pirate events, on the basis of which the study focused on the knowledge triple extraction method for pirate events. For entity extraction, the BERT-BiLSTM-CRF model was proposed for NER. In the process of entity linking, to address the lack of large amounts of labeled text data, the method of distant learning was introduced for entity linking model training. In addition, using the context attention mechanism, the basis was proposed so that the words that are helpful for disambiguation decision can be selected, enhancing the performance of entity linking model training. For relation extraction, on the basis of the traditional distant supervision relation extraction, which is based on the sentence-level attention mechanism, the bag-level and group-level attention mechanism methods were additionally proposed to further enhance the performance of distant supervision relation extraction. The experimental results demonstrate the effectiveness of the proposed method for the automated construction of knowledge graphs for pirate events.
The knowledge graphs for pirate events achieve the deep knowledge mining of maritime threat information, which can support the knowledge representation of pirate events, and comprehend spatiotemporal activity query and backtracking, so as to achieve the effect of semantic enhancement, which has a certain application value. Furthermore, the knowledge graphs for pirate events are conducive to subsequent maritime threat reasoning for maritime threat situational awareness in general. In future work, the author will conduct maritime threat reasoning research based on the existing knowledge graphs for pirate events, including identifying dangerous areas, predicting threat trends, and detecting potential hostile behaviors, so as to comprehensively achieve maritime threat situational awareness and assist in the detection of illegal maritime activities.

Author Contributions

Conceptualization, C.X.; methodology, C.X.; software, C.X.; validation, C.X., L.Z. and Z.Z.; writing—original draft preparation, C.X.; writing—review and editing, C.X., L.Z. and Z.Z.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China under Grant 91538201, in part by Taishan Scholar Project of Shandong Province under Grant ts201511020, and in part by Project supported by Chinese National Key Laboratory of Science and Technology on Information System Security under Grant 6142111190404.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rosenberg, D. The political economy of piracy in the south China sea. Nav. War Coll. Rev. 2009, 62, 43–58. [Google Scholar]
  2. Hurlburt, K. The human cost of somali piracy. In Piracy at Sea; Springer: Berlin/Heidelberg, Germany, 2013; pp. 289–310. [Google Scholar]
  3. Jin, J.; Techera, E. Strengthening Universal Jurisdiction for Maritime Piracy Trials to Enhance a Sustainable Anti-Piracy Legal System for Community Interests. Sustainability 2021, 13, 7268. [Google Scholar] [CrossRef]
  4. Song, H.; Gi, I.; Ryu, J.; Kwon, Y.; Jeong, J. Production Planning Forecasting System Based on M5P Algorithms and Master Data in Manufacturing Processes. Appl. Sci. 2023, 13, 7829. [Google Scholar] [CrossRef]
  5. Núñez, R.C.; Samarakoon, B.; Premaratne, K.; Murthi, M.N. Hard and soft data fusion for joint tracking and classification/intent detection. In Proceedings of the 16th International Conference on Information Fusion, Istanbul, Turkey, 9–12 July 2013; IEEE: New York, NY, USA, 2013. [Google Scholar]
  6. Dragos, V.; Lerouvreur, X.; Gatepaille, S. A critical assessment of two methods for heterogeneous information fusion. In Proceedings of the 18th International Conference on Information Fusion (FUSION), Washington, DC, USA, 6–9 July 2015; IEEE: New York, NY, USA, 2015; pp. 42–49. [Google Scholar]
  7. Abirami, T.; Taghavi, E.; Tharmarasa, R.; Kirubarajan, T.; Boury-Brisset, A.-C. Fusing social network data with hard data. In Proceedings of the 18th International Conference on Information Fusion (FUSION), Washington, DC, USA, 6–9 July 2015. [Google Scholar]
  8. Burks, L.; Ahmed, N. Collaborative semantic data fusion with dynamically observable decision processes. In Proceedings of the 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019; IEEE: New York, NY, USA, 2019. [Google Scholar]
  9. Plachkov, A. Soft Data-Augmented Risk Assessment and Automated Course of Action Generation for Maritime Situational Awareness. Ph.D. Dissertation, University of Ottawa, Ottawa, ON, Canada, 2016. [Google Scholar]
  10. Buffett, S.; Cherry, C.; Dai, C.; Désilets, A.; Guo, H.; McDonald, D.; Su, J.; Tulpan, D. Arctic Maritime Awareness for Safety and Security (AMASS); Final Report; Canadian National Research Council: Ottawa, ON, Canada, 2017. [Google Scholar]
  11. Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, Vancouver, BC, Canada, 10–12 June 2008; Association for Computing Machinery: New York, NY, USA, 2008; pp. 1247–1250. [Google Scholar]
  12. Vrandečić, D.; Krötzsch, M. Wikidata: A free collaborative KnowledgeBase. Commun. ACM 2014, 57, 78–85. [Google Scholar] [CrossRef]
  13. Suchanek, F.M.; Kasneci, G.; Weikum, G. YAGO: A large ontology from wikipedia and wordnet. J. Web Semant. 2008, 6, 203–217. [Google Scholar] [CrossRef]
  14. Rotmensch, M.; Halpern, Y.; Tlimat, A.; Horng, S.; Sontag, D. Learning a health knowledge graph from electronic medical records. Sci. Rep. 2017, 7, 5994. [Google Scholar] [CrossRef]
  15. Chen, P.; Lu, Y.; Zheng, V.W.; Chen, X.; Yang, B. KnowEdu: A system to construct knowledge graph for education. IEEE Access 2018, 6, 31553–31563. [Google Scholar] [CrossRef]
  16. Miao, R.; Zhang, X.; Yan, H.; Chen, C. A dynamic financial knowledge graph based on reinforcement learning and transfer learning. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 5370–5378. [Google Scholar]
  17. Lagoze, C.; Hunter, J. The ABC ontology and model. J. Digit. Inf. 2001, 79, 160–176. [Google Scholar]
  18. Hage, W.R.V.; Malaisé, V.; Segers, R.; Hollink, L.; Schreiber, G. Design and Use of the Simple Event Model (SEM). J. Web Semant. 2011, 9, 128–136. [Google Scholar] [CrossRef]
  19. Zahila, M.N.; Noorhidawati, A.; Aspura, M.K.; Idaya, Y. Content extraction of historical Malay manuscripts based on Event Ontology Framework. Appl. Ontol. 2021, 16, 249–275. [Google Scholar] [CrossRef]
  20. Doerr, M.; Ore, C.E.; Stead, S. The CIDOC conceptual reference model: A new standard for knowledge sharing. In Tutorials, Posters, Panels and Industrial Contributions, Proceeding of the 26th International Conference on Conceptual Modelling, Auckland, New Zealand, 5–9 November 2007; Springer: Berlin/Heidelberg, Germany, 2007; Volume 83, pp. 51–56. [Google Scholar]
  21. Chiu, J.P.C.; Nichols, E. Named entity recognition with bidirectional LSTM-CNNs. Trans. Assoc. Comput. Linguist. 2016, 4, 357–370. [Google Scholar] [CrossRef]
  22. Liu, L.Y.; Shang, J.B.; Ren, X.; Xu, F.; Gui, H.; Peng, J.; Han, J. Empower sequence labeling with task-aware neural language model. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI Press: Menlo Park, CA, USA, 2018; pp. 5253–5260. [Google Scholar]
  23. Ratinov, L.; Roth, D.; Downey, D.; Anderson, M. Local and global algorithms for disambiguation to Wikipedia. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics, Portland, OR, USA, 19–24 June 2011; ACL: Stroudsburg, PA, USA, 2011; pp. 1375–1384. [Google Scholar]
  24. Varma, V.; Bysani, P.; Reddy, K.; Reddy, V.B.; Kovelamudi, S.; Vaddepally, S.R.; Nanduri, R.; Kumar, N.; Gsk, S.; Pingali, P. IIIT hyderabad in guided summarization and knowledge base population. In Proceedings of the Text Analysis Conference 2009, Gaithersburg, MD, USA, 16–17 November 2009; TAC: Washington DC, USA, 2009; pp. 213–222. [Google Scholar]
  25. Gottipati, S.; Jiang, J. Linking entities to a knowledge base with query expansion. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Scotland, UK, 27–31 July 2011; ACL: Stroudsburg, PA, USA, 2011; pp. 804–813. [Google Scholar]
  26. Zhang, W.; Sim, Y.C.; Su, J.; Tan, C.L. Entity linking with effective acronym expansion, instance selection and topic modeling. In Proceedings of the 22th International Joint Conference on Artificial Intelligence, Barcelona, Catalonia, Spain, 16–22 July 2011; pp. 1909–1914. [Google Scholar]
  27. Honnibal, M.; Dale, R. DAMSEL: The DSTO/Macquarie system for entity-linking. In Proceedings of the Theory and Applications of Categories, Gaithersburg, ML, USA, 16–17 November 2009; TAC: Washington DC, USA, 2009; pp. 1–4. [Google Scholar]
  28. Bunescu, R.; Paşca, M. Using encyclopedic knowledge for named entity disambiguation. In Proceedings of the 11th Conference of the European Chapter of the Association for Computational Linguistics, Trento, Italy, 3–7 April 2006; ACL: Stroudsburg, PA, USA, 2006; pp. 9–16. [Google Scholar]
  29. Han, X.P.; Sun, L. A generative entity-mention model for linking entities with knowledge base. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics, Portland, OR, USA, 19–24 June 2011; ACL: Stroudsburg, PA, USA, 2011; pp. 945–954. [Google Scholar]
  30. Francis-Landau, M.; Durrett, G.; Klein, D. Capturing semantic similarity for entity linking with convolutional neural networks. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; ACL: Stroudsburg, PA, USA, 2016; pp. 1256–1261. [Google Scholar]
  31. Wei, C.H.; Lee, K.; Leaman, R.; Lu, Z. Biomedical mention disambiguation using a deep learning approach. In Proceedings of the 10th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics, Niagara Falls, NY, USA, 7–10 September 2019; ACM: New Nork, NY, USA, 2019; pp. 307–313. [Google Scholar]
  32. Zuheros, C.; Tabik, S.; Valdivia, A.; Martínez-Cámara, E.; Herrera, F. Deep recurrent neural network for geographical entities disambiguation on social media data. Knowl. Based Syst. 2019, 173, 117–127. [Google Scholar] [CrossRef]
  33. Ganea, O.E.; Hofmann, T. Deep joint entity disambiguation with local neural attention. In Proceedings of the Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; ACL: Stroudsburg, PA, USA, 2017; pp. 2619–2629. [Google Scholar]
  34. Sun, Y.M.; Ji, Z.Z.; Lin, L.; Wang, X.; Tang, D. Entity disambiguation with memory network. Neurocomputing 2018, 275, 2367–2373. [Google Scholar] [CrossRef]
  35. Raiman, J.R.; Raiman, O.M. Deeptype: Multilingual entity linking by neural type system evolution. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI: Palo Alto, CA, USA, 2018; pp. 5406–5413. [Google Scholar]
  36. Bunescu, R.; Mooney, R.J. Subsequence kernels for relation extraction. In Proceedings of the NIPS 2005, Vancouver, BC, Canada, 5–8 December 2005; pp. 171–178. [Google Scholar]
  37. Culotta, A.; Sorensen, J. Dependency tree kernels for relation extraction. In Proceedings of the ACL 2005, Ann Arbor, MI, USA, 25–30 June 2005; pp. 423–429. [Google Scholar]
  38. Mintz, M.; Bills, S.; Snow, R.; Jurafsky, D. Distant supervision for relation extraction without labeled data. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, Singapore, 2–7 August 2009; pp. 1003–1011. [Google Scholar]
  39. Hoffmann, R.; Zhang, C.; Ling, X.; Zettlemoyer, L.; Weld, D.S. Knowledge-based weak supervision for information extraction of overlapping relations. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; pp. 541–550. [Google Scholar]
  40. Riedel, S.; Yao, L.; McCallum, A. Modeling relations and their mentions without labeled text. In Proceedings of the Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2010, Barcelona, Spain, 20–24 September 2010; pp. 148–163. [Google Scholar]
  41. Surdeanu, M.; Tibshirani, J.; Nallapati, R.; Manning, C.D. Multiinstance multi-label learning for relation extraction. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Jeju Island, Republic of Korea, 12–14 July 2012; pp. 455–465. [Google Scholar]
  42. Yuan, Y.; Liu, L.; Tang, S.; Zhang, Z.; Zhuang, Y.; Pu, S.; Wu, F.; Ren, X. Cross-relation cross-bag attention for distantly-supervised relation extraction. In Proceedings of the AAAI 2018, New Orleans, LA, USA, 2–7 February 2018; pp. 419–426. [Google Scholar]
  43. Christou, D.; Tsoumakas, G. Improving distantly-supervised relation extraction through BERT-based label and instance embeddings. IEEE Access 2021, 9, 62574–62582. [Google Scholar] [CrossRef]
  44. Zhou, Y.; Pan, L.; Bai, C.; Luo, S.; Wu, Z. Self-selective attention using correlation between instances for distant supervision relation extraction. Neural Netw. 2021, 142, 213–220. [Google Scholar] [CrossRef] [PubMed]
  45. Gan, L.; Ye, B.; Huang, Z.; Xu, Y.; Chen, Q.; Shu, Y. Knowledge graph construction based on ship collision accident reports to improve maritime traffic safety. Ocean. Coast. Manag. 2023, 240, 106660. [Google Scholar] [CrossRef]
  46. Liu, C.; Zhang, X.; Xu, Y.; Xiang, B.; Gan, L.; Shu, Y. Knowledge graph for maritime pollution regulations based on deep learning methods. Ocean. Coast. Manag. 2023, 242, 106679. [Google Scholar] [CrossRef]
  47. Cao, Y.; Wang, X.; Wang, Y.; Fan, S.; Wang, H.; Yang, Z.; Liu, Z.; Wang, J.; Shi, R. Analysis of factors affecting the severity of marine accidents using a data-driven Bayesian network. Ocean. Eng. 2023, 269, 113563. [Google Scholar] [CrossRef]
  48. Kamal, B.; Cakir, E. Data-driven Bayes approach on marine accidents occurring in Istanbul strait. Appl. Ocean. Res. 2022, 123, 103180. [Google Scholar] [CrossRef]
  49. Lample, G.; Ballesteros, M.; Subramanian, S.; Kawakami, K.; Dyer, C. Neural architectures for named entity recognition. Computing Research Repository. arXiv 2016, arXiv:1603.01360. [Google Scholar]
  50. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 3111–3119. [Google Scholar]
  51. Zeng, D.; Liu, K.; Lai, S.; Zhou, G.; Zhao, J. Relation classification via convolutional deep neural network. In Proceedings of the COLING 2014, Dublin, Ireland, 23–29 August 2014; pp. 2335–2344. [Google Scholar]
  52. Zeng, D.; Liu, K.; Chen, Y.; Zhao, J. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of the EMNLP 2015, Lisbon, Portugal, 17–21 September 2015. [Google Scholar]
  53. Chen, T.; Shi, H.; Liu, L.; Tang, S.; Shao, J.; Chen, Z.; Zhuang, Y. Empower distantly supervised relation extraction with collaborative adversarial training. In Proceedings of the AAAI Conference on Artificial Intelligence 2021, Virtual, 2–9 February 2021; pp. 12675–12682. [Google Scholar]
  54. Christopoulou, F.; Miwa, M.; Ananiadou, S. Distantly supervised relation extraction with sentence reconstruction and knowledge base priors. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 11–26. [Google Scholar]
  55. Hao, K.; Yu, B.; Hu, W. Knowing false negatives: An adversarial training method for distantly supervised relation extraction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 9661–9672. [Google Scholar]
  56. Chen, T.; Shi, H.; Tang, S.; Chen, Z.; Wu, F.; Zhuang, Y. CIL: Contrastive instance learning framework for distantly supervised relation extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual, 1–6 August 2021; pp. 6191–6200. [Google Scholar]
  57. Rathore, V.; Badola, K.; Singla, P.; Mausam. PARE: A simple and strong baseline for monolingual and multilingual distantly supervised relation extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; pp. 340–354. [Google Scholar]
Figure 1. Ontology model for pirate events.
Figure 1. Ontology model for pirate events.
Applsci 14 06482 g001
Figure 2. NER results.
Figure 2. NER results.
Applsci 14 06482 g002
Figure 3. Comparison results of NER of each model.
Figure 3. Comparison results of NER of each model.
Applsci 14 06482 g003
Figure 4. Entity linking results for different entity types.
Figure 4. Entity linking results for different entity types.
Applsci 14 06482 g004
Figure 5. Precision–recall curves for the behavior relations extracted using our model.
Figure 5. Precision–recall curves for the behavior relations extracted using our model.
Applsci 14 06482 g005
Figure 6. Precision–recall curves of our model and competitors.
Figure 6. Precision–recall curves of our model and competitors.
Applsci 14 06482 g006
Figure 7. Knowledge graph for pirate event (partial).
Figure 7. Knowledge graph for pirate event (partial).
Applsci 14 06482 g007
Table 1. Entity relations of the PEM model.
Table 1. Entity relations of the PEM model.
RelationSubject(s)Object(s)
pem: Aggressor-LocationAggressorLocation
pem: Aggressor-DateAggressorDate
pem: Victim-LocationVictimLocation
pem: Victim-DateVictimDate
pem: AttackingAggressorVictim
pem: BoardingAggressorVictim
pem: Firing uponAggressorVictim
pem: KidnappingAggressorVictim
pem: HijackingAggressorVictim
pem: RobbingAggressorVictim
pem: Suspiciously approachingAggressorVictim
Table 2. BIO-labeled entities.
Table 2. BIO-labeled entities.
Serial NumberLabel
1B-Aggressor
2I-Aggressor
3B-Victim
4I-Victim
5B-Location
6I-Location
7B-Data
8I-Data
9O
Table 3. NER results.
Table 3. NER results.
Serial NumberLabelPrecisionRecallF1-Score
1B-Aggressor0.90450.92250.9134
2I-Aggressor0.91600.90470.9103
3B-Victim0.90220.91180.9070
4I-Victim0.92390.92030.9221
5B-Location0.95240.95240.9524
6I-Location0.94160.94840.9450
7B-Data0.98750.99250.9900
8I-Data0.99170.99130.9915
9O0.94410.92870.9363
10Overall0.94040.94140.9409
Table 4. Comparison results of NER of each model.
Table 4. Comparison results of NER of each model.
ModelPrecisionRecallF1-Score
HMM model0.81160.79610.8037
CRF model0.84360.83130.8374
BiLSTM model0.86710.87840.8664
BiLSTM-CRF model0.90290.89790.9004
BERT-BiLSTM-CRF model0.94040.94140.9409
Table 5. Entity linking results for different entity types.
Table 5. Entity linking results for different entity types.
Entity TypesPrecisionRecallF1-Score
pem: Aggressor0.89650.90170.8991
pem: Victim0.87390.86930.8716
pem: Location0.93620.95400.9450
pem: Data0.99120.99250.9919
Average0.92200.92440.9269
Table 6. AUC of seven precision–recall curves for the behavior relations.
Table 6. AUC of seven precision–recall curves for the behavior relations.
Relation TypesAUC
pem: Attacking0.733
pem: Boarding0.719
pem: Firing upon0.722
pem: Kidnapping0.754
pem: Hijacking0.744
pem: Robbing0.707
pem: Suspiciously approaching0.746
Table 7. AUC of seven precision–recall curves of our model and competitors.
Table 7. AUC of seven precision–recall curves of our model and competitors.
ModelAUC
SelfCON + SATT0.672
Multicast0.681
DSREVAE0.707
FAN0.656
CIL0.688
PARE0.625
Ours0.723
Table 8. Knowledge triples extracted based on pirate events 1 and 2.
Table 8. Knowledge triples extracted based on pirate events 1 and 2.
SubjectObjectRelation
Somali sailorscargo ship OYApem: boarding
Somali sailorsfive crewmenpem: kidnapping
Somali sailors30 July 2019pem: Aggressor-Date
Somali sailors04: 10N-006: 59Epem: Aggressor-Location
cargo ship OYA30 July 2019pem: Victim-Date
cargo ship OYA04: 10N-006: 59Epem: Victim-Location
five crewmen30 July 2019pem: Victim-Date
five crewmen04: 10N-006: 59Epem: Victim-Location
Somali sailorscontainer ship M/V MANDYpem: attacking
Somali sailorssix crewmenpem: kidnapping
Somali sailors2 January 2019pem: Aggressor-Date
Somali sailors05: 28N-002:21Epem: Aggressor-Location
container ship M/V MANDY2 January 0219pem: Victim-Date
container ship M/V MANDY05: 28N-002:21Epem: Victim-Location
six crewmen2 January 2019pem: Victim-Date
six crewmen05: 28N-002:21Epem: Victim-Location
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, C.; Zhong, Z.; Zhang, L. Automated Construction Method of Knowledge Graphs for Pirate Events. Appl. Sci. 2024, 14, 6482. https://doi.org/10.3390/app14156482

AMA Style

Xie C, Zhong Z, Zhang L. Automated Construction Method of Knowledge Graphs for Pirate Events. Applied Sciences. 2024; 14(15):6482. https://doi.org/10.3390/app14156482

Chicago/Turabian Style

Xie, Cunxiang, Zhaogen Zhong, and Limin Zhang. 2024. "Automated Construction Method of Knowledge Graphs for Pirate Events" Applied Sciences 14, no. 15: 6482. https://doi.org/10.3390/app14156482

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop