Next Article in Journal
A Study of Weather-Image Classification Combining VIT and a Dual Enhanced-Attention Module
Previous Article in Journal
A Tile-Based, Adaptable Boost Converter with Fast Transient Response and Small Voltage Ripples in 40 nm CMOS Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Language Entity Alignment Based on Dual-Relation Graph and Neighbor Entity Screening

School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050013, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(5), 1211; https://doi.org/10.3390/electronics12051211
Submission received: 14 January 2023 / Revised: 20 February 2023 / Accepted: 1 March 2023 / Published: 3 March 2023
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Graph convolutional network-based methods have become mainstream for cross-language entity alignment. The graph convolutional network has multi-order characteristics that not only process data more conveniently but also reduce the interference of noise effectively. Although the existing methods have achieved good results for the task of cross-language entity alignment, they have often overlooked the same entity names in the real corpus, resulting in an entity-matching result that was not ideal. Therefore, this study proposed a neighboring-entity-screening rule by combining the entity name and the attribute (NENA) to reduce the influence of these issues. We used the NENA-screening rule to filter and delete redundant equivalent entities and to construct a dual-relation graph as auxiliary evidence for scenarios when the attribute information may be insufficient.This study adopted a graph convolutional network in order to embed knowledge graphs and entity names into a unified vector space, and then a down-sampling method was used to extract the neighboring entities of each entity, thus forming sub-graphs of the two knowledge graphs. We embedded the sub-graphs into the GCN, as the new input, and then we used a cross-graph-matching module to finally achieve alignment. Our results on the DBP15K dataset showed that our approach significantly improved the overall entity alignment.On the sub-dataset ZH-EN of DBP15K, the value of Hits@1 improved by 1.38%, as compared to the best approach mentioned in this paper, and it was useful for the construction and completion of the open knowledge graph.

1. Introduction

A knowledge graph is a graph-based data structure. Specifically, a knowledge graph is a modern theory that combines theories and methods of applied mathematics, graphics, information visualization technology, and other disciplines with metrological citation analysis and other methods, and it uses a visualization graph to display the complete knowledge framework of the disciplines to achieve the purpose of multidisciplinary integration. Knowledge graphs encode unstructured data into structured triples and provide an effective scheme to describe the complex relations between concepts and entities. With the continuous deepening and improvement of the knowledge-graph research, knowledge graphs have been widely applied to various tasks related to artificial intelligence, such as intelligent question-answering, intelligent recommendations, and semantic searches. In recent years, knowledge graphs have been constructed continuously, and the number and scale of the knowledge graphs have increased rapidly. Extensive data sources, uneven quality, and diverse languages are the main factors contributing to the diversity and heterogeneity of the knowledge-graph information. Therefore, to solve these issues, cross-language tasks have been invented. Cross-linguistic research are studies developed from contrastive linguistics. Cross-linguistic tasks are based on the large-scale corpora of two or more languages, and the research objects extend from traditional phonemes, vocabulary, and syntax, to pragmatics, rhetorical structures, and even language using in special scenarios. In recent years, in order to support multi-lingual applications, integrating cross-language knowledge graphs has become a necessary step in the knowledge graph research, and entity alignment is critical for this step.
The goal of the entity alignment task [1,2,3,4] has been to find entities that represent the same thing in the real world in multiple knowledge graphs. Formally, it is necessary to establish the relation between these entities and then connect the multi-source and heterogeneous knowledge graphs in order to form a unified knowledge graph, which provides support for various applications based on knowledge graphs. Cross-language entity alignment task refers to the task of finding matching entity-pairs in different languages in a multi-language scenario. For example, the knowledge graphs of different languages are shown in Figure 1. In shown in Figure 1, ZH-KG1 is a Chinese knowledge graph, and EN-KG2 is an English knowledge graph. ZH-KG1 and EN-KG2 not only have different language forms, but they also have different structures. The entities, relations, and attributes in the two knowledge graphs were represented differently by each language.Even if two entities belong to different knowledge graphs, it does not mean that the two entities refer to different things. The “Husak” entity in the English knowledge graph is different from the “胡萨克” entity in the Chinese knowledge graph. However, since they refer to and represent the same concept in the real world, they were regarded as align-able entity-pairs.
In order to achieve the integration of cross-language knowledge graphs more accurately and efficiently on the basis of the classic model TransE [5,6], researchers have proposed approaches such as TransD [7], TransR [8], MtransE [1] and other representative cross-language entity alignment methods. These approaches have made positive contributions to the research of cross-language entity alignment. However, they have relied on machine translations to eliminate language barriers, and therefore, these methods often result in uneven translation quality and lack context to employ as a reference.
At present, most approaches have used a graph convolutional network (GCN), which can greatly reduce the influence of noise caused by language differences. GCN is a graph domain-information-processing method based on deep learning. Because of its good performance and interpretability, it has become a widely used graph analysis method. It has efficiently combined entity-adjacent node information and learning structures to encode knowledge graphs.It is worth noting that although GCN has been widely used for the task of cross-language entity alignment at present, the existing methods have not considered that entities in the real corpora may have the same name but refer to different things. Insufficient supporting materials has easily led to a significant amount of incorrectly matched results.
After studying and summarizing previous research, this study then applied the multi-order characteristics of graph convolutional networks and then used the entity attribute information as evidence to screen high-quality triples. In order to reduce the influence caused by insufficient attribute information, we constructed a dual-relation graph [9,10]. Meanwhile, the central entity was defined as the head entity that occurred at least once in all triples.Then, the cross-language entity alignment was completed by neighboring-sampling and similarity calculations. Our main contributions were the following:
  • We proposed a neighboring-entity-screening rule that used the entity name and its attributes as the main evidence for screening. In the neighboring entity set of the central entity, we identified all entities with the same entity name as the central entity. Then, the attribute information of the entities was used as the main evidence to judge whether two entities with the same entity name referred to the same concept in the reality corpus.
  • Dual-relation graph was used to make full use of relations. This paper used the dual-relation graph not only to strengthen the role of entity relations but also to avoid the impact of insufficient attribute information and to reduce the errors in the entity screening process.
  • A cross-language entity alignment method based on neighboring-entity screening and a dual-relation graph was proposed. We used a graph convolutional network, combined with neighboring-entity-screening rules and a dual-relation graph to realize cross-language entity alignment and achieve excellent results on public datasets, such as DBP15K, which proved the importance of entity screening.

2. Related Work

The emergence of a large number of heterogeneous, multifarious, and complementary knowledge graphs have emphasized the importance of integrating multi-lingual knowledge graphs to achieve comprehensive and complete knowledge graphs. Entity alignment is a key technology to achieve this goal. The existing entity alignment methods can be divided into two categories, traditional entity alignment methods and representation learning-based entity alignment methods.

2.1. Traditional Entity Alignment Methods

Traditional entity alignment methods have mostly applied used feature-engineering techniques, such as string similarity and rule mining, to complete alignment tasks. For example, Ngomo et al. [11] proposed using triangle inequality to calculate the entity similarity, and NIU [12] optimized artificially defined entity-matching rules by using the maximization expectation algorithm.
Traditional entity alignment methods have manually constructed a large amount of feature information and designed attribute-matching rules. Datasets in different domain and attribute categories have required different attribute-matching rules. Therefore, designing different attribute-matching rules for the attribute categories contained in datasets in different domains have been crucial for realizing traditional entity alignment technology. However, these approaches not only require a significant amount of manpower but are also difficult to migrate between multiple datasets.

2.2. Entity Alignment Based on Representation Learning

With the rapid development of entity alignment, traditional approaches have been unable to meet the requirements of high-efficiency and high-accuracy matching. After significant exploration, some researchers found that approaches based on representation learning were more suitable for the task of entity alignment, and the results have been significantly better than the traditional approaches. Entity alignment based on representation learning has had good results for modeling the semantic information of the knowledge graphs. Many entity alignment algorithms have been implemented based on graph representation learning, which encodes all entities in a knowledge graph in order to extract the graph features of the entities. Entity alignment based on representation learning has been divided into two categories, namely translation model-based methods [2,13,14] and graph convolutional network (GCN)-based methods [2,3,4,15]. It is worth noting that most of the following methods have used Chinese–English, French–English, and Japanese–English datasets.

2.2.1. Entity Alignment Based on Translation Model

Translation-based approaches have originated from cross-language word-embedding. Therefore, they also have a core assumption that the entity-embedding of different knowledge graphs have similar distributions, similar to the word-embedding of different languages [16,17]. Translation-based methods usually include two modules: a translation module (TM) and an alignment module (AM) [16], as shown in Figure 2. TM usually refers to translation models, such as TransE [5]. AM is usually divided into two different alignment modules: mapping and sharing. The mapping approach embeds different knowledge graphs into a unified vector space through a linear transformation matrix. For example, KDCoE [18] minimized the distances between the pre-aligned pairs by optimizing one or two linear transformation matrices. The sharing approach embeds different KGs into a unified vector space by allowing each pre-aligned pair to directly share the same embedding [17]. For example, MTransE [5] proposed to minimize the equation e 3 e 5 for each pre-aligned entity-pair.
Representation learning approaches based on translation models are able to learn vector representations of entities in low-dimensional dense spaces. Inspired by Word2Vec [19], TransE [5,6] interprets relations as head-to-tail translations. The representation learning approach, represented by TransE, was originally used for the structural embedding of entities in a single knowledge graph. TransE’s extension models TransD [7], TransH [20], etc., have performed entity alignment and entity inference in a unified vector space on the basis of learning entity-embedding representations. The closer the distance between the entity-pair in the embedding space, the higher the semantic similarity of the two entities, and the greater the possibility that the entity-pair could be aligned between knowledge graphs.
Inspired by the above approaches, many researchers [1,9,21,22,23] have explored improvements. MtransE [1] employed the TransE model to encode the structure of the knowledge graph into a specific language vector space. However, the model relied on a large number of previously aligned entities, which were often difficult to obtain.In order to reduce the dependence on a priori aligned entities, IPTransE [22] and BootsTransE [23] adopted a semi-supervised iterative strategy, which predicted new alignment entities with a small number of previously aligned entities, used them to augment the previously aligned entity set, and iterated repeatedly until no new alignment entities appeared. However, the iterative process often resulted in error propagation.

2.2.2. Entity Alignment Based on Graph Convolutional Networks

The entity alignment methods based on translation model used the sum of the head entity vector and the relation vector to predict the tail entity vector, but in actual situations, the links between entities and relations may be more complicated. For example, in the process of using this kind of approach for alignment tasks, a significant amount of noise may be generated, which may then affect the accuracy of alignment. The discovery of this issue introduced new challenges. More effective alignment has become the focus to overcome this challenge.In recent years, many researchers have introduced graph convolutional networks (GCNs). The concept of these approaches was derived from the utilization of the global information modeling of graphs. GCN is a multi-layer neural network that operates directly on a graph, which induces node-embedding vectors based on the attributes of its neighborhoods [24].
Approaches based on GCN, such as [9,10,25,26,27,28,29,30,31,32,33,34,35,36,37,38], have enhanced the embedding of entities and their neighbors’ information, and they only require a small number of aligned seed-pairs to transfer similar information to the whole graph. As compared to the traditional entity-alignment methods, approaches based on GCNs not only require relatively less human involvement in the process of feature construction, but also such approaches could be extended to large knowledge graphs. GCN applications have typically embedded the data to be processed into a unified vector space [39,40].
JAPE [9] introduced two embedding modules, namely structure embedding (SE) and attribute embedding (AE), which jointly embedded the structures of two knowledge graphs into a unified vector space, and then they used the attribute correlation in the knowledge graph for further improvements. As compared to the traditional alignment approach, this has provided significant improvements, but it has not been able to effectively address the one-versus-many issue of structural embedding.In response to this issue, Wang [25] proposed the GCN-align model. It embedded the entities of each language into a unified vector space and then learned the embeddings from the structure and attribute information of the entities. This combination resulted in precise alignment results.Inspired by Wang [25], Fan [30] proposed the assignment of different weight values to different attributes on the basis of the GCN-align. This method effectively utilized attribute information and reduced the computational complexity in the matching process, resulting in a more precise alignment.
The introduction and application of the GCN has promoted the rapid development of entity alignment task. In addition to the aforementioned approaches related to the GCN, the RDGCN [10] was a novel relation-aware dual-graph convolutional network, and it integrated relational information by exploiting the close interaction between the knowledge graph and its dual-relations and then further captured neighboring structures to better understand the entity representations. However, the RDGCN ignored the structural differences between different knowledge graphs. The multi-channel graph neural network model (MuGNN) [32] adopted the judgment and the comparison of the knowledge graphs to coordinate the structural differences between different knowledge graphs and used graph-based models to better utilize the seed-pair information. Although MUGNN reduced the impact of the structural differences between the different knowledge graphs, it was unable to fully utilize the attribute information of the entities. Tam [34] proposed an end-to-end and unsupervised cross-language knowledge-graph entity alignment framework, which captured the relation-based correlations between entities by exploiting the multi-order properties of the GCN and then fused different types of information to take advantage of the richness of the knowledge graph. After Tam [34], Wu et al. [41] proposed a new entity alignment framework (neighborhood matching network, NMN) to address the challenge of structural heterogeneity. The NMN estimated the similarities between entities to capture both the topological structures and the neighborhood differences. It provided two innovative components that enabled a better learning representation for entity alignments. Such strategies allowed the NMN to efficiently construct matching-oriented entity representations while ignoring noisy neighbors that would have a negative impact on the alignment results [41].
These approaches have made useful contributions towards advancing entity alignment, and at the same time, they have opened up new directions for the application of GCNs in entity alignment tasks. However, these works have overlooked the existence of triples of the same entity name in the real corpus. The entities in these triples contain a significant amount of relation and attribute information, and this has lead to inexact matches between similar entities. Therefore, based on the above research and the NMN [41] model, we used the multi-order properties of the GCN and combined the dual-relations with neighboring-entity-screening to achieve cross-language entity alignment.

3. Methods

3.1. Problem Definition

Entity alignment is the process of integrating entities that mean the same thing in the real world, according to different knowledge bases, into a unified knowledge base. Although many entity alignment methods for cross-language knowledge graphs have emerged, there is still a significant amount of advancement required based on the matching scores of similar entities, whether in specific or open domains.Therefore, how to effectively improve the accuracy of similar entities is a hotspot in current alignment-task research. The following questions addressed in this paper concerned cross-language entity alignment: (1) how to exploit the advantages of the GCN in cross-language entity alignment; (2) how to use the entities, relations, and attributes more effectively; and (3) how to handle entities with the same entity name. The problem definition diagram is shown in Figure 3. Without the loss of generality, we implemented the task of entity alignment between the two knowledge graphs KG1 and KG2, based on a set of pre-aligned equivalent entities. Our goal was to identify the equivalent entity-pair between KG1 and KG2.

3.2. Symbol Definition

Knowledge graphs represent knowledge about real-world entities as triples. According to the experimental requirements, two kinds of triples in the knowledge graph were formally defined, which were relation triples and attribute triples. A relation triple represented the relation between two entities and consisted of a head entity, a relation, and a tail entity (which could also be understood as a subject, a predicate, and an object). Its basic definition format was ( h e a d , r e l a t i o n , t a i l ) . For example, the relation triple < M o u n t F u j i , i s , m o u n t a i n >, from which we could determine the relation between Mount Fuji and mountain. The attribute triple described the attribute characteristics of the entity, and its basic definition format was ( e n t i t y , a t t r i b u t e , v a l u e ) . For convenience, we denoted the knowledge graph as K G = ( E , R , T ) , where E, R, and T represent the entity set, the relation set, and the triple set, respectively. For the convenience of representation, in form, we denoted the two heterogeneous knowledge graphs as K G 1 = ( E 1 , R 1 , T 1 ) and K G 2 = ( E 2 , R 2 , T 2 ) , where E i , R i , and T i represent the entity set, the relation set, and the triple set in K G i ( i = 1 , 2 ) , respectively. Taking Figure 3 as an example, we obtained two knowledge graphs K G 1 = ( E 1 , R 1 , T 1 ) and K G 2 = ( E 2 , R 2 , T 2 ) . As shown in Figure 3, we found that E 1 and E 2 were composed of five entities, namely E 1 = { 1 , 2 , 3 , 4 , 5 } and E 2 = { 11 , 12 , 13 , 14 , 15 } . Variables R 1 and R 2 were composed of relations connecting entities in E 1 and E 2 , respectively, and T 1 and T 2 were composed of triples formed by entities in KG1 and KG2 that have some relations, respectively.
Definition 1.
We combined the two knowledge graphs to form a large knowledge graph and referred to this large knowledge graph as the primal graph (PG).
Definition 2.
The dual-relation graph (DRG) was an un-directed graph, which assumed that relations sharing the same head or tail entity were a vertex and the shared edge’s weight value as the correlation degree of the two relations.The dual-relation graph was then used to strengthen the role of relation so that it could be fully utilized and serve as auxiliary evidence for entity screening.
Definition 3.
Two entities were said to be equivalent entities if they had the same entity name and an inclusion relation between their attributes.

3.3. Entity Alignment Based on Entity Screening and Dual-Relation Graph

In a practical situation, the heterogeneity and the noise of common neighboring entities in the knowledge graphs greatly reduce the potential of entity alignments and make it difficult to obtain useful information. We used the two knowledge graphs in Figure 1 for this example. In EN-KG2, entity “Husak” and entity “Havel” had the same neighboring entity “president”, and we assumed that there was no relation “successor” between “Husak” and “Havel”. In entity alignment, due to the heterogeneity of neighboring entities, it was difficult for us to determine whether neighboring entities could be aligned due to insufficient evidence, and this process would inevitably produce noise, precluding the extraction of useful information to prove that entity “哈维尔” and entity “Havel” could be aligned.In fact, we knew that “电影” and “film” could also be aligned, but as opposed to “胡萨克” and “Husak”, the former had a significantly lower number of neighboring entities and could provide relatively little useful information, making it difficult or even impossible to align the former. The entity’s nearest neighboring entity played an important role in determining whether the entity should be aligned with other entities. However, not all nearest neighboring entities positively contributed to entity alignment. Therefore, we needed to select informative entities as alignment samples. The whole research procedure is shown in Figure 4.
Using the entities that were part of the two knowledge graphs, in Figure 1, for this example, the steps of our method were as follows:
Step 1: Dual-relation graph (DRG) construction. The dual-relation graph used the entity relation as vertex, and an edge was established between the two relations if they shared the same head or tail entity. In order to ensure the obtained dual-relation graph fully included the source and target knowledge graphs, a graph attention network (GAT) was used to ensure the dual-relation graph interacted with the PG for several rounds. The dotted lines in Figure 4 represent multiple rounds of interaction between the PG and the DRG.
Step 2: Neighboring-entity screening. A neighboring-entity-screening rule based on entity name and attribute was proposed. The attribute information of the entity was used to judge whether the neighboring entity was equivalent to the corresponding central entity. If two entities had the same entity name and there was an inclusion relation for their attributes, we considered the two entities to be equivalent. When the attribute information was insufficient, we used the dual diagram as auxiliary evidence for neighboring-entity screening.
Step 3: Knowledge embedding. We embedded the primal graph structure with the names of the entities obtained in step 2 into a unified vector space, and then the resulting entity feature representations in the primal graph were fed to the GCN layers with highway gates in order to capture the neighboring structural information. In order to strengthen the role of the relation between entities and their closer connections, we embedded the primal graph into the GCN.
Step 4: Entity neighbor-sampling. In order to facilitate the matching of the corresponding entities of the two knowledge graphs, we calculated the probability of each neighboring entity being sampled and then use a down-sampling method to obtain the neighboring entity with higher probability ranking to finally obtain the two knowledge sub-graphs.
Step 5: Entity alignment. We embedded the sub-graph obtained in step 4 into the GCN, used the cross-graph matching module [41] for the neighboring-entity matching, and then aggregated the matched neighboring-entity information. Finally, we further calculated the distance between the two central entities to achieve alignment.

3.3.1. Dual Relation Graph Construction

Formally, this paper defined a primal graph (PG) K G p = ( E p , R p , T p ) , where E p , R p , T p represent the entity set, the relation set, and the triple set of the PG, respectively. For a given arbitrary K G p , its DRG was defined as K G d = ( E d , W d , T d ) , where E d represents the set of entities with relations in the PG as nodes, W d represents the correlation degree of the two relations, and T d represents a correlation degree triple of the format ( E d 1 , W d , E d 2 ) . If two different relations were connected by the same head entity (or tail entity), we built an edge between the two relations, and these edges formed the set W d . For each type of relation r in K G p , there was a vertex e d ( e d R p ) in K G d that corresponded to r. In order to ensure the DRG more accurately expressed the relation between different e d in K G d , based on the possibility that the two relations in K G d could share a similar head or tail entity in K G p , the Jaccard similarity coefficient was used to set the weight value w i j r for each edge in W d . Inspired by the literature [24], the calculation of weight w i j r was:
H e ( r i , r j ) = H e i H e j H e i H e j , T e ( r i , r j ) = T e i T e j T e i T e j
w i j r = H e ( r i , r j ) + T e ( r i , r j )
where H e i and T e i represent the set of head and tail entities connected with relation r i ( r i R p ), respectively. The expressions H e ( r i , r j ) and T e ( r i , r j ) , respectively, represent the probability that relations r i and r j share the same head entity and tail entity.
In order to facilitate this understanding, this study provided an example of constructing a dual-relation graph, as shown in Figure 5. Since it was necessary to judge whether two relations shared the same head entity (or tail entity) when constructing a dual-relation graph, three possible scenarios are summarized in Figure 5. To illustrate the calculation of weight values, we used relations “career” ( r 1 ) and “serve as” ( r 2 ) in Figure 5 as an example. The head and tail entities connected by “career” were { H u s a k , S k a d a , H a v e l } and { p o l i t i c i a n , d i r e c t o r } , respectively, and those connected by “serve as” were { H u s a k , H a v e l } and { p r e s i d e n t } , respectively. Relations “career” and “serve as” had the same head and tail entities; therefore, the weight value of the edge between the two relations was w 12 r = 0.6 , according to Equations (1) and (2).
It is worth noting that the dashed and solid lines in Figure 5 represent the same correlation degree. Since all edges between any two relations in the dual-relation graph had equal weight values, it was sufficient to connect an edge directly. In order to gain insight into the construction of the dual-relation graph, all edges are marked in Figure 5, and the case of more than one edge is indicated by a dashed line. The use of a dual-relation graph was not only to strengthen the role of relations between entities in entity alignment but also to serve as auxiliary evidence for entity screening in the case of insufficient entity attribute information.

3.3.2. Neighboring-Entity Screening

In the task of entity alignment, it was difficult to obtain accurate matching results only by the information contained alongside the corresponding entities in two knowledge graphs. Therefore, it has been typical to rely on the entity’s neighboring entities to provide more useful information to support the implementation of entity alignment. However, the existing methods [26,27,28,29] directly overlooked situations where the head and tail entities had the same entity name in the relation triple when obtaining the neighboring entities of each entity, but this situation could exist in the real corpus. For example, with relational triples ( B o b , k n o w , B o b ) , it was obvious that the head and tail entities in this triple had the same name, but it could not be determined that the two entities referred to the same person or thing in the real world. What issues presented when we assumed that two entities were equivalent without comparing their overall information? As shown in Figure 6, we present an example.
In Figure 6, two knowledge graphs of different languages are shown. In the knowledge graph EN-KG, there were two entities whose entity names were both “ T o m ”. According to their attribute information, we found that the two entities were not equivalent. If we assumed that two entities were equivalent simply because their entity names were the same, then we would obtain a new knowledge graph for EN-KG’ after deleting one of the two entities. By the entity alignment of the knowledge graph EN-KG’ with the knowledge graph JA-KG, we were likely to obtain entity “雄猫” and entity “ T o m ” referring to the same thing. However, in the actual situation, the entity “ T o m ” in English and the entity “雄猫” in Japanese are not equivalent entities; therefore, the matching result obtained by the processing method in the figure was incorrect. In order to solve this issues, this study proposed a neighboring-entity-screening rule for joint entity names and attributes.
The neighboring-entity-screening rule for joints entity names and attributes was designed to remove neighboring entities that were equivalent to the central entity. Figure 7 shows the steps of the entity screening rule, for which the implementation steps were as follows (for ease of illustration, we set the IDs of the two “Tom” entities to “1” and “2”):
(1)
We first had to determine whether the name of the neighboring entity was the same as the central entity.
(2)
On the premise that the verification result in the first step was true (that is, the entity names of the two entities were the same), we compared the attribute information of the two entities.
As shown in Figure 7, if entity e 1 and entity e 2 had the same entity name and their attributes were the same (or there was an inclusion relation), then we considered the two entities to be equivalent and retained entity e 1 and deleted entity e 2 . If entities e 1 and e 2 were only the same entity name, and there was no inclusion relation between their attributes, we assumed that the two entities were not equivalent. However, in this scenario, there could be a special case where an entity would not have any attribute information, and we could then use the dual-relation graph as auxiliary evidence for screening. In the dual-relation graph, (1) when the sum of the correlation degrees of all corresponding relations of entity e 1 was equal to the sum of the correlation degrees of all corresponding relations of entity e 2 , we assumed that the two entities were equivalent. Conversely, (2) if the sum of the corresponding relations was not equal, then we considered the two entities as not equivalent.

3.3.3. Knowledge Embedding

(1). Knowledge Graph Structure Embedding and Entity Name Embedding. In order to learn knowledge graph structure embedding, we were inspired by the literature [41]. Therefore, the multi-order characteristics of a graph convolutional network were used to aggregate the higher degree neighboring structural information of entities. Using pre-trained entity name-embedding to initialize the input entity features of the GCN led to the model learning contextual information about adjacent structures.The GCN embedding structure of our method is shown in Figure 8. We embedded the knowledge graph structure and entity names into the GCN, so that each layer of the GCN contained entities, relations, and attributes of the knowledge graph, which also retained the original structures and entity names.
The knowledge graph structure embedding embedded the entities and relations in the knowledge graph into a continuous vector space, which was convenient for computing, and preserved the structural information in the knowledge graph. In order to facilitate data processing and retain the structural information of the knowledge graph, we used the primal graph as a large input graph, and each GCN layer used a set of entity features as input to update the entity feature representations, as shown in Equation (3).
x i ( l ) = R e L U ( k N i { i } C W ( l ) x k ( l 1 ) )
Among them, { x 1 ( l ) , x 2 ( l ) , x 3 ( l ) , . . . , x n ( l ) | x i ( l ) R d ( l ) ( 1 i n ) } ( d ( l ) represents the dimension of the l-th GCN layer) represents the output node of the l-th GCN layer, namely, the entity features; N i is the index set of the neighboring entities of the entity e i ; and C = 1 φ i is a constant while φ i represents the normalization constant; W ( l ) R d ( l ) × d ( l + 1 ) represents the weight matrix of the l-th layer; and i and k both represent the index value of the entity. We used the product of the constant C, the weight matrix, and the entity e k ( k N i { i } ) as the output node of the l 1 -th layer, and then all the obtained products were summed to obtain the output entity features of the l-th GCN layer.
(2). Dual-relation graph embedding. We built a dual-relation graph (DRG) to not only enable entity screening but also to enhance the role of relations in entity alignment so that relations were fully utilized. To further verify whether the DRG could improve the performance of the entity alignment method, we embedded the DRG in the same vector space as the knowledge graph. In order to distinguish it from x ( l ) , we denoted the output entity feature using the DRG as x ( l ) , which was calculated as:
x i ( l ) = R e L U ( C D ˜ 1 2 A ˜ D ˜ 1 2 W ( l ) x ( l 1 ) )
where A ˜ = A + I is the adjacency matrix of the primal graph K G p with added self-connections, I is an identity matrix as well the degree matrix D ˜ i i = j A ˜ i j (both i and j represent the index of the entity). We treated K G p as an un-directed graph when constructing A in order to allow the information to flow in both directions. We iteratively compared the results obtained by Equations (3) and (4) and then selected the maximum value as the output vector, as shown in Equation (5).
x i ( l ) = x i ( l ) · x i ( l )
Among them, x i ( l ) represents the output vector generated by the combination of structure embedding and entity-name embedding, x ( l ) represents the output vector iteratively generated by using the dual-relation graph, and · is a defined operation symbol that compares the values of two output vectors and outputs the maximum value.

3.3.4. Neighborhood Entity Sampling

The neighboring entities were key to determining whether an entity should be aligned with other entities. However, not all the neighboring entities contributed exactly to entity alignment. To extract useful neighboring entities, we used a down-sampling method (also known as an extraction method) to reduce the amount of calculation and prevent over-fitting in order to sample informative data from the neighboring entity of the central entity, as the neighboring entities of the central entity had already been processed by the entity-screening rules.
This study employed the pre-trained entity name-embedding to initialize the input node of the graph neural network, so that the entity-embedding learned by each layer of the GCN could contain rich structural and semantic information. The neighboring entities of the target entity were sampled through this information. In this way, neighboring entities that had been closer to the contextual semantic information of the target entity were more likely to be sampled. The contextual information of two equivalent entities in a real corpus is usually similar, so the stronger the correlation between the neighboring entity and the central entity, the more meaningful the information it contained, and the more clues it could provide for alignment. Therefore, we needed to sample neighboring entities with strong correlations to the central entity. In order to ensure the sampled neighboring entities were relevant, we defined a conditional probability to calculate the probability of the neighboring entity being sampled, and the adjacent entities with higher sampling probability ranking were selected as the candidate entity set.Formally, for a given arbitrary entity e i , the probability expression for its neighboring entity e i j to be sampled are shown in Equation (6).
p ( x i j x i ) = s o f t m a x ( x i W s x i j T ) = e x p ( x i W s x i j T ) k N e i e x p ( x i W s x i k T )
The softmax function is an activation function that uses an exponential function to convert a neural network output value into a probability value. The variables x i and x i j represented the learned embedding of entity e i and entity e i j , respectively, and i and j were indexes of entity e i and e j , respectively. The variable W s denoted the shared weight matrix, and N e i was the index of the neighboring entity of entity e i .
For each candidate entity in the knowledge graph, whether the neighboring entities of the entity and the neighboring entities that were closely related to the target entity could be found was critical to deciding whether to align the two entities. For an entity in the knowledge graph, if we wanted to find its equivalent entity in the target knowledge graph, we needed to compare its sampled neighboring entities with each candidate entity in the target knowledge graph in order to select the optimal matching entity.
In order to achieve the matching of neighboring entities, we used a low-cost approximate-matching method. Entities in E 2 that were closer to e i ( e i E 1 ) were more likely to align with e i in the embedding space. For an entity e j E 2 , its probability of being sampled as a candidate for e i is shown in Equation (7).
p ( x j x i ) = e x p ( x i x j L 1 ) k E 2 e x p ( x i x k L 1 )
Among them, x i x j L 1 used Manhattan distance and the L 1 -norm to calculate the distance between the learning embedding layers of entities e i and e j . The L 1 -norm was chosen because it streamlined the weights to prevent over-fitting.By selectively sampling the nearest neighboring entities, our approach essentially constructed a neighbor discriminant sub-graph for each entity. Through Equations (6) and (7), the entities with high similarity could be preliminarily obtained, which could effectively improve the matching accuracy and reduce the influence of noise.

3.3.5. Entity Alignment

In Section 3.3.4, we obtained the sub-graphs formed by entities with the higher similarities in the two knowledge graphs. We used the sub-graphs of the two knowledge graphs as input to then identify the entity-pairs that could be aligned. We obtained the sparse vector of neighboring entity-embeddings in the sub-graphs and then utilized a cross-graph model [32] for neighboring entity-matching. The cross-graph model [32] utilized the structural information of the knowledge graph, including the cross-graph structure, and embedded entities into a unified vector space to achieve alignment between the knowledge graphs. Formally, we defined ( e i , e i _ j ) as the entity-pair to be measured, e i E 1 and e i _ j E 2 , respectively. The variables n i and n i _ j represented the neighboring entities of e i and e i _ j , respectively. The probability of neighboring entities could be obtained from Equations (6) and (7), and the neighboring entity with the largest probability was then extracted. On the basis of the literature [10,27,41], we obtained the calculation formula of the cross-graph matching vector of the neighbor n i , as shown in Equations (8) and (9).
W n i n i _ j = e x p ( x n i · x n i _ j ) n i _ j N i _ j s e x p ( x n i · x n i _ j )
m n i = n i _ j N i _ j s W n i n i _ j ( x n i x n i _ j )
where W n i n i _ j represents the attention weight, m n i represents the matching vector of n i , and N i _ j s represents the set of sampled neighboring entities. The variables x n i and x n i _ j denote the GCN output embeddings of n i and n i _ j , respectively. Then, the GCN output embeddings of neighboring entities n i were concatenated by a weighted matching vector m n i as:
x n i = [ x n i β m n i ]
where [ a b ] represents the OR operation of a and b. Since the matching vector captured the difference between the two nearest neighboring entities, for each target neighboring entity in the neighboring entity, the attention mechanism could accurately detect the neighboring entity that was most likely to match the target neighboring entity in the entity of another knowledge graph.
Since the neighborhood sampling was based on the GCN output embedding, it was first necessary to pre-train the GCN-based knowledge-graph-embedding model to generate high-quality entity representations. Then, by measuring the distance between the two entities, we could determine if they should be aligned, which was calculated as:
d i s t ( e i , e j ) = x e i x e j L 1
Equation (11) uses Manhattan distance to calculate the distance between two entities to judge whether the two entities could be aligned.
During training, the distance between the aligned entity-pairs was expected to be as close as possible, and the distance between the negative entity-pairs was expected to be as far as possible; thus, an edge-based scoring function was used, which was calculated as shown in Equation (12).
L = ( i , j ) S ( i , j ) S m a x { 0 , d ( i , j ) d ( i , j ) + γ }
where γ > 0 is an edge hyper parameter, S is the alignment seed, S is the set of negatively aligned entity-pairs produced by nearest-neighbor sampling, and m a x { 0 , d ( i , j ) d ( i , j ) + γ } indicates the value cannot be less than 0.

4. Experimental Setup

4.1. Datasets

This study used DBP15K [10] as the experimental dataset. DBP15K was a subset of DBpedia and has been commonly used as a benchmark dataset for entity alignment. DBpedia was a special example of a semantic web application that extracted structured data from Wikipedia entries in order to enhance Wikipedia’s search capabilities and to link other datasets to Wikipedia. DBP15K contained three sub-datasets, namely Chinese–English (ZH-EN), French–English (FR-EN) and Japanese–English (JA-EN). Each of these datasets was constructed by extracting 15,000 aligned entity links from the multi-lingual version of Dbpedia and contained the knowledge in two languages. The statistics of the three datasets are shown in Table 1.
In this paper, we present only part of the data in DBP15K ZH-EN, and the data are presented in the format of the dataset, as shown in Figure 9. In Figure 9, ent_ids_1 and ent_ids_2 represent the correspondence between IDs and entities in the Chinese and English datasets, respectively. Furthermore, rel_ids_1 and rel_ids_2 represent the correspondence between id and relation in Chinese dataset and English dataset, respectively. In addition, triples_1 and triples_2 represent the triples in the Chinese and English datasets, respectively, which were composed of head entity ID, relation, and tail entity ID.

4.2. Comparison Model

Our approach was compared to several existing graph neural network-based approaches for cross-language entity alignment:
(1)
RDGCN [10]: A deep embedding-based technique that considered the attentive interaction between each KG and its dual-relation by passing them through a two-layer GCN with highway gates. The embeddings were then compared directly to obtain the alignment results [10,42].
(2)
JAPE [9]: A shallow embedding-based technique that generated structural (using a TransE model) and attribute embedding. The two types of embedding were then used simultaneously to compute the similarity score between the entities [9,42].
(3)
MUGNN [32]: A novel multi-channel graph neural-network model (MuGNN) that remembered alignment-oriented knowledge graph embeddings by robustly encoding two knowledge graphs through multiple channels.
(4)
KECG [34]: A semi-supervised entity-alignment method combining a knowledge-embedding model and cross-graph model in order to better utilize seed alignments in order to propagate the entire graphs under KG-based constraints.
(5)
GCN-Align [21]: A novel approach for cross-lingual knowledge-graph alignment based on graph convolutional networks that could learn embeddings from the structural and attribute information of entities and then combine the results to obtain accurate alignment.
(6)
NMN [41]: A novel entity-alignment framework, neighborhood-matching network, that captured the topology structure and neighborhood differences of entities by estimating the similarity between entities.
(7)
Dual-AMN [43]: A new KG encoder, dual-attention-matching network, that not only modeled both intra-graph and cross-graph information smartly, but also greatly reduced the computational complexity [43].
(8)
PSR [44]: A novel entity-alignment approach with three new components, which enabled high performance, high scalability, and high robustness.
It is worth noting that our method was the DRG+ESGCN, which was based on entity alignment with a dual-relation graph (DRG) and neighboring-entity-screening rule, combining entity name and attribute (NENA) data. In addition, we conducted an ablation experiment to evaluate the effectiveness of our proposed approach based on the dual-relation graph with neighboring-entity screening. In the ablation experiment, according to whether the experimental method adopted DRG or NENA, the variants of the DRG+ESGCN were obtained as the ESGCN and the DRGCN, respectively, where the ESGGN represented the entity alignment method using NENA-screening rules, and the DRGCN represented the entity alignment using DRG. In addition to this, we considered the baseline method NMN as a variant of the DRG+ESGCN to illustrate the effectiveness of our method. Since our method was an improvement based on the NMN method, it was more effective to illustrate the improvements generated by our method by using the NMN as a variant model.

4.3. Implementation Details

The configuration we used was as follows: β = 0.9, and γ = 1.0, and the neighbor-sampling stage sampled 5 entities for each entity. In order to set a reasonable number of sampling neighboring entities, this paper used samples 1, 5, 10, and 15 for comparison. The experimental results showed that it was optimal when the parameter was set to 5. The matching effect did not change significantly when the parameter was adjusted from 5 to 10, or even 15. However, considering that there were some entities whose number of neighbors would be less than 10, the number of sampled neighboring entities was set to 5. The dimensions of the embedding layer of the primal graph and the dual-relation graph were d = 300, d = 600, and d ˜ = 300. All dimensions of the hidden representations in GCN layers were 300. The learning rate was set to 0.005, and the number of negative samples of each positive one was set to k = 125 (as shown in Figure 10). In Figure 8, when k (k = 75, 100, 125, 150) were different values, the value of Hits@n (n = 1, 10) changed as k changes, and the peak value was reached when k = 125, which was the most appropriate value for k. This study, as well as the literature [9,10,21,32,34,41], used the same training–testing split: A total of 30% of the data were used for training and 70% of the data were used for testing.

4.4. Metrics

This study used Hits@k (k = 1, 10), a widely used metric in existing entity-alignment methods [9,10,21,32,34,41], to evaluate the performance of the entity alignment. A Hits@k score (higher was better) was computed by measuring the proportion of correctly aligned entities, as ranked at the top of the k list.Furthermore, we chose mean rank (MR) and mean reciprocal rank (MRR) as the evaluation metrics. Higher Hits@ k and MRR scores, as well as lower MR scores, indicated better performance.In comparison, Hits@1 represented the degree of matching with the most similar entity to the target entity, and it was equivalent to the precision of widely used conventional entity-alignment methods.

5. Results and Discussion

On the DBP15K dataset, we compared the performance of our approach with several of the models mentioned above, as shown in Table 2 and Table 3. It is worth noting that since, during the experiment, we divided the dataset into training and testing sets, in order to ensure the authenticity and accuracy of the experimental results, we cross-validated the divided training and testing sets on our model. Then, we averaged the results in order to finally obtain the data shown in Table 2 and Table 3.
As shown in Table 2, the DRG+ESGCN achieved better Hits@1 values on DBP15K ZH-EN, which was 0.024 higher than the NMN and 0.042 higher than the RDGCN. In the above comparison method, the RDGCN used the dual-relation graph, but the neighboring entity was not utilized efficiently. Furthermore, although the NMN used neighborhood sampling to optimize the selection of the neighboring entities, it overlooked the negative effect of entities with the same entity name on entity alignment. When comparing the DRG+ESSGCN with the PSR, we found that the PSR results were significantly better than our method, on the dataset DBP15K ZH-EN. However, on the datasets DBP15K FR-EN and DBP15K JA-EN, our method had a significant improvement, as compared to PSR. This showed that although our method outperformed the NMN, its ability to address the language differences between Chinese and English was still insufficient, and we needed to further improve its performance.
As shown in Table 3, as compared to the PSR, the MR and MRR values improved.Based on the methods of [9,10,19,30,32,39], higher MRR scores, combined with lower MR scores, indicated better performance. On the sub-dataset ZH-EN, the MR’s value with the DRG+ESGCN decreased by 9.674, as compared to that of the PSR, while the MRR value increased by 0.001. Furthermore, as compared to the NMN, the MRR value of the DRG+ESGCN was increased by 0.032. The results fully demonstrated the positive contribution of filtering out triples with a head entity name that was the same as the tail entity name for entity alignment.
The Ablation Experiment. Table 4 shows the results of the ablation experiment. It should be noted that in order to illustrate the advantages of the GCN, we exchanged the GCN for a GNN in our method and obtained a new variant, DRG+GNN. As expected, the performance of the different variants without the NENA-screening rule led to the DRG results decreasing. According to the results of the NMN, the ESGCN, and the DRGCN, we found that both the NENA-screening rules and the DRG have a significant improvement on the entity alignment, which also illustrated the effectiveness of our method. By comparing the results of the DRG+ESGNN and the DRG+ESGCN, we found that the results of the DRG+ESGCN were significantly better than those of the DRG+ESGNN. This also showed that the GCN had obvious advantages over the GNN, even under the same conditions.The GCN not only had multi-order characteristics, but it also effectively avoided the translation noise in the translation model. Our proposed NENA-screening rule used the entities and attributes to remove the negative impact of the equivalent entities, to a certain extent, and we used the DRG to strengthen the role of the relations, creating closer connections between entities, so that the entities, the relations, and the attributes could be fully utilized.
In order to further compare the performance of the DRG+ESGCN and the NMN, the specific gravity of the pre-aligned entity-pairs (denoted by SEED) was changed for comparison, as shown in Figure 11. As expected, the results of both models on the three sub-datasets gradually improved with the increasing proportion of previously aligned information. As shown in Figure 11a–c, the value of Hits@1 increased with the increase in SEED proportion. As compared to the sub-datasets JA-EN and FR-EN, the growth trend of the Hits@1 value of the ESGCN in sub-dataset ZH-EN was more obvious than that of the NMN, which was due to the significant language differences between Chinese and English. By the interaction between the dual-relation graph and the primal graph, the information of the neighboring entity was richer, and the connection with the central entity was closer, which reduced the influence of the language differences to a certain extent. As shown in Figure 11, our method reduced the noise caused by the differences between Chinese and English, to some extent, and significantly improved the Chinese–English entity alignment task.
In order to verify the validity of the results in Figure 11a–c, the MR and MRR values were obtained under different SEED values of specific gravity, as shown in Figure 11d–f. As shown in the figure, as the SEED value increased, the value of MR (MRR) continuously decreased (increased), and under the same conditions, the MR and MRR values of the ESGCN were better than those of the NMN. As shown in Figure 11, it was fully verified that our method made a significant contribution to the cross-language entity alignment task, and it also proved the effectiveness of the dual-relation graph and the entity screening for cross-language entity alignment.

6. Conclusions

This paper presented a neighboring-entity-screening rule based on the entity name and attributes for cross-language entity alignment. In addition, the dual-relation graph was used as auxiliary evidence when attribute information was insufficient. Our approach was designed to reduce the influence of noise caused by heterogeneity and linguistic differences between different knowledge graphs, and it effectively promoted the close associations between entities. Our method not only took full advantage of entities, relations, and attributes, but it also effectively eliminated the equivalent neighboring entities, yielding a significant improvement. We performed extensive experiments on real-world datasets and compared the DRG+ESGCN against seven recent GCN-based methods, and we conducted ablation experiments as well. The experimental results showed that the DRG+ESGCN achieved the best performance by consistently outperforming the competitive methods on the same datasets and evaluation metrics. The GCN not only has multi-order characteristics, but it also effectively avoids the interference of translation noise, giving it a significant advantage over GNNs and other methods in entity alignment tasks.
Although our method achieved superior results in cross-language entity alignment, there were still some limitations. In this study, we used a definition of equivalent entities that conformed to our proposed neighboring-entity-screening rules. Whether the attributes of the two entities had an inclusion relation was the basis for determining the equivalence of the two entities. However, in an actual situation, there could be an inclusion relation between the attributes of two entities, yet they still may not be equivalent; thus, this situation would requires deeper and more detailed verification.Although our method reduced the impact of noise caused by language differences to a certain extent, that was only a small improvement, and it could still not effectively mitigate the negative impact of this issue.
In our future work, we will introduce image information. Images contain rich information about entity features in a visual format.We will use multi-modal data that combines text and imagery to further reduce the negative impact of heterogeneity and language differences between knowledge graphs.

Author Contributions

Conceptualization, W.Z.; methodology, W.Z.; software, W.Z.; validation, W.Z., X.Z. and H.W.; formal analysis, W.Z.; investigation, W.Z.; resources, W.Z.; data curation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, W.Z. and X.Z.; visualization, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science and Technology Project of Hebei Education Department (Key Program), grant number ZD2021048, and Hebei Natural Science Foundation, grant number F2022208002.

Data Availability Statement

DBP15K can be downloaded at https://github.com/nju-websoft/JAPE (accessed on 2 December 2017).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, M.; Tian, Y.; Yang, M.; Zaniolo, C. Multilingual Knowledge Graph Embeddings for Cross-lingual Knowledge Alignment. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  2. Zeng, K.; Li, C.; Hou, L.; Li, J.; Feng, L. A comprehensive survey of entity alignment for knowledge graphs. AI Open 2021, 2, 1–13. [Google Scholar] [CrossRef]
  3. Zhang, R.; Trisedya, B.D.; Li, M.; Jiang, Y.; Qi, J. A benchmark and comprehensive survey on knowledge graph entity alignment via representation learning. VLDB J. 2022, 31, 1143–1168. [Google Scholar] [CrossRef]
  4. Jiang, T.; Bu, C.; Zhu, Y.; Wu, X. Combining embedding-based and symbol-based methods for entity alignment. Pattern Recognit. 2022, 124, 108433. [Google Scholar] [CrossRef]
  5. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Proceedings of the NIPS’13: Proceedings of the 26th International Conference on Neural Information Processing Systems, Stateline, NV, USA, 5–10 December 2013; Volume 26. [Google Scholar]
  6. Nayyeri, M.; Vahdati, S.; Lehmann, J.; Yazdi, H.S. Soft marginal transe for scholarly knowledge graph completion. arXiv 2019, arXiv:1904.12211. [Google Scholar]
  7. Ji, G.; He, S.; Xu, L.; Liu, K.; Zhao, J. Knowledge graph embedding via dynamic mapping matrix. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; pp. 687–696. [Google Scholar]
  8. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  9. Sun, Z.; Hu, W.; Li, C. Cross-lingual entity alignment via joint attribute-preserving embedding. In Proceedings of the International Semantic Web Conference, Vienna, Austria, 21–25 October 2017; pp. 628–644. [Google Scholar]
  10. Wu, Y.; Liu, X.; Feng, Y.; Wang, Z.; Yan, R.; Zhao, D. Relation-Aware Entity Alignment for Heterogeneous Knowledge Graphs. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macau, China, 10–16 August 2019. [Google Scholar]
  11. Ngomo, A.C.N.; Auer, S. LIMES—A time-efficient approach for large-scale link discovery on the web of data. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Catalonia, Spain, 16–22 July 2011. [Google Scholar]
  12. Niu, X.; Rong, S.; Wang, H.; Yu, Y. An effective rule miner for instance matching in a web of data. In Proceedings of the 21st ACM International Conference on Information and Knowledge Management, Maui, HI, USA, 29 October–2 November 2012; pp. 1085–1094. [Google Scholar]
  13. Yu, J.; Ma, X. English translation model based on intelligent recognition and deep learning. Wirel. Commun. Mob. Comput. 2022, 2022, 3079775. [Google Scholar] [CrossRef]
  14. Wang, X. Design of English Translation Model Based on Recurrent Neural Network. Math. Probl. Eng. 2022, 2022, 5177069. [Google Scholar] [CrossRef]
  15. Zhang, R.; Trisedy, B.D.; Li, M.; Jiang, Y.; Qi, J. A comprehensive survey on knowledge graph entity alignment via representation learning. arXiv 2021, arXiv:2103.15059. [Google Scholar]
  16. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  18. Chen, M.; Tian, Y.; Chang, K.-W.; Skiena, S.; Zaniolo, C. Co-training Embeddings of Knowledge Graphs and Entity Descriptions for Cross-Lingual Entity Alignment. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence Main Track, IJCAI 2018, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  19. Mao, X.; Wang, W.; Xu, H.; Wu, Y.; Lan, M. Relational reflection entity alignment. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Newport Beach, CA, USA, 17–21 October 2020; pp. 1095–1104. [Google Scholar]
  20. Mikolov, T.; Le, Q.V.; Sutskever, I. Exploiting Similarities among Languages for Machine Translation. arXiv 2013, arXiv:1309.4168. [Google Scholar]
  21. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014. [Google Scholar]
  22. Zhu, H.; Xie, R.; Liu, Z.; Sun, M. Iterative entity alignment via knowledge embeddings. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  23. Sun, Z.; Hu, W.; Zhang, Q.; Qu, Y. Bootstrapping Entity Alignment with Knowledge Graph Embedding. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence Main Track, IJCAI 2018, Stockholm, Sweden, 13–19 July 2018; pp. 4396–4402. [Google Scholar]
  24. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Berrendorf, M.; Faerman, E.; Melnychuk, V.; Tresp, V.; Seidl, T. Knowledge graph entity alignment with graph convolutional networks: Lessons learned. In Advances in Information Retrieval, Proceedings of the 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, 14–17 April 2020; Springer: Cham, Switzerland, 2020; pp. 3–11. [Google Scholar]
  26. Wang, Z.; Lv, Q.; Lan, X.; Zhang, Y. Cross-lingual knowledge graph alignment via graph convolutional networks. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 349–357. [Google Scholar]
  27. Akhtar, M.U.; Liu, J.; Xie, Z.; Liu, X.; Ahmed, S.; Huang, B. Entity alignment based on relational semantics augmentation for multi-lingual knowledge graphs. Knowl.-Based Syst. 2022, 252, 109494. [Google Scholar] [CrossRef]
  28. Jozashoori, S.; Sakor, A.; Iglesias, E.; Vidal, M.E. EABlock: A declarative entity alignment block for knowledge graph creation pipelines. In Proceedings of the 37th ACM/SIGAPP Symposium on Applied Computing, Newport Beach, CA, USA, 25–29 April 2022; pp. 1908–1916. [Google Scholar]
  29. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2019, arXiv:1609.02907. [Google Scholar]
  30. Xiong, F.; Gao, J. Entity Alignment for Cross-lingual Knowledge Graph with Graph Convolutional Networks. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macau, China, 10–16 August 2019; pp. 6480–6481. [Google Scholar]
  31. Guo, H.; Tang, J.; Zeng, W.; Zhao, X.; Liu, L. Multi-modal entity alignment in hyperbolic space. Neurocomputing 2021, 461, 598–607. [Google Scholar] [CrossRef]
  32. Cao, Y.; Liu, Z.; Li, C.; Li, J.; Chua, T.S. Multi-Channel Graph Neural Network for Entity Alignment. arXiv 2019, arXiv:1908.09898. [Google Scholar]
  33. Nguyen, T.T.; Huynh, T.T.; Yin, H. Entity alignment for knowledge graphs with multi-order convolutional networks. IEEE Trans. Knowl. Data Eng. 2020, 34, 4201–4214. [Google Scholar]
  34. Li, C.; Cao, Y.; Hou, L.; Shi, J.; Li, J.; Chua, T.S. Semi-supervised entity alignment via joint knowledge embedding model and cross-graph model. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 2723–2732. [Google Scholar]
  35. Huang, H.; Li, C.; Peng, X. Cross-knowledge-graph entity alignment via relation prediction. Knowl.-Based Syst. 2022, 240, 107813. [Google Scholar] [CrossRef]
  36. Mao, X.; Wang, W.; Xu, H.; Lan, M.; Wu, Y. MRAEA: An efficient and robust entity alignment approach for cross-lingual knowledge graph. In Proceedings of the 13th International Conference on Web Search and Data Mining, Houston, TX, USA, 3–7 February 2020; pp. 420–480. [Google Scholar]
  37. Chen, M.; Shi, W.; Zhou, B.; Roth, D. Cross-Lingual Entity Alignment with Incidental Supervision; EACL: Kiev, Ukraine, 2021; pp. 645–658. [Google Scholar]
  38. Zhu, Y.; Liu, H.; Wu, Z.; Du, Y. Relation-aware neighborhood matching model for entity alignment. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; pp. 4749–4756. [Google Scholar]
  39. Zeng, W.; Zhao, X.; Tang, J.; Lin, X.; Groth, P. Reinforcement Learning–based Collective Entity Alignment with Adaptive Features. ACM Trans. Inf. Syst. (TOIS) 2021, 39, 1–31. [Google Scholar] [CrossRef]
  40. Zhu, R.; Ma, M.; Wang, P. Raga: Relation-aware graph attention networks for global entity alignment. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Delhi, India, 11–14 May 2021; pp. 501–513. [Google Scholar]
  41. Wu, Y.; Liu, X.; Feng, Y.; Wang, Z.; Zhao, D. Neighborhood Matching Network for Entity Alignment. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6477–6487. [Google Scholar]
  42. Tam, N.T.; Trung, H.T.; Yin, H.; Vinh, T.V.; Sakong, D.; Zheng, B.; Hung, N.Q.V. Multi-order graph convolutional networks for knowledge graph alignment. In Proceedings of the 37th IEEE Int Conf Data Eng, Chania, Greece, 19–22 April 2021; pp. 19–22. [Google Scholar]
  43. Mao, X.; Wang, W.; Wu, Y.; Lan, M. Boosting the speed of entity alignment 10×: Dual attention matching network with normalized hard sample mining. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 821–832. [Google Scholar]
  44. Mao, X.; Wang, W.; Wu, Y.; Lan, M. From Alignment to Assignment: Frustratingly Simple Unsupervised Entity Alignment. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 2843–2853. [Google Scholar]
Figure 1. Chinese and English knowledge graph example.
Figure 1. Chinese and English knowledge graph example.
Electronics 12 01211 g001
Figure 2. An example of the description of triple entities in Freebase.
Figure 2. An example of the description of triple entities in Freebase.
Electronics 12 01211 g002
Figure 3. Problem definition.
Figure 3. Problem definition.
Electronics 12 01211 g003
Figure 4. Method overview. ZH-KG1 and EN-KG2 represent the Chinese knowledge graph and English knowledge graph, respectively.
Figure 4. Method overview. ZH-KG1 and EN-KG2 represent the Chinese knowledge graph and English knowledge graph, respectively.
Electronics 12 01211 g004
Figure 5. An example of constructing dual-relation graph. zh-KG1 and en-KG2 represent the Chinese knowledge graph and English knowledge graph, respectively.
Figure 5. An example of constructing dual-relation graph. zh-KG1 and en-KG2 represent the Chinese knowledge graph and English knowledge graph, respectively.
Electronics 12 01211 g005
Figure 6. Entity alignment of the knowledge graphs containing the same head and tail entity names. JA-KG and EN-KG represent the Japanese knowledge graph and English knowledge graph, respectively.
Figure 6. Entity alignment of the knowledge graphs containing the same head and tail entity names. JA-KG and EN-KG represent the Japanese knowledge graph and English knowledge graph, respectively.
Electronics 12 01211 g006
Figure 7. Entity screening rule.
Figure 7. Entity screening rule.
Electronics 12 01211 g007
Figure 8. The GCN embedding structure of our method.
Figure 8. The GCN embedding structure of our method.
Electronics 12 01211 g008
Figure 9. The data format of DBP15K ZH-EN (partial data).
Figure 9. The data format of DBP15K ZH-EN (partial data).
Electronics 12 01211 g009
Figure 10. Effect of different k values on Hits@n (n = 1, 10).
Figure 10. Effect of different k values on Hits@n (n = 1, 10).
Electronics 12 01211 g010
Figure 11. Performance comparison of the ESGCN and the NMN under different ratios of SEED. (a) ZH-EN (b) JA-EN (c) FR-EN (d) ZH-EN (e) JA-EN (f) FR-EN.
Figure 11. Performance comparison of the ESGCN and the NMN under different ratios of SEED. (a) ZH-EN (b) JA-EN (c) FR-EN (d) ZH-EN (e) JA-EN (f) FR-EN.
Electronics 12 01211 g011
Table 1. DBP15K dataset.
Table 1. DBP15K dataset.
DBP15KEntityRelationAttributeRel.TriplesAtt. Triples
ZH-ENZH66,46928308113153,929379,684
EN98,12523177173237,674567,755
JA-ENJA65,74420435882164,373354,619
EN95,68020966066233,319497,230
FR-ENFR66,85813794547192,191528,665
EN105,88922096422278,590576,543
Table 2. Performance on DBP15K dataset.
Table 2. Performance on DBP15K dataset.
ModelZH-ENJA-ENFR-EN
Hits@1Hits@10Hits@1Hits@10Hits@1Hits@10
JAPE0.40780.73210.37230.68190.32210.6658
KECG0.47700.83500.48470.84990.49290.8442
MUGNN0.47730.84210.48660.85730.48900.8681
RDGCN0.71050.85290.77900.90690.88830.9602
NMN0.73300.86050.78610.90130.90310.9662
Dual−AMN0.74030.90190.75980.94900.72930.9284
PSR0.80240.91400.73100.93110.80330.9380
DRG+ESGCN0.75700.90730.80700.93300.97010.9730
Table 3. Performance on DBP15K dataset (MR/MRR).
Table 3. Performance on DBP15K dataset (MR/MRR).
ModelZH-ENJA-ENFR-EN
MRMRRMRMRRMRMRR
JAPE640.490990.476920.430
KECG71.8020.59859.7060.61141.9250.609
RDGCN68.8290.76345.7280.82517.6640.915
Dual−AMN28.6300.80511.7970.83020.0560.801
PSR11.4560.81010.9310.8447.5320.852
ESGCN1.5370.7891.5110.8341.3330.927
DRG+ESGCN1.7820.8111.6150.8531.3590.927
Table 4. The comparison results of the DRG+ESGCN and its variants in the ablation experiments.
Table 4. The comparison results of the DRG+ESGCN and its variants in the ablation experiments.
ModelZH-ENJA-ENFR-EN
Hits@10MRRHits@10MRRHits@10MRR
NMN0.86050.7990.90310.8270.96620.926
ESGCN0.87860.7890.90990.8340.96890.927
DRGCN0.86790.7910.90540.8300.96730.926
DRG+ESGNN0.83250.6850.88100.7610.92770.839
DRG+ESGCN0.90730.8110.93300.8530.97300.927
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Zhang, W.; Wang, H. Cross-Language Entity Alignment Based on Dual-Relation Graph and Neighbor Entity Screening. Electronics 2023, 12, 1211. https://doi.org/10.3390/electronics12051211

AMA Style

Zhang X, Zhang W, Wang H. Cross-Language Entity Alignment Based on Dual-Relation Graph and Neighbor Entity Screening. Electronics. 2023; 12(5):1211. https://doi.org/10.3390/electronics12051211

Chicago/Turabian Style

Zhang, Xiaoming, Wencheng Zhang, and Huiyong Wang. 2023. "Cross-Language Entity Alignment Based on Dual-Relation Graph and Neighbor Entity Screening" Electronics 12, no. 5: 1211. https://doi.org/10.3390/electronics12051211

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop